ai-avatar-video — community ai-avatar-video, Procedio-AI, community, ide skills, Claude Code, Cursor, Windsurf

v1.0.0
GitHub

About this Skill

Ideal for Media Agents needing advanced AI avatar and talking head video generation capabilities via CLI. Procedio AI

RomainGRAS42 RomainGRAS42
[0]
[0]
Updated: 3/5/2026

Agent Capability Analysis

The ai-avatar-video skill by RomainGRAS42 is an open-source community AI agent skill for Claude Code and other IDE workflows, helping agents execute tasks with better context, repeatability, and domain-specific guidance.

Ideal Agent Persona

Ideal for Media Agents needing advanced AI avatar and talking head video generation capabilities via CLI.

Core Value

Empowers agents to create AI avatars and talking head videos using inference.sh CLI, supporting image and audio inputs via protocols like HTTPS, and providing a seamless video creation experience with libraries like bytedance/omnihuman-1-5.

Capabilities Granted for ai-avatar-video

Generating AI avatar videos from image and audio inputs
Creating talking head videos for social media platforms
Automating video content creation for marketing campaigns

! Prerequisites & Limits

  • Requires inference.sh CLI installation and login
  • Dependent on bytedance/omnihuman-1-5 library
  • Needs image and audio URLs as input
Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

ai-avatar-video

Install ai-avatar-video, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly

AI Avatar & Talking Head Videos

Create AI avatars and talking head videos via inference.sh CLI.

AI Avatar & Talking Head Videos

Quick Start

bash
1curl -fsSL https://cli.inference.sh | sh && infsh login 2 3# Create avatar video from image + audio 4infsh app run bytedance/omnihuman-1-5 --input '{ 5 "image_url": "https://portrait.jpg", 6 "audio_url": "https://speech.mp3" 7}'

Available Models

ModelApp IDBest For
OmniHuman 1.5bytedance/omnihuman-1-5Multi-character, best quality
OmniHuman 1.0bytedance/omnihuman-1-0Single character
Fabric 1.0falai/fabric-1-0Image talks with lipsync
PixVerse Lipsyncfalai/pixverse-lipsyncHighly realistic

Search Avatar Apps

bash
1infsh app list --search "omnihuman" 2infsh app list --search "lipsync" 3infsh app list --search "fabric"

Examples

OmniHuman 1.5 (Multi-Character)

bash
1infsh app run bytedance/omnihuman-1-5 --input '{ 2 "image_url": "https://portrait.jpg", 3 "audio_url": "https://speech.mp3" 4}'

Supports specifying which character to drive in multi-person images.

Fabric 1.0 (Image Talks)

bash
1infsh app run falai/fabric-1-0 --input '{ 2 "image_url": "https://face.jpg", 3 "audio_url": "https://audio.mp3" 4}'

PixVerse Lipsync

bash
1infsh app run falai/pixverse-lipsync --input '{ 2 "image_url": "https://portrait.jpg", 3 "audio_url": "https://speech.mp3" 4}'

Generates highly realistic lipsync from any audio.

Full Workflow: TTS + Avatar

bash
1# 1. Generate speech from text 2infsh app run infsh/kokoro-tts --input '{ 3 "text": "Welcome to our product demo. Today I will show you..." 4}' > speech.json 5 6# 2. Create avatar video with the speech 7infsh app run bytedance/omnihuman-1-5 --input '{ 8 "image_url": "https://presenter-photo.jpg", 9 "audio_url": "<audio-url-from-step-1>" 10}'

Full Workflow: Dub Video in Another Language

bash
1# 1. Transcribe original video 2infsh app run infsh/fast-whisper-large-v3 --input '{"audio_url": "https://video.mp4"}' > transcript.json 3 4# 2. Translate text (manually or with an LLM) 5 6# 3. Generate speech in new language 7infsh app run infsh/kokoro-tts --input '{"text": "<translated-text>"}' > new_speech.json 8 9# 4. Lipsync the original video with new audio 10infsh app run infsh/latentsync-1-6 --input '{ 11 "video_url": "https://original-video.mp4", 12 "audio_url": "<new-audio-url>" 13}'

Use Cases

  • Marketing: Product demos with AI presenter
  • Education: Course videos, explainers
  • Localization: Dub content in multiple languages
  • Social Media: Consistent virtual influencer
  • Corporate: Training videos, announcements

Tips

  • Use high-quality portrait photos (front-facing, good lighting)
  • Audio should be clear with minimal background noise
  • OmniHuman 1.5 supports multiple people in one image
  • LatentSync is best for syncing existing videos to new audio
bash
1# Full platform skill (all 150+ apps) 2npx skills add inferencesh/skills@inference-sh 3 4# Text-to-speech (generate audio for avatars) 5npx skills add inferencesh/skills@text-to-speech 6 7# Speech-to-text (transcribe for dubbing) 8npx skills add inferencesh/skills@speech-to-text 9 10# Video generation 11npx skills add inferencesh/skills@ai-video-generation 12 13# Image generation (create avatar images) 14npx skills add inferencesh/skills@ai-image-generation

Browse all video apps: infsh app list --category video

Documentation

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is ai-avatar-video?

Ideal for Media Agents needing advanced AI avatar and talking head video generation capabilities via CLI. Procedio AI

How do I install ai-avatar-video?

Run the command: npx killer-skills add RomainGRAS42/Procedio-AI. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for ai-avatar-video?

Key use cases include: Generating AI avatar videos from image and audio inputs, Creating talking head videos for social media platforms, Automating video content creation for marketing campaigns.

Which IDEs are compatible with ai-avatar-video?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for ai-avatar-video?

Requires inference.sh CLI installation and login. Dependent on bytedance/omnihuman-1-5 library. Needs image and audio URLs as input.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add RomainGRAS42/Procedio-AI. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use ai-avatar-video immediately in the current project.

Related Skills

Looking for an alternative to ai-avatar-video or another community skill for your workflow? Explore these related open-source skills.

View All

widget-generator

Logo of f
f

f.k.a. Awesome ChatGPT Prompts. Share, discover, and collect prompts from the community. Free and open source — self-host for your organization with complete privacy.

149.6k
0
AI

flags

Logo of vercel
vercel

flags is a Next.js feature management skill that enables developers to efficiently add or modify framework feature flags, streamlining React application development.

138.4k
0
Browser

zustand

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
AI

data-fetching

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
AI