fal-ai-media — Claude Code MCP fal-ai-media, everything-claude-code, official, Claude Code MCP, ide skills, media generation workflow, automation with fal.ai, browser automation for developers, install fal-ai-media skill, AI-powered media creation, Claude Code

Verified
v1.0.0
GitHub

About this Skill

Ideal for Media Generation Agents requiring unified content creation capabilities via fal.ai MCP, covering text-to-image, text-to-video, text-to-speech, and video-to-audio tasks. fal-ai-media is a unified media generation skill, creating images, videos, and audio using fal.ai models and MCP.

Features

Generate images using Nano Banana
Create videos from text or images with Seedance, Kling, and Veo 3
Convert text to speech with CSM-1B
Extract audio from videos using ThinkSound
Utilize MCP for seamless media generation

# Core Topics

affaan-m affaan-m
[105.1k]
[13661]
Updated: 3/25/2026

Agent Capability Analysis

The fal-ai-media skill by affaan-m is an open-source official AI agent skill for Claude Code and other IDE workflows, helping agents execute tasks with better context, repeatability, and domain-specific guidance. Optimized for Claude Code MCP, media generation workflow, automation with fal.ai.

Ideal Agent Persona

Ideal for Media Generation Agents requiring unified content creation capabilities via fal.ai MCP, covering text-to-image, text-to-video, text-to-speech, and video-to-audio tasks.

Core Value

Empowers agents to generate images, videos, and audio using fal.ai models, including Nano Banana for text-to-image, Seedance, Kling, and Veo 3 for text/image-to-video, and CSM-1B for text-to-speech, all via the MCP protocol, enabling advanced media generation capabilities with support for various formats and libraries.

Capabilities Granted for fal-ai-media

Generating images from text prompts using Nano Banana
Creating videos from text or images with Seedance, Kling, and Veo 3
Producing speech, music, or sound effects with CSM-1B and ThinkSound
Automating thumbnail creation for videos and images

! Prerequisites & Limits

  • Requires fal.ai MCP server configuration
  • Needs addition of fal.ai settings to ~/.claude.json
  • Dependent on specific fal.ai models and libraries
Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

fal-ai-media

Install fal-ai-media, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly

fal.ai Media Generation

Generate images, videos, and audio using fal.ai models via MCP.

When to Activate

  • User wants to generate images from text prompts
  • Creating videos from text or images
  • Generating speech, music, or sound effects
  • Any media generation task
  • User says "generate image", "create video", "text to speech", "make a thumbnail", or similar

MCP Requirement

fal.ai MCP server must be configured. Add to ~/.claude.json:

json
1"fal-ai": { 2 "command": "npx", 3 "args": ["-y", "fal-ai-mcp-server"], 4 "env": { "FAL_KEY": "YOUR_FAL_KEY_HERE" } 5}

Get an API key at fal.ai.

MCP Tools

The fal.ai MCP provides these tools:

  • search — Find available models by keyword
  • find — Get model details and parameters
  • generate — Run a model with parameters
  • result — Check async generation status
  • status — Check job status
  • cancel — Cancel a running job
  • estimate_cost — Estimate generation cost
  • models — List popular models
  • upload — Upload files for use as inputs

Image Generation

Nano Banana 2 (Fast)

Best for: quick iterations, drafts, text-to-image, image editing.

generate(
  app_id: "fal-ai/nano-banana-2",
  input_data: {
    "prompt": "a futuristic cityscape at sunset, cyberpunk style",
    "image_size": "landscape_16_9",
    "num_images": 1,
    "seed": 42
  }
)

Nano Banana Pro (High Fidelity)

Best for: production images, realism, typography, detailed prompts.

generate(
  app_id: "fal-ai/nano-banana-pro",
  input_data: {
    "prompt": "professional product photo of wireless headphones on marble surface, studio lighting",
    "image_size": "square",
    "num_images": 1,
    "guidance_scale": 7.5
  }
)

Common Image Parameters

ParamTypeOptionsNotes
promptstringrequiredDescribe what you want
image_sizestringsquare, portrait_4_3, landscape_16_9, portrait_16_9, landscape_4_3Aspect ratio
num_imagesnumber1-4How many to generate
seednumberany integerReproducibility
guidance_scalenumber1-20How closely to follow the prompt (higher = more literal)

Image Editing

Use Nano Banana 2 with an input image for inpainting, outpainting, or style transfer:

# First upload the source image
upload(file_path: "/path/to/image.png")

# Then generate with image input
generate(
  app_id: "fal-ai/nano-banana-2",
  input_data: {
    "prompt": "same scene but in watercolor style",
    "image_url": "<uploaded_url>",
    "image_size": "landscape_16_9"
  }
)

Video Generation

Seedance 1.0 Pro (ByteDance)

Best for: text-to-video, image-to-video with high motion quality.

generate(
  app_id: "fal-ai/seedance-1-0-pro",
  input_data: {
    "prompt": "a drone flyover of a mountain lake at golden hour, cinematic",
    "duration": "5s",
    "aspect_ratio": "16:9",
    "seed": 42
  }
)

Kling Video v3 Pro

Best for: text/image-to-video with native audio generation.

generate(
  app_id: "fal-ai/kling-video/v3/pro",
  input_data: {
    "prompt": "ocean waves crashing on a rocky coast, dramatic clouds",
    "duration": "5s",
    "aspect_ratio": "16:9"
  }
)

Veo 3 (Google DeepMind)

Best for: video with generated sound, high visual quality.

generate(
  app_id: "fal-ai/veo-3",
  input_data: {
    "prompt": "a bustling Tokyo street market at night, neon signs, crowd noise",
    "aspect_ratio": "16:9"
  }
)

Image-to-Video

Start from an existing image:

generate(
  app_id: "fal-ai/seedance-1-0-pro",
  input_data: {
    "prompt": "camera slowly zooms out, gentle wind moves the trees",
    "image_url": "<uploaded_image_url>",
    "duration": "5s"
  }
)

Video Parameters

ParamTypeOptionsNotes
promptstringrequiredDescribe the video
durationstring"5s", "10s"Video length
aspect_ratiostring"16:9", "9:16", "1:1"Frame ratio
seednumberany integerReproducibility
image_urlstringURLSource image for image-to-video

Audio Generation

CSM-1B (Conversational Speech)

Text-to-speech with natural, conversational quality.

generate(
  app_id: "fal-ai/csm-1b",
  input_data: {
    "text": "Hello, welcome to the demo. Let me show you how this works.",
    "speaker_id": 0
  }
)

ThinkSound (Video-to-Audio)

Generate matching audio from video content.

generate(
  app_id: "fal-ai/thinksound",
  input_data: {
    "video_url": "<video_url>",
    "prompt": "ambient forest sounds with birds chirping"
  }
)

ElevenLabs (via API, no MCP)

For professional voice synthesis, use ElevenLabs directly:

python
1import os 2import requests 3 4resp = requests.post( 5 "https://api.elevenlabs.io/v1/text-to-speech/<voice_id>", 6 headers={ 7 "xi-api-key": os.environ["ELEVENLABS_API_KEY"], 8 "Content-Type": "application/json" 9 }, 10 json={ 11 "text": "Your text here", 12 "model_id": "eleven_turbo_v2_5", 13 "voice_settings": {"stability": 0.5, "similarity_boost": 0.75} 14 } 15) 16with open("output.mp3", "wb") as f: 17 f.write(resp.content)

VideoDB Generative Audio

If VideoDB is configured, use its generative audio:

python
1# Voice generation 2audio = coll.generate_voice(text="Your narration here", voice="alloy") 3 4# Music generation 5music = coll.generate_music(prompt="upbeat electronic background music", duration=30) 6 7# Sound effects 8sfx = coll.generate_sound_effect(prompt="thunder crack followed by rain")

Cost Estimation

Before generating, check estimated cost:

estimate_cost(
  estimate_type: "unit_price",
  endpoints: {
    "fal-ai/nano-banana-pro": {
      "unit_quantity": 1
    }
  }
)

Model Discovery

Find models for specific tasks:

search(query: "text to video")
find(endpoint_ids: ["fal-ai/seedance-1-0-pro"])
models()

Tips

  • Use seed for reproducible results when iterating on prompts
  • Start with lower-cost models (Nano Banana 2) for prompt iteration, then switch to Pro for finals
  • For video, keep prompts descriptive but concise — focus on motion and scene
  • Image-to-video produces more controlled results than pure text-to-video
  • Check estimate_cost before running expensive video generations
  • videodb — Video processing, editing, and streaming
  • video-editing — AI-powered video editing workflows
  • content-engine — Content creation for social platforms

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is fal-ai-media?

Ideal for Media Generation Agents requiring unified content creation capabilities via fal.ai MCP, covering text-to-image, text-to-video, text-to-speech, and video-to-audio tasks. fal-ai-media is a unified media generation skill, creating images, videos, and audio using fal.ai models and MCP.

How do I install fal-ai-media?

Run the command: npx killer-skills add affaan-m/everything-claude-code/fal-ai-media. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for fal-ai-media?

Key use cases include: Generating images from text prompts using Nano Banana, Creating videos from text or images with Seedance, Kling, and Veo 3, Producing speech, music, or sound effects with CSM-1B and ThinkSound, Automating thumbnail creation for videos and images.

Which IDEs are compatible with fal-ai-media?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for fal-ai-media?

Requires fal.ai MCP server configuration. Needs addition of fal.ai settings to ~/.claude.json. Dependent on specific fal.ai models and libraries.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add affaan-m/everything-claude-code/fal-ai-media. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use fal-ai-media immediately in the current project.

Related Skills

Looking for an alternative to fal-ai-media or another official skill for your workflow? Explore these related open-source skills.

View All

flags

Logo of facebook
facebook

Use when you need to check feature flag states, compare channels, or debug why a feature behaves differently across release channels.

243.6k
0
Developer

extract-errors

Logo of facebook
facebook

Use when adding new error messages to React, or seeing unknown error code warnings.

243.6k
0
Developer

fix

Logo of facebook
facebook

Use when you have lint errors, formatting issues, or before committing code to ensure it passes CI.

243.6k
0
Developer

flow

Logo of facebook
facebook

Use when you need to run Flow type checking, or when seeing Flow type errors in React code.

243.6k
0
Developer