video-generation
Verifiedby bytedance · Updated Apr 16, 2026
Use this skill when the user requests to generate, create, or imagine videos. Supports structured prompts and reference image for guided generation.
See It In Action
Let me find the reference frame in the project first.
Found the reference frame at hero-button-reference.png. Let me invoke the video generation skill.
Let me view the reference image first, then set up the prompt and generate.
Nice — a teal-to-purple gradient reference frame. I'll create the prompt file and run generation.
The /mnt paths aren't available on this system. Let me use local paths instead.
Now let me check for the generation script and run it.
Found the script locally. Let me run the generation.
Video generated. Let me confirm the output.
Your hero video is ready at outputs/hero-button-morph.mp4.
What was generated:
- 6-second clip, 16:9 aspect ratio
- Floating pill-shaped button with the teal-to-purple gradient from your reference frame
- Smooth morph sequence: breathing button (0-1s) → elastic scale-up with emerging UI elements (1-3s) → full design system component card assembly (3-5s) → shine sweep + hero-shot hold (5-6s)
- Clean light neutral background with soft shadows and depth
Good to go for the design systems review this afternoon. Let me know if you want any tweaks to the motion or styling.
External Tools
| Tool | Type |
|---|---|
| python3 | binary |
Permissions
| Scope | Description |
|---|---|
| filesystem:read | |
| filesystem:write | |
| network:outbound | |
| process:spawn |
SKILL.md
Video Generation Skill
Overview
This skill generates high-quality videos using structured prompts and a Python script. The workflow includes creating JSON-formatted prompts and executing video generation with optional reference image.
Core Capabilities
- Create structured JSON prompts for AIGC video generation
- Support reference image as guidance or the first/last frame of the video
- Generate videos through automated Python script execution
Workflow
Step 1: Understand Requirements
When a user requests video generation, identify:
- Subject/content: What should be in the image
- Style preferences: Art style, mood, color palette
- Technical specs: Aspect ratio, composition, lighting
- Reference image: Any image to guide generation
- You don't need to check the folder under
/mnt/user-data
Step 2: Create Structured Prompt
Generate a structured JSON file in /mnt/user-data/workspace/ with naming pattern: {descriptive-name}.json
Step 3: Create Reference Image (Optional when image-generation skill is available)
Generate reference image for the video generation.
- If only 1 image is provided, use it as the guided frame of the video
Step 3: Execute Generation
Call the Python script:
python /mnt/skills/public/video-generation/scripts/generate.py \
--prompt-file /mnt/user-data/workspace/prompt-file.json \
--reference-images /path/to/ref1.jpg \
--output-file /mnt/user-data/outputs/generated-video.mp4 \
--aspect-ratio 16:9
Parameters:
--prompt-file: Absolute path to JSON prompt file (required)--reference-images: Absolute paths to reference image (optional)--output-file: Absolute path to output image file (required)--aspect-ratio: Aspect ratio of the generated image (optional, default: 16:9)
[!NOTE] Do NOT read the python file, instead just call it with the parameters.
Video Generation Example
User request: "Generate a short video clip depicting the opening scene from "The Chronicles of Narnia: The Lion, the Witch and the Wardrobe"
Step 1: Search for the opening scene of "The Chronicles of Narnia: The Lion, the Witch and the Wardrobe" online
Step 2: Create a JSON prompt file with the following content:
{
"title": "The Chronicles of Narnia - Train Station Farewell",
"background": {
"description": "World War II evacuation scene at a crowded London train station. Steam and smoke fill the air as children are being sent to the countryside to escape the Blitz.",
"era": "1940s wartime Britain",
"location": "London railway station platform"
},
"characters": ["Mrs. Pevensie", "Lucy Pevensie"],
"camera": {
"type": "Close-up two-shot",
"movement": "Static with subtle handheld movement",
"angle": "Profile view, intimate framing",
"focus": "Both faces in focus, background soft bokeh"
},
"dialogue": [
{
"character": "Mrs. Pevensie",
"text": "You must be brave for me, darling. I'll come for you... I promise."
},
{
"character": "Lucy Pevensie",
"text": "I will be, mother. I promise."
}
],
"audio": [
{
"type": "Train whistle blows (signaling departure)",
"volume": 1
},
{
"type": "Strings swell emotionally, then fade",
"volume": 0.5
},
{
"type": "Ambient sound of the train station",
"volume": 0.5
}
]
}
Step 3: Use the image-generation skill to generate the reference image
Load the image-generation skill and generate a single reference image narnia-farewell-scene-01.jpg according to the skill.
Step 4: Use the generate.py script to generate the video
python /mnt/skills/public/video-generation/scripts/generate.py \
--prompt-file /mnt/user-data/workspace/narnia-farewell-scene.json \
--reference-images /mnt/user-data/outputs/narnia-farewell-scene-01.jpg \
--output-file /mnt/user-data/outputs/narnia-farewell-scene-01.mp4 \
--aspect-ratio 16:9
Do NOT read the python file, just call it with the parameters.
Output Handling
After generation:
- Videos are typically saved in
/mnt/user-data/outputs/ - Share generated videos (come first) with user as well as generated image if applicable, using
present_filestool - Provide brief description of the generation result
- Offer to iterate if adjustments needed
Notes
- Always use English for prompts regardless of user's language
- JSON format ensures structured, parsable prompts
- Reference image enhance generation quality significantly
- Iterative refinement is normal for optimal results
FAQ
What does video-generation do?
Use this skill when the user requests to generate, create, or imagine videos. Supports structured prompts and reference image for guided generation.
When should I use video-generation?
Use it when you need a repeatable workflow that produces text report, downloadable file.
What does video-generation output?
In the evaluated run it produced text report, downloadable file.
How do I install or invoke video-generation?
Ask the agent to use this skill when the task matches its documented workflow.
Which agents does video-generation support?
Agent support is inferred from the source, but not explicitly declared.
What tools, channels, or permissions does video-generation need?
It uses python3; channels commonly include text, file; permissions include filesystem:read, filesystem:write, network:outbound, process:spawn.
Is video-generation safe to install?
Static analysis marked this skill as medium risk; review side effects and permissions before enabling it.
How is video-generation different from an MCP or plugin?
A skill packages instructions and workflow conventions; tools, MCP servers, and plugins are dependencies the skill may call during execution.
Does video-generation outperform not using a skill?
About video-generation
When to use video-generation
When a user wants a generated video clip from a described scene or concept. When you need to create a structured prompt file and run the bundled generator script. When reference images should guide the visual result of the generated video.
When video-generation is not the right choice
When the task is to edit existing video footage rather than generate a new one. When no Python runtime is available to execute the bundled generation script.
What it produces
Produces text report and downloadable file.