English | ไธญๆ | ๆฅๆฌ่ช | ํ๊ตญ์ด | Espaรฑol | Portuguรชs | Franรงais
Cinematic Shot Language Designer
Turn your vague video ideas into cinema-grade video prompts ready for Jimeng Seedance 2.0 โ in one shot.
A Claude Skill built on the Agent Skills specification, blending Hollywood-level cinematography aesthetics with Chinese film industry practices. It's designed to help creators break free from the "looks nice but random" AI video trap and achieve precise, controllable visual storytelling.
| Capability | Description |
|---|---|
| ๐ญ AI Comic Drama & Short Drama Production | Full-pipeline support for AI comic strips (ๆผซๅง) and AI short dramas โ character dialogue / voiceover / actor blocking / exaggerated expression close-ups / narrative-motivated camera / short drama style quick-selector / 4 prompt template variants (CN/EN ร dialogue/voiceover), with dedicated scenario templates and complete examples |
| ๐จ 28+ Director & Style Presets | Nolan / Villeneuve / Fincher / Deakins / Kurosawa / Makoto Shinkai / Wong Kar-wai / Zhang Yimou / Xianxia / Cel-Shaded CG / Anime / Xiaohongshuโฆ |
| ๐ฌ Pro Camera Movement Dictionary | 3-tier camera system + 14 focal lengths + 6 focus controls + 7 physical mounts, with bilingual CN/EN references |
| ๐ก Three-Layer Lighting Structure | Light Source โ Light Behavior โ Color Tone โ no more vague "add a light" |
| ๐ Timestamped Storyboarding | 0-3s / 3-8s / โฆ precise timeline control to prevent visual bleeding between shots |
| ๐ฏ Six-Element Precision Assembly | Subject / Action / Scene / Lighting / Camera / Sound โ a structured, high-conversion formula |
| ๐ฌ Smart Multi-Segment Storyboard | Videos >15s are automatically split into independent prompt segments with unified style, lighting, sound, and seamless transition frames |
| ๐ฆ 20 Scenario Templates | E-commerce / Xianxia / Short Drama / Food / MV / One-Take / Automotive / Macro / Nature / Game PV / Horror / Travel / Pets / Transformation / Loop / Video Editing / Video Extension / Story Completion / Multiframe Storytelling |
| ๐ต Sound & ASMR Vocabulary | Physics-based onomatopoeia library covering ambient / action / vocal / music sounds |
| ๐ค Voice & Language Control | Timbre cloning via video reference, dialect/accent control (Sichuan/Cantonese/Northeast/Taiwanese etc.), multilingual dialogue mixing, special voice styles (documentary/stand-up/opera/ASMR) |
| ๐น Multimodal Reference Guide | 6 core reference patterns (first frame / camera replication / action replication / camera+action separation / timbre reference / effect replication), multi-asset character control, consistency preservation |
| ๐ Bilingual Prompt Output | Chinese users โ Chinese prompts, non-Chinese users โ English prompts, auto-detected |
| ๐ก๏ธ Copyright-Safe IP Fallback | Three-tier progressive IP fallback strategy to prevent platform content blocks |
| ๐ Structured Hard Validation | Word count / camera moves / temporal logic / filler detection / optical physics conflicts / style conflict matrix โ 7-rule checklist applied before every delivery |
| ๐ CLI Integration | Jimeng CLI command mapping (text2video / image2video / multiframe2video / multimodal2video), async task management, VIP channel routing |
| ๐๏ธ Multiframe Storytelling | Upload 2-9 keyframe images โ engine auto-composes coherent story video via multiframe2video, with decision matrix for choosing multiframe vs. multi-segment storyboard |
Claude Code
Place the seedance-shot-design/ folder under .claude/skills/ in your project root:
# Clone into your project's Skill directory
git clone https://github.com/woodfantasy/Seedance2.0-ShotDesign-Skills.git .claude/skills/seedance-shot-designClaude Code will automatically detect and load the Skill.
OpenClaw
In your connected IM app (WeChat, Feishu, etc.), send a message to the OpenClaw Agent:
Please learn this skill: https://github.com/woodfantasy/Seedance2.0-ShotDesign-Skills
The Agent will fetch and learn the Seedance Shot Design skill automatically. You can start making requests right away.
Codex
Place the Skill folder under Codex's agents instruction directory:
git clone https://github.com/woodfantasy/Seedance2.0-ShotDesign-Skills.git agents/skills/seedance-shot-designThen invoke it within a Codex conversation.
Cursor
Place the Skill folder under .cursor/skills/ in your project root:
git clone https://github.com/woodfantasy/Seedance2.0-ShotDesign-Skills.git .cursor/skills/seedance-shot-designCursor Agent mode will automatically read the Skill instructions.
Just tell Claude:
Write a 15-second cyberpunk rain chase video prompt
The Skill auto-activates and generates the prompt through a 5-step workflow:
- Requirement Analysis โ Confirm duration / aspect ratio / assets / style
- Visual Diagnosis โ Select camera language & director style
- Six-Element Assembly โ Compose the prompt using the structured formula
- Mandatory Validation โ Apply 7-rule structured validation checklist
- Professional Delivery โ Director's statement + complete prompt
User Input:
Write a 10-second Eastern Xianxia short film video prompt
Skill Output:
Seedance Video Prompt
Theme: A young man in white catches a red leaf in a misty ancient temple at dawn, reaching enlightenment
Director's Statement (for creative intent only โ do not copy): Uses an aerial โ dolly โ slow push three-phase camera progression, transitioning from grand vista to intimate emotion. 35mm film grain adds a handcrafted texture, and the gold-teal palette echoes the Eastern philosophy of harmony with nature.
Full Prompt (copy directly into Jimeng's input field):
10-second Chinese fantasy, realistic Eastern cinematic look, gold-teal palette, ethereal ambient sound.
0-3s: High-angle aerial shot of an ancient temple amid clouds, slow aerial push, morning mist drifting through valleys, distant bell faintly ringing, Tyndall light beams piercing through cloud layers.
3-7s: Dolly push through the temple gate into the courtyard, a young man in white raises his hand to catch a falling red leaf, 35mm film grain texture, shallow depth of field focusing on hand details.
7-10s: Close-up of the young man looking up, slow push-in, wind rises, sleeves and hair sweep to the right side of frame, spiritual light spiraling upward in the courtyard.
Sound: Ambient sound contracts into a single clear sword ring.
Forbidden: Any text, subtitles, logos, or watermarks
# AI Comic Drama
Write a 10-second AI comic-style bossy CEO short, vertical 9:16, with dialogue and exaggerated close-up expressions
# E-commerce Ad
Write an 8-second luxury watch product ad video prompt, 9:16 vertical
# Short Drama Dialogue
Write a 12-second plot-twist short drama scene with dialogue
# One-Take
Write a 15-second one-take museum walkthrough video prompt
# Image-to-Video (I2V)
I have a character design image โ generate a 10-second animation from this first frame
# Video Extension
Extend this video by 10 seconds with a sunset ending
# Multiframe Storytelling
I have 5 storyboard keyframe images โ create a coherent story video that flows through all frames
# With Reference Assets
I've uploaded 3 character design images and 1 reference video โ generate a 15-second Xianxia fight scene
seedance-shot-design/
โโโ SKILL.md # Core instructions (the Skill's brain)
โโโ README.md # This file
โโโ scripts/
โ โโโ validate_prompt.py # Industrial-grade prompt validation script
โ โโโ test_validate.py # Validation script test cases
โโโ references/
โโโ cinematography.md # Camera & focal length dictionary (incl. physical mounts & focal psychology)
โโโ director-styles.md # Director style parameterized mapping (28+ styles, incl. Cel-Shaded CG)
โโโ seedance-specs.md # Seedance 2.0 official platform specs
โโโ quality-anchors.md # Quality anchors & lighting library (incl. NPR materials / lighting / conflict matrix)
โโโ scenarios.md # Vertical scenario templates (20 scenarios + anime variants + video editing + multiframe storytelling + physics damping toolkit)
โโโ audio-tags.md # Audio & sound effect tag specs (incl. spatial acoustics & material-based onomatopoeia)
A standalone Python validation tool for developers and CI/CD pipelines. The AI agent applies these same rules natively via its built-in 7-rule validation checklist โ no Python execution required during prompt generation.
# Validate text directly
python scripts/validate_prompt.py --text "your prompt"
# Validate from file
python scripts/validate_prompt.py --file prompt.txt
# Specify language (auto=auto-detect, cn=Chinese, en=English)
python scripts/validate_prompt.py --text "your prompt" --lang en
# JSON output (for programmatic processing)
python scripts/validate_prompt.py --text "your prompt" --jsonValidation Checks:
- โ Over word limit (Chinese >500 chars / English >1000 words)
- โ Missing professional camera terminology
- โ Filler word hard-block (masterpiece / ultra-clear, etc. โ error)
- โ Optical physics conflicts (ultra-wide + bokeh, handheld + perfect symmetry)
- โ Style conflict matrix (IMAX vs VHS, film vs digital, ink-wash vs UE5, Cel-Shaded vs realistic PBR, Slow Motion vs Speed Ramp)
- โ Asset reference overflow (images >9 / videos >3 / audio >3 / total >12)
- โ Long video (>5s) without time-slice hard-block
โ ๏ธ Time-slice gaps or overlapsโ ๏ธ Declared duration vs slice endpoint mismatchโ ๏ธ In-segment motion logic conflictsโ ๏ธ Seedance review-risk bare English camera terms detection (Dolly / Aerial / Crane / Pan / Arc / Dutch / Steadicam)- ๐ Auto language detection (Chinese / English), adapting length standards & detection strategies per language
- ๐ฌ Multi-segment cross-segment consistency checks (style preamble / lighting structure / forbidden items)
Run Tests:
python -m unittest scripts.test_validate -v
# 54 tests pass (covering 11 test classes)Following Agent Skills best practices:
- SKILL.md (~4000 tokens): Core workflow + structural templates + quality checklist
- references/ (three-layer routing): Camera dictionary and quality anchors are always loaded (Always-On); other knowledge bases auto-matched via semantic inference or loaded on explicit user request
- scripts/ (executed on demand): Validation runs only after prompt generation
| Dimension | Common Approach | This Skill |
|---|---|---|
| Compliance | Plain-text suggestions | Python hard validation (incl. optical / style conflict matrix + review safety detection) |
| Director Styles | International directors only | International + Chinese + Short Drama + AI Comic + Social Media + Anime + Cel-Shaded CG + Xiaohongshu |
| Scene Coverage | Biased toward epic films | 20 vertical scenarios + anime variants + video editing + physics damping toolkit |
| Sound Design | Brief mentions | Spatial acoustics + material-based onomatopoeia library |
| Lighting | "Add a light" | Source โ Behavior โ Tone three-layer + lighting recipes + material library |
| Multilingual | Chinese only | Chinese / English bilingual output, auto language detection |
| Review Safety | Not considered | Camera term disambiguation rules + bare-word auto-detection |
-
๐ Runway Platform Adaptation: Clarified native asset limits for Runway users (โค5 images, โค3 videos) and provided explicit moderation bypass strategies for realistic human faces (blurring or NPR restyling).
-
๐๏ธ Start & End Frame Interpolation: Added a 7th multimodal reference pattern (
@Image1 as start frame, @Image2 as end frame) to support precise point-to-point transitions. -
๐ฌ Two New Effects Scenarios:
Freeze Time (Bullet Time): Camera dramatically weaves through completely frozen scene elements.Multishot Video: Bypasses the "one-take" limitation by triggering the model to auto-generate sharp montage cuts in a single generation.
-
๐ Extreme POV Scenarios: Added new 21st scenario template focusing on "human head-tracking logic", "high-speed projectile FPV (flying swords/arrows)", and "creature flight".
-
๐ง Immersive Audio Exclusion: Introduced strict audio-exclusion directives for POV templates (ONLY ambient sound, NO BGM or dialogue) to prevent AI from breaking immersion.
-
๐งน Background Purification Rule: Clarified that reference entity images must use a "pure white/blank background" to avoid contaminating the video environment in Image-to-Video generation.
- ๐ CLI Integration Guide: New
seedance-specs.mdsection mapping Shot Design modes to Jimeng CLI commands (text2video/image2video/multiframe2video/multimodal2video), with async task management and VIP channel documentation - ๐๏ธ Multiframe Storytelling Template: New scenario template (#20) for
multiframe2videoโ upload 2-9 keyframe images and let the engine auto-compose a coherent story video. Includes decision matrix for choosing between multiframe vs. multi-segment storyboard - ๐ Knowledge Base Routing: Added multiframe and CLI routing entries to Step 2 semantic inference table
- ๐ญ Descriptive Over Narrative Rule: New core rule (#12) โ only write what the camera SEES (visual words), never what characters FEEL (emotion words). All emotions must be converted to visible physical expressions (facial micro-expressions, body language, breathing rhythm, gaze direction)
- โ๏ธ English Present Progressive: Assembly rules now mandate
-ingform for English action descriptions (runningnotruns) โ progressive tense implies continuous motion, matching video's dynamic nature - ๐ฏ Motion Tone Front-Loading: Style preamble now explicitly declares overall motion energy (e.g.,
dynamic motion, high energyorserene, slow-paced atmosphere) to lock motion baseline early in generation
- ๐ฅ One-Shot-One-Move Rule: New core rule (#10) enforcing a single camera movement per time segment โ combining movements (e.g., push-in + pan) causes jitter. Subject motion and camera motion must be described separately
- ๐ผ๏ธ I2V Golden Rule: New core rule (#11) and dedicated I2V section โ when generating video from an image, only describe motion/changes, never re-describe static content already in the first frame. Introduces
preserve composition and colorsanchor phrase - ๐ Optimal Prompt Length: Added 60โ100 word sweet spot guidance โ below is vague, above causes concept drift and conflicting instructions
- ๐ช Motion Intensity Modifiers: New bilingual quick-reference table in cinematography dictionary with 6 intensity tiers (violent โ gentle โ gradual) and do/don't examples to eliminate "mushy motion"
- ๐ค Rhythm Over Specs: Assembly rules now explicitly prefer semantic rhythm words (gentle/gradual/smooth) over technical parameters (24fps/f2.8) that Seedance cannot parse
- ๐ฌ Reference Video Best Practices: New practical constraints for reference clips โ ideal 3โ8s length, continuous shot (no cuts), single intent (subject OR camera, not both)
- ๐ก๏ธ Security Compliance: Resolved ClawHub OpenClaw "Suspicious patterns" flag by converting Python-based validation to LLM-native structured 7-rule validation checklist. Python scripts remain as standalone developer tools but are no longer executed by the agent
- ๐ฏ Trigger Phrase Optimization: Reduced activation trigger phrases from 40+ to 15 high-signal professional terms, lowering unintended activation surface while preserving core discoverability
- ๐ค Voice & Language Control System: New timbre cloning via video reference (
่ฏญๆฐๅ้ณ่ฒๅ่@่ง้ข1), dialect/accent control (Sichuan/Cantonese/Northeast/Taiwanese etc.), multilingual dialogue mixing, special voice styles (documentary narration / stand-up comedy / opera / ASMR) - ๐น Multimodal Reference Guide: Upgraded from 4 brief tips to a structured guide with 6 core reference patterns (first frame / camera replication / action replication / camera+action separation / timbre reference / effect replication), plus multi-asset character control and consistency preservation guidance
- ๐ Video Extension Scenario: New forward/backward extension templates, seamless continuation techniques, duration cognition correction (generation duration = added seconds, not total)
- ๐ Story Completion Scenario: New storyboard-to-video, comic panel animation, and image-to-emotion-video creative modes
- ๐ฌ Creative Effects Quick Reference: New VFX trigger keywords โ Hitchcock zoom, fisheye lens, particle effects, speed ramp, freeze transition, ink wash style, morphing effects (bilingual)
- ๐ญ Emotion Performance Guidance: New emotion specificity table, emotion transition trigger words, emotion reference video usage in short drama chapter
- ๐ฏ Trigger Word Expansion: Massively expanded Skill activation coverage โ added 20+ colloquial Chinese triggers (ๅธฎๆๅไธช่ง้ข, ๆไธไธช, ๅๅ้, ็ญ่ง้ข, AI่ง้ข, ๆ้ณ่ง้ข, vlog, ่ฟ้...) and 10+ English triggers (generate a video, make a clip, shoot a scene, video script, drone shot, camera movement...) so the Skill auto-activates on natural, everyday user expressions โ not just professional terminology
- ๐ Security Compliance Optimization: Resolved ClawHub security flagging issues for shell execution, process control, and file access patterns while maintaining full functionality
- ๐จ Step 3 Mandatory Assembly Rules: Three-layer lighting must be on its own line with all three layers complete; SFX line must start with
SFX:; prohibition line standardized (no custom additions); freestyle non-template sections forbidden - โ Step 4 Validation Blocking: Prompts failing validation are now forbidden from being shown to users; clear 5-step validation flow
- ๐ Step 5 Format Enforcement: Output must follow template exactly (Theme + Director's Note + code-block-wrapped prompt); missing any section = non-compliant
- ๐ฏ Step 2 Parameter Extraction Directive: Knowledge bases must not just be "loaded" โ specific parameters must be extracted and embedded into the prompt
- ๐ง Smart Semantic Intent Routing: Step 2 knowledge base loading upgraded from "explicit trigger" to three-layer routing โ Always-On loads camera dictionary & quality anchors every time, Semantic Intent Inference auto-detects needed knowledge bases from user's natural language, Explicit Override preserves direct user specification
- ๐ฏ Step 1 Smart Inference Principle: Agent proactively infers parameters (duration / style / scene) from a single user sentence, only asking about genuinely unknown info, limiting follow-up questions to 1-2
- ๐ Design philosophy updated from "loaded on demand" to "three-layer routing" ensuring every prompt has a quality foundation
- ๐ญ Actor Blocking System: Three-element positioning (placement + face direction + gaze focus) with emotion modifier vocabulary for multi-character scenes
- ๐๏ธ Voiceover / Dialogue Split: Distinct templates for on-screen dialogue vs. off-screen voiceover / inner monologue, with anti-lip-sync directive for VO scenes
- ๐ Camera Angle Specificity: Vague โ specific angle mapping (e.g., "close-up" โ "over-shoulder medium close-up, focus on listener") with 5 comparison pairs
- ๐ฌ Narrative-Motivated Camera Movement: Camera moves now paired with storytelling purpose (e.g., "slow push-in โ revealing inner turmoil")
- ๐ Segment Transition Strategy: 6 transition types (gaze continuity / emotional escalation / contrast cut / spatial leap / temporal ellipsis / sensory bridge) for multi-shot coherence
- ๐จ Short Drama Style Quick-Selector: 4-dimension combo system (visual type ร render style ร color tone ร genre)
- ๐ Short drama prompt templates expanded from 1 to 4 variants (CN dialogue / CN voiceover / EN dialogue / EN voiceover)
- ๐ Multi-segment Director's Note template adds transition strategy declaration
- ๐ 5 complete short drama examples covering: plot-twist dialogue / voiceover monologue / action conflict / 2D anime / transition strategy
- โ 54 tests pass
- ๐ฌ Smart Multi-Segment Storyboard: Videos >15s auto-split into multiple independent prompts (each โค15s, min โฅ8s)
- ๐ Multi-segment coherence: unified style preamble / three-layer lighting / sound design / transition frames / forbidden items
- ๐ Step 5 adds multi-segment output format template (CN / EN)
- ๐ New 60-second desert Kali/Escrima 4-segment full example
- ๐ง Validation script adds
validate_multi_segment()cross-segment consistency check - โ 54 tests pass (incl. 4 new multi-segment validation tests)
- ๐ Bilingual Prompt Output: Chinese users โ Chinese, non-Chinese โ English, with auto language detection
- ๐ All structural templates, delivery formats, and multimodal tips now include English versions
- ๐ก๏ธ Camera Term Disambiguation (Rule 9): Chinese uses Chinese camera terms, English uses full phrases โ avoids Seedance review false positives
- ๐ง Validation adds
check_ambiguous_terms()bare-word detection +--langflag + English word-count length check - ๐ง New Slow Motion vs Speed Ramp conflict detection
- ๐ง
detect_language()expanded with CJK Extension A + full-width punctuation support - ๐
cinematography.mdadds "Seedance Safe Phrasing" column - โ 50 tests pass (incl. bilingual + review safety tests)
- ๐จ Cel-Shaded CG Style: New complete four-axis parameterized entry (distinct from anime's explosive energy โ positioned for contemplative narrative)
- ๐งฑ Anime / NPR Material Library: Anime skin / hair / cartoon metal / cartoon fabric โ 4 non-photorealistic materials
- ๐ฆ Anime Game PV Variant: Scenario template adds Cel-Shaded sub-template + ice-attribute character example
โ ๏ธ Conflict matrix adds: Cel-Shade vs Realistic PBR material- ๐ง Validation adds Cel-Shade vs PBR style conflict detection
- ๐ฌ Camera Upgrade: New focal length narrative psychology, dynamic focus paradigms, physical mount chapter (7 specialty rigs)
- ๐จ Director Styles: New Fincher / Deakins / Kurosawa / Makoto Shinkai + Anime Explosion / Xiaohongshu Aesthetic (incl. de-named safe prompts + forbidden items)
- ๐ก Quality Upgrade: Anti-plastic manifesto, film stock library (5 types), material texture library (8 types), lighting combo quick-reference (4 sets), organic imperfection library, quality conflict matrix
- ๐ฌ Scene Expansion: New Game PV / Horror-Thriller / Travel-City / Pet-Cute / Before-After / Meme-Loop, totaling 16 scenarios + physics damping appendix
- ๐๏ธ Sound Upgrade: Spatial acoustic modifiers (7 types), material-based onomatopoeia refinement (7 pairs)
- ๐ง Validation Enhancement: Filler word warning โ error hard-block, optical physics conflict detection, style conflict matrix, duration-aware time-slicing, 35 tests pass
- ๐ Initial release
- SKILL.md core workflow
- 6 professional knowledge base files
- Python validation script + test cases
- 20+ director style mappings
- 10 vertical scenario templates
MIT-0 (MIT No Attribution) License
