Authentic UGC-style videos with real face references, native voice and ambient audio via Seedance 2.0 Omni-Reference (MuAPI). Drop in a persona headshot and get clips with consistent face, natural speech and room tone. For marketers and DTC brands who want real character consistency, not dubbed TTS.
Step 1: Research audience pain points, language, triggers, and objections. Async (1-3 min). Do NOT skip. Pass full output to generate_creative.
Generate UGC video hooks using the 3-variable framework (Angle + Aesthetic + Action). Uses RAG from the UGC playbook. ⏱ ~5 seconds. Requires research output from step 1.
Generate authentic UGC scripts using But/Therefore zigzag structure. ⏱ 5-10s per hook. Requires hooks from write_hooks.
Generate visual scene descriptions for each script, optimized for AI image/video generation. Set "format" for format-specific direction. Takes ~5s per script.
Generate hooks, scripts, and scene descriptions in one call (~8s). Set "format" for format-specific output. Photo formats return compositions instead of hooks/scripts. Pass FULL research output.
Generate first-frame images from scenes. ~30s async. Pass existing_frames to skip frames that already have image_url. persona_image_url for face consistency.
Generate UGC video clips via MuAPI Seedance 2.0 Omni-Reference. Accepts real face photos as persona_image_url (no fal.ai face block). Takes 2-10 MINUTES. Seedance generates speech and ambient natively — voice is steered by persona.tone + persona.accent. Use check_video for pending clips.
Check on a pending MuAPI video that was still generating when generate_videos returned. Pass the muapi_request_id. Returns the video URL if ready, or current status.
Check on a pending image that was still generating when generate_frames timed out. Pass the fal_request_id from pending_frames.
Re-generate a single frame without re-running the full pipeline. Accepts optional revision notes for targeted edits.
Composite real product into frames. Inline chat images CANNOT be passed as URLs. Direct user to toolrouter.com/dashboard/files to upload their image.
Stitch video clips into one final video with text overlays and transitions. Last step after generate_videos. Takes ~30-120s. Async.
List available models for this tool, sorted by popularity. Returns provider details and pricing.
Loading activity...
- Product image now composited into frames and passed as @Image3 reference to Seedance so the real product appears in every clip
- Tightened iPhone front-camera sensor aesthetic — grain, highlight clipping, compressed dynamic range
- Renamed to UGC Video Turbo — accept scenes, products, and outfits from your file library
- Initial release — Seedance 2.0 Omni-Reference via MuAPI, face-reference persona images, ElevenLabs TTS lip-sync