Use when the user wants to narrate a demo, add voice-over to a screen recording, or create AI narration for a silent video. End-to-end pipeline that extracts frames, analyzes with parallel subagents, writes a word-budgeted voice-over script, generates TTS audio per act, and merges everything back.
94
94%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
extract <video> [fps] -- extract frames and contact sheetstts-acts <dir> <timing.txt> [voice] -- generate per-act audiotts-acts --dry-run <dir> <timing.txt> -- show budgets without TTSmerge-acts <video> <dir> <timing.txt> [out] -- merge audio onto videofade-intro <video> [secs] [out] -- add fade-in from blackdeps -- check dependenciesvoices -- list available TTS voicestts <file> [voice] -- single-file TTS (manual workflow, not used by pipeline)merge <video> <audio> [out] -- single-file merge (manual workflow, not used by pipeline)Default voices for Step 4a preferences:
For the full list of available voices, run "$NARRATE_SH" voices.
<output_dir>/
act1_opening.txt <- plain text, spoken words only
act1_opening.mp3 <- TTS audio for act 1
act2_extraction.txt
act2_extraction.mp3
...
timing.txt <- "filename offset_seconds" per line
script_final.txt <- master reference with all acts + metadataThis skill can also be used without the full plugin:
SKILL.md to ~/.claude/commands/demo-narrate.mdscripts/demo-narrate.sh to ~/.local/bin/demo-narrate.sh
and chmod +x it (must be on your PATH)command -v demo-narrate.sh -- no path edits needed