Generate agent-ready prompts from existing task specification files. Use when regenerating prompts after editing tasks, updating prompt templates, or preparing tasks for cpo execution.
Install with Tessl CLI
npx tessl i github:jpoutrin/product-forge --skill parallel-prompt-generator80
Does it follow best practices?
If you maintain this skill, you can automatically optimize it using the tessl CLI to improve its score:
npx tessl skill review --optimize ./path/to/skillValidation for skill structure
Generate agent-ready prompts from existing task specification files in tasks/.
Every generated prompt file MUST include ALL of these sections (see Step 4 for full template):
| Section | Purpose | Required |
|---|---|---|
=== REQUIRED SKILLS === | Skills to invoke at start | YES |
=== CONTEXT === | Shared project context | YES |
=== OBJECTIVE === | Task goal | YES |
=== CONTRACTS === | Contract file references | YES |
=== FILES TO CREATE === | Scope CREATE | YES |
=== FILES TO MODIFY === | Scope MODIFY | YES |
=== DO NOT MODIFY === | Scope BOUNDARY | YES |
=== IMPLEMENTATION REQUIREMENTS === | Requirements | YES |
=== ACCEPTANCE CRITERIA === | Checklist | YES |
=== EXECUTION INSTRUCTIONS === | How to implement | MANDATORY |
=== IMPORTANT RULES === | Constraints | MANDATORY |
=== COMPLETION SIGNAL === | touch .claude-task-complete | MANDATORY |
DO NOT generate prompts without the EXECUTION INSTRUCTIONS, IMPORTANT RULES, and COMPLETION SIGNAL sections.
tasks/cpo run if prompts are missingThe parallel directory must have:
parallel/TS-XXXX-slug/
manifest.json # Required: wave structure, agents
context.md # Required: shared context
contracts/ # Required: types.py, api-schema.yaml
tasks/ # Required: task-*.md files
task-001-users.md
task-002-products.mdGenerates:
parallel/TS-XXXX-slug/
prompts/
agent-prompts.md # Wave summary and launch commands
task-001.txt # Individual agent prompt
task-002.txt
...flowchart TB
Start[("tasks/ folder")]
Read["1. Read manifest.json"]
Context["2. Load context.md"]
Parse["3. Parse each task file"]
Generate["4. Generate task-NNN.txt"]
Summary["5. Create agent-prompts.md"]
Done[("prompts/ ready")]
Start --> Read
Read --> Context
Context --> Parse
Parse --> Generate
Generate --> Summary
Summary --> DoneExtract:
The shared context is injected into every prompt for efficiency.
For each tasks/task-NNN-*.md:
Extract YAML frontmatter:
id: task-001
component: users
wave: 1
deps: []
agent: python-experts:django-expert
contracts: [contracts/types.py, contracts/api-schema.yaml]Extract Markdown sections:
## Scope -> CREATE, MODIFY, BOUNDARY## Requirements -> Implementation requirements## Checklist -> Acceptance criteriaResolve skills for each task:
parallel-agents/agent-skills-mapping.yaml for agent-to-skills mappingpython-experts:django-expert)plugin:skill-nameNote: The Output Format section is not in task files. It's a static template that the prompt generator always includes in generated prompts.
CRITICAL: Copy content EXACTLY from task files. Do not summarize or rewrite.
When generating prompts:
principal_id not user_id)Create prompts/task-NNN.txt using this template:
TASK-{id}: {title}
Agent: {agent}
Wave: {wave}
Dependencies: {deps or "None"}
=== REQUIRED SKILLS ===
Before starting implementation, invoke these skills to load best practices:
{list of skills from agent-skills-mapping.yaml, formatted as bullets}
Invoke skills using: skill: "plugin:skill-name"
=== CONTEXT ===
{contents of context.md}
=== OBJECTIVE ===
{First line of task description or component name}
=== CONTRACTS ===
Reference these contract files before implementing:
{list of contract files with full paths}
=== FILES TO CREATE ===
{CREATE section from Scope}
=== FILES TO MODIFY ===
{MODIFY section from Scope}
=== DO NOT MODIFY ===
{BOUNDARY section from Scope}
=== IMPLEMENTATION REQUIREMENTS ===
{Requirements section content}
=== ACCEPTANCE CRITERIA ===
{Checklist section content}
=== EXECUTION INSTRUCTIONS ===
You MUST write code, not just describe it. Follow these steps:
1. **Read contracts first**: Read all files in contracts/ to understand interfaces
2. **Read existing code**: If modifying files, read them first
3. **Implement**: Use Write/Edit tools to create/modify code
4. **Test & Fix Loop**:
- Run tests for your implementation
- If ANY test fails, FIX the issue and rerun tests
- REPEAT until ALL tests pass
- Do NOT proceed to step 5 until tests pass
5. **Lint**: Run linters (ruff, mypy, eslint as appropriate)
6. **Commit**: Create atomic commit with conventional format (only after tests pass)
=== IMPORTANT RULES ===
- Do NOT ask for confirmation - proceed immediately with implementation
- Do NOT modify files listed in DO NOT MODIFY section
- Do NOT skip writing tests
- Do NOT deviate from contract interfaces
- Do NOT commit or signal completion until ALL tests pass
- If tests fail, you MUST fix the issues and keep running until they pass
- STOP if you encounter blocking issues and report in output
**CRITICAL - CONTRACTS ARE DESIGN DOCUMENTS:**
- Contracts in `parallel/` are DESIGN SPECIFICATIONS ONLY
- READ them to understand the interface you must implement
- RECREATE the types/interfaces in your actual project code
- NEVER import from `parallel/` directory
- NEVER use sys.path manipulation to access parallel/ files
- The `parallel/` directory is NOT part of the deployed codebase
=== COMPLETION SIGNAL ===
Upon successful completion (ALL tests passing), run: touch .claude-task-complete
CRITICAL: Only signal completion if:
- All tests pass (pytest exits with 0)
- All linting passes
- Code is committedSummary file with launch commands:
# Agent Prompts for {name}
Generated: {timestamp}
Tech Spec: {tech_spec_id}
Total Tasks: {count}
## Wave Execution Order
### Wave 1 (Parallel)
| Task | Agent | Component |
|------|-------|-----------|
| task-001 | python-experts:django-expert | users |
| task-002 | python-experts:django-expert | products |
### Wave 2 (Parallel, depends on Wave 1)
| Task | Agent | Component |
|------|-------|-----------|
| task-003 | python-experts:django-expert | orders |
## Launch Commands
### Using cpo (Recommended)
```bash
cpo run parallel/{slug}/# Run in parallel terminals or with &
claude --prompt-file prompts/task-001.txt --agent python-experts:django-expert &
claude --prompt-file prompts/task-002.txt --agent python-experts:django-expert &
wait# After Wave 1 completes
claude --prompt-file prompts/task-003.txt --agent python-experts:django-expert| File | Task | Lines |
|---|---|---|
| prompts/task-001.txt | task-001-users | ~150 |
| prompts/task-002.txt | task-002-products | ~145 |
| prompts/task-003.txt | task-003-orders | ~160 |
## Validation Before Generation
Check these conditions:
- [ ] `manifest.json` exists and is valid JSON
- [ ] `context.md` exists
- [ ] `contracts/` directory exists with at least one file
- [ ] `tasks/` directory has at least one task file
- [ ] All tasks referenced in manifest exist in `tasks/`
## Error Handling
### Task Not in ManifestWARNING: task-005-api.md exists but not in manifest.json -> Skipping prompt generation (add to manifest first)
### Missing Contract ReferenceWARNING: task-001 references contracts/types.py but file doesn't exist -> Generate prompt with warning comment
## Regeneration Mode
When prompts already exist:
1. **Backup existing**: Move to `prompts.backup-{timestamp}/`
2. **Generate fresh**: Create new prompts from current tasks
3. **Report diff**: Show what changed
```bash
# Command triggers regeneration
/parallel-generate-prompts parallel/TS-0042-inventory/ --forceThe generated prompts are optimized for token efficiency:
The cpo run command expects:
manifest.json with wave structureprompts/task-NNN.txt for each task in manifestThis skill ensures prompts match what cpo expects.
After generating prompts, verify EACH prompt file contains:
# Quick validation - each prompt should have these markers
for f in prompts/task-*.txt; do
echo "=== $f ==="
grep -c "=== REQUIRED SKILLS ===" "$f" || echo "MISSING: REQUIRED SKILLS"
grep -c "=== EXECUTION INSTRUCTIONS" "$f" || echo "MISSING: EXECUTION INSTRUCTIONS"
grep -c "=== COMPLETION SIGNAL" "$f" || echo "MISSING: COMPLETION SIGNAL"
doneIf any prompt is missing these sections, regenerate it using the exact template from Step 4.
Note: Output format and JSON summary blocks are managed via system prompt by the external execution tool, not included in generated prompts.
| Skill | Relationship |
|---|---|
parallel-decompose | Creates tasks/ that this skill reads |
parallel-task-format | Defines the task YAML format parsed here |
parallel-execution | Consumes the prompts/ this skill generates |
parallel-agents | Overall workflow context + agent-skills-mapping.yaml |
parallel-agents/agent-skills-mapping.yaml - Maps agents to their recommended skills0ebe7ae
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.