Optimize your skills and tiles: review SKILL.md quality, generate eval scenarios, run evals, compare across models, diagnose gaps, and re-run until scores improve.
88
94%
Does it follow best practices?
Impact
88%
1.07xAverage score across 24 eval scenarios
Passed
No known issues
You handle tile eval setup — scenario generation from a tile, running evals, and presenting results.
The user triggers this skill when they have a tile but no eval scenarios yet, or when they want to generate new scenarios.
Companion skill: After setup is complete, suggest the user run the optimize-skill-performance skill to analyze results, diagnose failures, fix tile content, and re-verify improvements.
Time expectations: Set these upfront so the user isn't surprised:
Before diving in, figure out what the user wants to accomplish in this session. If the user's request already makes the scope clear (e.g., "run my evals", "generate scenarios"), skip the question and go straight to the relevant phase.
Otherwise, ask:
"What would you like to do?
- Full pipeline — generate scenarios, run evals, and see results (start-to-finish, ~1 hour)
- Generate scenarios only — generate and download scenarios, but don't run evals yet
- Run evals on existing scenarios — skip generation, just run and compare results on scenarios already in
evals/- Something else — tell me what you need"
Map the user's choice to phases:
| Choice | Phases to run |
|---|---|
| Full pipeline | 1 → 2 → 3 → 4 → 5 → 6 |
| Generate scenarios only | 1 → 2 → 3 |
| Run evals on existing scenarios | 1 → 4 → 5 → 6 |
For partial runs, skip phases not in scope — don't load their reference files.
Locate the tile and check for existing scenarios.
Read references/phase1-gather-context.md for the full procedure.
Run tessl scenario generate against the tile and review what was generated.
Read references/phase2-generate-scenarios.md for the full procedure.
Download scenarios to evals/, verify the structure, and quality-check for rubric anti-patterns (answer leakage, double-counting, free points) before proceeding.
Read references/phase3-download-scenarios.md for the full procedure.
Choose agents/models, run tessl eval run, and poll for completion.
Read references/phase4-run-evals.md for the full procedure.
Show baseline vs. with-context scores and per-scenario breakdown.
Read references/phase5-view-results.md for the full procedure.
Summarize the setup, suggest next actions based on scores, and offer to continue.
Read references/phase6-next-steps.md for the full procedure.
Stop when:
optimize-skill-performance or stopevals
scenario-1
scenario-2
scenario-3
scenario-4
scenario-5
scenario-6
scenario-7
scenario-8
scenario-9
scenario-10
scenario-11
scenario-12
scenario-13
scenario-14
scenario-15
scenario-16
scenario-17
scenario-18
scenario-19
scenario-20
scenario-21
scenario-22
scenario-23
scenario-24
skills
compare-skill-model-performance
optimize-skill-instructions
references
optimize-skill-performance
optimize-skill-performance-and-instructions