Writes, refactors, and evaluates prompts for LLMs — generating optimized prompt templates, structured output schemas, evaluation rubrics, and test suites. Use when designing prompts for new LLM applications, refactoring existing prompts for better accuracy or token efficiency, implementing chain-of-thought or few-shot learning, creating system prompts with personas and guardrails, building JSON/function-calling schemas, or developing prompt evaluation frameworks to measure and improve model performance.
100
100%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that excels across all dimensions. It uses proper third-person voice, provides comprehensive specific capabilities, includes rich natural trigger terms that practitioners would use, and has an explicit 'Use when...' clause with multiple concrete scenarios. The description is distinctive enough to avoid conflicts with other skills while being comprehensive about its scope.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'Writes, refactors, and evaluates prompts', 'generating optimized prompt templates, structured output schemas, evaluation rubrics, and test suites'. These are clear, actionable capabilities. | 3 / 3 |
Completeness | Clearly answers both what ('Writes, refactors, and evaluates prompts for LLMs') and when with explicit 'Use when...' clause covering six distinct trigger scenarios including designing new prompts, refactoring, implementing specific techniques, and building evaluation frameworks. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural terms users would say: 'prompts', 'LLM', 'chain-of-thought', 'few-shot learning', 'system prompts', 'JSON', 'function-calling', 'token efficiency', 'evaluation frameworks'. These match how practitioners naturally discuss prompt engineering. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive niche focused specifically on prompt engineering for LLMs. The specific triggers like 'chain-of-thought', 'few-shot learning', 'system prompts', and 'function-calling schemas' clearly differentiate this from general coding or documentation skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
100%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is an exemplary skill file that demonstrates best practices across all dimensions. It provides actionable guidance with concrete examples, maintains token efficiency by avoiding unnecessary explanations, includes clear validation checkpoints in the workflow, and uses progressive disclosure effectively through a well-organized reference table. The MUST DO/MUST NOT DO constraints are specific and practical.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The content is lean and efficient, avoiding explanations of concepts Claude already knows. Each section serves a clear purpose without padding or unnecessary context about what prompts or LLMs are. | 3 / 3 |
Actionability | Provides concrete, copy-paste ready prompt examples with clear before/after comparisons. The zero-shot vs few-shot examples and the optimization example are fully executable templates with placeholder syntax. | 3 / 3 |
Workflow Clarity | The 5-step core workflow is clearly sequenced with an explicit validation checkpoint at step 3 (accuracy < 80% triggers failure pattern analysis). The iterate step explicitly states 'one change at a time' for debugging discipline. | 3 / 3 |
Progressive Disclosure | Excellent structure with a clear reference table pointing to 6 separate files for detailed guidance, each with a 'Load When' context. The main skill provides a complete overview while appropriately deferring deep dives to one-level-deep references. | 3 / 3 |
Total | 12 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
5b76101
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.