CtrlK
BlogDocsLog inGet started
Tessl Logo

prompt-engineer

Writes, refactors, and evaluates prompts for LLMs — generating optimized prompt templates, structured output schemas, evaluation rubrics, and test suites. Use when designing prompts for new LLM applications, refactoring existing prompts for better accuracy or token efficiency, implementing chain-of-thought or few-shot learning, creating system prompts with personas and guardrails, building JSON/function-calling schemas, or developing prompt evaluation frameworks to measure and improve model performance.

100

Quality

100%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

SKILL.md
Quality
Evals
Security

Quality

Discovery

100%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is an excellent skill description that excels across all dimensions. It uses proper third-person voice, provides comprehensive specific capabilities, includes rich natural trigger terms that practitioners would use, and has an explicit 'Use when...' clause with multiple concrete scenarios. The description is distinctive enough to avoid conflicts with other skills while being comprehensive about its scope.

DimensionReasoningScore

Specificity

Lists multiple specific concrete actions: 'Writes, refactors, and evaluates prompts', 'generating optimized prompt templates, structured output schemas, evaluation rubrics, and test suites'. These are clear, actionable capabilities.

3 / 3

Completeness

Clearly answers both what ('Writes, refactors, and evaluates prompts for LLMs') and when with explicit 'Use when...' clause covering six distinct trigger scenarios including designing new prompts, refactoring, implementing specific techniques, and building evaluation frameworks.

3 / 3

Trigger Term Quality

Excellent coverage of natural terms users would say: 'prompts', 'LLM', 'chain-of-thought', 'few-shot learning', 'system prompts', 'JSON', 'function-calling', 'token efficiency', 'evaluation frameworks'. These match how practitioners naturally discuss prompt engineering.

3 / 3

Distinctiveness Conflict Risk

Highly distinctive niche focused specifically on prompt engineering for LLMs. The specific triggers like 'chain-of-thought', 'few-shot learning', 'system prompts', and 'function-calling schemas' clearly differentiate this from general coding or documentation skills.

3 / 3

Total

12

/

12

Passed

Implementation

100%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is an exemplary skill file that demonstrates best practices across all dimensions. It provides actionable guidance with concrete examples, maintains token efficiency by avoiding unnecessary explanations, includes clear validation checkpoints in the workflow, and uses progressive disclosure effectively through a well-organized reference table. The MUST DO/MUST NOT DO constraints are specific and practical.

DimensionReasoningScore

Conciseness

The content is lean and efficient, avoiding explanations of concepts Claude already knows. Each section serves a clear purpose without padding or unnecessary context about what prompts or LLMs are.

3 / 3

Actionability

Provides concrete, copy-paste ready prompt examples with clear before/after comparisons. The zero-shot vs few-shot examples and the optimization example are fully executable templates with placeholder syntax.

3 / 3

Workflow Clarity

The 5-step core workflow is clearly sequenced with an explicit validation checkpoint at step 3 (accuracy < 80% triggers failure pattern analysis). The iterate step explicitly states 'one change at a time' for debugging discipline.

3 / 3

Progressive Disclosure

Excellent structure with a clear reference table pointing to 6 separate files for detailed guidance, each with a 'Load When' context. The main skill provides a complete overview while appropriately deferring deep dives to one-level-deep references.

3 / 3

Total

12

/

12

Passed

Validation

100%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation11 / 11 Passed

Validation for skill structure

No warnings or errors.

Repository
jeffallan/claude-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.