CtrlK
BlogDocsLog inGet started
Tessl Logo

prompt-engineering-patterns

Master advanced prompt engineering techniques to maximize LLM performance, reliability, and controllability in production. Use when optimizing prompts, improving LLM outputs, or designing production prompt templates.

67

1.69x
Quality

54%

Does it follow best practices?

Impact

83%

1.69x

Average score across 3 eval scenarios

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/llm-application-dev/skills/prompt-engineering-patterns/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

67%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description has good structural completeness with an explicit 'Use when' clause, which is its strongest aspect. However, it leans toward high-level buzzwords ('maximize LLM performance, reliability, and controllability') rather than listing concrete, specific techniques. The trigger terms cover the basics but miss many natural variations users might employ when seeking prompt engineering help.

Suggestions

Replace abstract phrases like 'maximize LLM performance, reliability, and controllability' with specific techniques such as 'chain-of-thought prompting, few-shot example design, system prompt structuring, output format control'.

Expand trigger terms in the 'Use when' clause to include natural user phrases like 'system prompt', 'few-shot examples', 'reduce hallucinations', 'prompt template', 'AI responses', or 'prompt debugging'.

DimensionReasoningScore

Specificity

Names the domain ('prompt engineering') and mentions some actions ('optimizing prompts', 'improving LLM outputs', 'designing production prompt templates'), but these are fairly high-level and not concrete specific actions like 'chain-of-thought structuring, few-shot example selection, system prompt design'.

2 / 3

Completeness

Clearly answers both 'what' ('Master advanced prompt engineering techniques to maximize LLM performance, reliability, and controllability in production') and 'when' ('Use when optimizing prompts, improving LLM outputs, or designing production prompt templates') with an explicit 'Use when' clause.

3 / 3

Trigger Term Quality

Includes some relevant keywords like 'prompt engineering', 'LLM', 'prompts', 'production', but misses many natural user terms like 'system prompt', 'few-shot', 'chain of thought', 'prompt template', 'AI output quality', 'hallucination', 'prompt optimization'.

2 / 3

Distinctiveness Conflict Risk

The domain of 'prompt engineering' is somewhat specific, but terms like 'improving LLM outputs' and 'maximize LLM performance' are broad enough to potentially overlap with skills related to general LLM usage, AI coding assistance, or model evaluation.

2 / 3

Total

9

/

12

Passed

Implementation

42%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

The skill provides excellent, executable code examples across many prompt engineering patterns, but is severely bloated with content Claude already knows (definitions of few-shot learning, CoT, best practices lists, success metrics). The monolithic structure with no external file references makes it a poor fit for a SKILL.md that should be a concise overview. The ratio of novel, actionable content to general knowledge padding is low.

Suggestions

Remove the 'Core Capabilities' overview section, 'Best Practices', 'Common Pitfalls', and 'Success Metrics' sections entirely — these describe concepts Claude already knows and add ~80 lines of padding.

Split the detailed code patterns into separate files (e.g., patterns/structured-output.md, patterns/few-shot.md) and keep SKILL.md as a concise index with one-line descriptions and links.

Add an explicit decision workflow: 'Start with simple prompt → measure → if inconsistent, add examples → if reasoning errors, add CoT → if parsing fails, add structured output' with validation checkpoints at each stage.

Remove the 'When to Use This Skill' section — it's a bullet list of obvious use cases that wastes tokens.

DimensionReasoningScore

Conciseness

The skill is extremely verbose at ~350+ lines, with extensive sections explaining concepts Claude already knows (what few-shot learning is, what chain-of-thought is, best practices like 'be specific', common pitfalls, success metrics). The 'When to Use This Skill' and 'Core Capabilities' sections are pure padding that describe rather than instruct. Much of this content is general prompt engineering knowledge that Claude inherently possesses.

1 / 3

Actionability

The skill provides numerous fully executable Python code examples with real libraries (langchain, anthropic, pydantic), concrete patterns with copy-paste ready implementations, and specific schemas. The code examples are complete and functional with proper imports and error handling.

3 / 3

Workflow Clarity

While individual patterns are clear, there's no overarching workflow for when to apply which pattern or how to iterate through prompt optimization. The 'Progressive Disclosure' pattern (Pattern 4) shows escalation levels but lacks explicit validation checkpoints or decision criteria for moving between levels. The iterative refinement process mentioned in Core Capabilities is never actually detailed.

2 / 3

Progressive Disclosure

This is a monolithic wall of content with no references to external files. All patterns, examples, integration guides, best practices, and metrics are inlined in a single massive document. Content like the role-based system prompts, RAG patterns, and detailed code examples should be split into separate referenced files.

1 / 3

Total

7

/

12

Passed

Validation

100%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation11 / 11 Passed

Validation for skill structure

No warnings or errors.

Repository
wshobson/agents
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.