Master advanced prompt engineering techniques to maximize LLM performance, reliability, and controllability in production. Use when optimizing prompts, improving LLM outputs, or designing production prompt templates.
67
54%
Does it follow best practices?
Impact
83%
1.69xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./plugins/llm-application-dev/skills/prompt-engineering-patterns/SKILL.mdStructured output pipeline with LangChain
ChatAnthropic import
0%
100%
Correct model name
0%
100%
with_structured_output usage
100%
100%
ChatPromptTemplate.from_messages
100%
100%
Pydantic Field descriptions
100%
70%
Async ainvoke
100%
100%
ValidationError catch
0%
0%
JSONDecodeError catch
0%
0%
Fallback on error
0%
50%
System message included
100%
100%
Chain composition operator
100%
100%
Chain-of-thought reasoning with self-consistency
Zero-shot CoT trigger
0%
100%
Steps section label
44%
100%
Answer section label
77%
100%
Verification section label
44%
100%
Verification triggers revision
60%
70%
Self-consistency sample count
100%
100%
Self-consistency temperature
0%
0%
Majority vote selection
100%
100%
Few-shot CoT step format
0%
0%
Confidence reported
100%
100%
Separate verification call
100%
100%
Dynamic few-shot selection with prompt caching
VoyageAI embeddings import
0%
100%
voyage-3-large model
0%
100%
Chroma vectorstore
0%
100%
k=2 examples selected
100%
100%
User/Assistant example format
0%
100%
Prompt caching cache_control
100%
100%
Direct Anthropic SDK
100%
100%
claude-sonnet-4-6 model
0%
100%
RAG context-only instruction
100%
100%
RAG fallback phrase
50%
100%
Citation notation
0%
66%
Hard vs soft constraints
0%
50%
91fe43e
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.