Build Retrieval-Augmented Generation (RAG) systems for LLM applications with vector databases and semantic search. Use when implementing knowledge-grounded AI, building document Q&A systems, or integrating LLMs with external knowledge bases.
71
66%
Does it follow best practices?
Impact
70%
2.12xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./tests/ext_conformance/artifacts/agents-wshobson/llm-application-dev/skills/rag-implementation/SKILL.mdLangGraph RAG pipeline with Anthropic stack and citation prompting
LangGraph StateGraph
0%
100%
Claude model version
0%
0%
Voyage embedding model
0%
100%
RAGState schema
0%
100%
Retriever k=4
62%
100%
Local vector store
87%
100%
RecursiveCharacterTextSplitter separators
50%
100%
Chunk size range
100%
100%
Chunk overlap ratio
100%
100%
Citation format in prompt
37%
100%
Fallback phrase
100%
100%
Graph edge order
0%
100%
Hybrid search with cross-encoder reranking and metadata filtering
EnsembleRetriever usage
0%
70%
Hybrid search weights
0%
0%
Component k values
0%
80%
Cross-encoder model
100%
100%
Reranking candidate pool
0%
0%
Top-k after reranking
0%
0%
Metadata fields
100%
70%
Voyage embeddings
0%
0%
Claude model
0%
50%
LangGraph or async pattern
0%
0%
HyDE retrieval, parent document retriever, structured output, and RAG evaluation
HyDE hypothetical generation
100%
100%
HyDE retrieval uses hypothetical doc
100%
100%
HyDE state fields
0%
100%
HyDE graph ordering
0%
100%
Parent doc child chunk size
0%
0%
Parent doc parent chunk size
0%
0%
RAGResponse Pydantic model
30%
100%
Structured output binding
0%
100%
Faithfulness metric
0%
100%
Retrieval quality metrics
100%
100%
Answer quality metrics
25%
100%
47823e3
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.