Build Retrieval-Augmented Generation (RAG) systems for LLM applications with vector databases and semantic search. Use when implementing knowledge-grounded AI, building document Q&A systems, or integrating LLMs with external knowledge bases.
67
56%
Does it follow best practices?
Impact
83%
2.07xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./plugins/llm-application-dev/skills/rag-implementation/SKILL.mdLangGraph RAG pipeline with preferred libraries
LangGraph import
100%
100%
ChatAnthropic client
0%
100%
Correct Claude model
0%
100%
VoyageAI embeddings import
0%
100%
voyage-3-large model
0%
100%
RAGState TypedDict
66%
100%
Retriever k=4
0%
100%
Async invocation
0%
100%
Cannot-answer fallback
100%
100%
Hybrid search with chunking and structured output
EnsembleRetriever
80%
70%
BM25 weight 0.3
0%
100%
Dense weight 0.7
0%
100%
RecursiveCharacterTextSplitter
100%
100%
Chunk size 1000
0%
0%
Chunk overlap 200
0%
0%
Chroma vector store
0%
100%
Chroma persist_directory
0%
0%
Citation format in prompt
0%
0%
Structured output model
30%
100%
Structured model fields
100%
100%
Production Pinecone setup and reranking configuration
Pinecone dimension=1024
0%
100%
Pinecone metric=cosine
100%
100%
ServerlessSpec aws us-east-1
100%
100%
CrossEncoder model name
100%
100%
MMR fetch_k=20
50%
100%
MMR lambda_mult=0.5
0%
0%
Metadata source field
100%
100%
Metadata category field
0%
100%
Metadata date field
100%
50%
Faithfulness metric
100%
100%
Five eval metrics
37%
100%
91fe43e
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.