Optimize vector index performance for latency, recall, and memory. Use when tuning HNSW parameters, selecting quantization strategies, or scaling vector search infrastructure.
77
71%
Does it follow best practices?
Impact
84%
1.27xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./plugins/llm-application-dev/skills/vector-index-tuning/SKILL.mdHNSW parameter benchmarking
hnswlib used
100%
100%
Cosine space
0%
100%
Recall via set intersection
100%
100%
Memory formula
60%
100%
M recommendation for 500K
100%
0%
ef_construction for 500K
100%
0%
ef_search for 95% recall
0%
0%
Results artifact produced
100%
100%
Benchmark sweeps M and ef_construction
100%
100%
Benchmark sweeps ef_search
100%
100%
Recommendation provided
100%
100%
Search time in ms
100%
100%
Qdrant collection configuration
qdrant_client used
75%
100%
COSINE distance
100%
100%
Recall HNSW config
50%
100%
Recall: no quantization
100%
100%
Speed HNSW config
25%
0%
Speed: INT8 always_ram=True
77%
0%
Memory HNSW config
42%
100%
Memory: ProductQuantization X16
0%
100%
Balanced HNSW config
44%
100%
Optimizer thresholds
37%
100%
Search params: 0.99 threshold
25%
100%
Search params: 0.95 threshold
44%
100%
Vector quantization and memory estimation
FP32 bytes=4
100%
100%
FP16 bytes=2
100%
100%
INT8 bytes=1
100%
100%
PQ bytes~0.05
37%
100%
Binary bytes=0.125
100%
100%
HNSW overhead formula
70%
100%
IVF overhead formula
20%
100%
Index recommendation: <10K → Flat
100%
100%
Index recommendation: 10K-1M → HNSW
100%
100%
Index recommendation: >100M → IVF+PQ/DiskANN
40%
80%
Index recommendation: 1M-100M → HNSW+Quantization
0%
62%
memory_report.json produced
100%
100%
M parameter in HNSW overhead
100%
100%
91fe43e
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.