This skill should be used when the user asks to "optimize context", "reduce token costs", "improve context efficiency", "implement KV-cache optimization", "partition context", or mentions context limits, observation masking, context budgeting, or extending effective context capacity.
49
36%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./skills/context-optimization/SKILL.mdQuality
Discovery
37%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This description is heavily lopsided: it provides excellent trigger term coverage for when the skill should be activated, but completely fails to describe what the skill actually does. A reader or Claude selecting from multiple skills would know when to pick it but not what it delivers. The description needs to be restructured to lead with concrete capabilities before listing trigger conditions.
Suggestions
Add a clear 'what it does' section listing specific concrete actions, e.g., 'Partitions prompt context into static and dynamic sections, implements KV-cache-friendly layouts, masks redundant observations, and budgets token allocation across context windows.'
Restructure to lead with capabilities first, then follow with the 'Use when...' clause, following the pattern: '[What it does]. Use when [triggers].'
Ensure the description uses third-person voice for the capability portion (e.g., 'Optimizes LLM context windows by...' rather than starting with 'This skill should be used when...').
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | The description lists no concrete actions or capabilities. It only describes when to use the skill via trigger phrases, but never explains what the skill actually does (e.g., 'partitions prompts into static/dynamic sections', 'masks observations to reduce token count'). The actions are entirely absent. | 1 / 3 |
Completeness | While the 'when' is thoroughly addressed with explicit trigger phrases, the 'what' (what the skill actually does) is completely missing. The description never explains the concrete actions or outputs of the skill. Per the rubric, missing 'what OR when' warrants a score of 1. | 1 / 3 |
Trigger Term Quality | The description includes a rich set of natural trigger terms: 'optimize context', 'reduce token costs', 'improve context efficiency', 'KV-cache optimization', 'partition context', 'context limits', 'observation masking', 'context budgeting', 'extending effective context capacity'. These cover a good range of terms a user might naturally say. | 3 / 3 |
Distinctiveness Conflict Risk | The trigger terms are fairly specific to context optimization and token management, which is a relatively narrow niche. However, without knowing what the skill actually does, there's some risk of overlap with general prompt engineering or LLM optimization skills. The specificity of terms like 'KV-cache optimization' and 'observation masking' helps, but the lack of concrete capability description weakens distinctiveness. | 2 / 3 |
Total | 7 / 12 Passed |
Implementation
35%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill is comprehensive in coverage but suffers from significant verbosity, explaining concepts Claude already knows at length rather than providing terse, actionable instructions. The code examples are pseudocode calling undefined functions rather than executable patterns. The structure would benefit greatly from moving detailed topic sections into separate reference files and keeping the main skill as a lean decision-making guide.
Suggestions
Cut the body by 50%+: remove explanations of what KV-cache is, what context windows do, and why optimization matters — Claude knows these. Focus on the specific thresholds, ordering rules, and decision table.
Replace pseudocode examples with executable, complete code snippets or remove them — `compact_context()` and `store_observation()` are undefined and not actionable. Either provide real implementations or link to a reference file that does.
Move the detailed subsections (Compaction Strategies, Observation Masking, KV-Cache Optimization, Context Partitioning, Budget Management) into the referenced optimization_techniques.md file, keeping only the priority ordering and decision table in the main skill.
Add explicit validation/verification steps to workflows — e.g., after compaction, verify summary against task goals; after partitioning, validate all sub-agent results before merging.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is extremely verbose at ~300+ lines, explaining many concepts Claude already understands (what KV-cache is, what context windows are, how summarization works). Sections like 'Core Concepts' and 'Detailed Topics' extensively explain mechanisms rather than providing terse, actionable instructions. The 'When to Activate' section lists obvious triggers. Much of this could be cut by 60%+ without losing actionable value. | 1 / 3 |
Actionability | The code examples are simplistic pseudocode-level snippets (e.g., `if context_tokens / context_limit > 0.8: context = compact_context(context)`) that call undefined functions and aren't executable. The decision framework table and performance targets provide some concrete guidance, but the skill mostly describes strategies conceptually rather than providing copy-paste-ready implementations or specific commands. | 2 / 3 |
Workflow Clarity | There is a clear priority ordering (KV-cache → masking → compaction → partitioning) and a decision framework table, but the multi-step processes lack explicit validation checkpoints and feedback loops. For example, the compaction workflow doesn't include a 'verify summary quality' step, and the partitioning workflow mentions validation but doesn't provide concrete verification steps. The Gotchas section partially compensates by describing failure modes. | 2 / 3 |
Progressive Disclosure | The skill references external files (optimization_techniques.md, related skills) with 'Read when' annotations, which is good. However, the body itself is monolithic — the detailed topics on compaction, masking, KV-cache, partitioning, and budget management could each be separate reference files, with the SKILL.md serving as a concise overview. Instead, everything is inline, creating a wall of text. | 2 / 3 |
Total | 7 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
7a95d94
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.