CtrlK
BlogDocsLog inGet started
Tessl Logo

context-engineering

Understand the components, mechanics, and constraints of context in agent systems. Use when writing, editing, or optimizing commands, skills, or sub-agents prompts.

56

Quality

47%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Advisory

Suggest reviewing before use

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/customaize-agent/skills/context-engineering/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

67%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description has a clear structure with both 'what' and 'when' clauses, which is its strongest aspect. However, the capabilities described are somewhat abstract ('understand the components, mechanics, and constraints') rather than listing concrete actions, and the trigger terms could be broader to capture more natural user queries. The domain is moderately distinct but could overlap with general prompt-engineering or skill-authoring skills.

Suggestions

Replace the abstract 'understand the components, mechanics, and constraints' with specific concrete actions like 'manage token budgets, structure system prompts, optimize context windows, prioritize information in agent context'.

Add more natural trigger terms users might say, such as 'context window', 'token limit', 'system prompt', 'prompt engineering', 'agent memory', or 'context optimization'.

DimensionReasoningScore

Specificity

Names the domain (context in agent systems) and some actions (writing, editing, optimizing), but doesn't list specific concrete capabilities like 'manage token budgets', 'structure system prompts', or 'design context windows'. The phrase 'understand the components, mechanics, and constraints' is somewhat abstract.

2 / 3

Completeness

Clearly answers both 'what' (understand components, mechanics, and constraints of context in agent systems) and 'when' (Use when writing, editing, or optimizing commands, skills, or sub-agents prompts), with an explicit 'Use when...' clause.

3 / 3

Trigger Term Quality

Includes relevant terms like 'commands', 'skills', 'sub-agents', 'prompts', and 'context', but misses common natural variations users might say such as 'system prompt', 'token limit', 'context window', 'prompt engineering', or 'agent prompt design'.

2 / 3

Distinctiveness Conflict Risk

The focus on 'context in agent systems' provides some specificity, but terms like 'writing' and 'editing' 'skills' or 'prompts' could overlap with general prompt-writing skills or skill-authoring skills. The niche of 'context mechanics and constraints' is somewhat distinct but not sharply delineated.

2 / 3

Total

9

/

12

Passed

Implementation

27%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill reads like an academic paper or textbook chapter on context engineering rather than an operational skill for Claude. It is extremely verbose, explaining many concepts Claude already knows (attention mechanisms, what system prompts are, progressive disclosure as a concept), while burying actionable multi-agent workflow patterns deep in the middle of an 800+ line document—ironically falling victim to the very 'lost-in-middle' problem it describes. The content would be far more effective split across multiple files with a lean overview.

Suggestions

Reduce the skill to a concise overview (<100 lines) covering key principles and decision frameworks, moving detailed workflows (hallucination detection, lost-in-middle detection, error propagation, relevance scoring, health monitoring) into separate referenced files.

Remove explanations of concepts Claude already knows: what attention mechanisms are, what system prompts are, what tool outputs are, how position encoding works, etc. Focus only on novel operational guidance.

Replace pseudocode patterns like 'spawn_parallel([scoring_agent(...)])' with concrete Claude Code implementation examples using actual tool calls or bash commands.

Add a quick-reference decision table at the top: 'If symptom X → apply technique Y → see FILE.md' to make the skill immediately actionable without reading 800+ lines.

DimensionReasoningScore

Conciseness

This skill is extremely verbose at ~800+ lines, extensively explaining concepts Claude already understands well (attention mechanisms, what system prompts are, what tool outputs are, how PDFs work metaphorically). It includes lengthy explanations of research findings, cognitive metaphors, and conceptual frameworks that don't add actionable value. Much of the content reads like a textbook chapter rather than operational guidance.

1 / 3

Actionability

The multi-agent workflow sections provide concrete prompt templates and structured output formats that are somewhat actionable, but most are pseudocode patterns rather than executable code. The hallucination detection, lost-in-middle detection, and error propagation workflows give specific steps but rely on abstract 'spawn agent' directives without executable implementation. The earlier conceptual sections are largely descriptive rather than instructive.

2 / 3

Workflow Clarity

The multi-agent workflow sections have clear numbered steps with decision thresholds and structured outputs, which is good. However, validation checkpoints are mostly theoretical (e.g., 'calculate poisoning risk' formulas without real implementation), and the earlier conceptual sections lack any workflow structure. The health monitoring workflow mentions intervention but the restart intervention is vague ('ask user to start a new session').

2 / 3

Progressive Disclosure

This is a monolithic wall of text with no references to external files despite being extremely long. All content—fundamentals, degradation patterns, multi-agent workflows, and optimization techniques—is crammed into a single file. There are no bundle files and no references to separate documents. The content would benefit enormously from splitting into focused files with a concise overview in the main SKILL.md.

1 / 3

Total

6

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

skill_md_line_count

SKILL.md is long (1262 lines); consider splitting into references/ and linking

Warning

Total

10

/

11

Passed

Repository
NeoLabHQ/context-engineering-kit
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.