CtrlK
BlogDocsLog inGet started
Tessl Logo

learning-opportunities

Facilitates deliberate skill development during AI-assisted coding. Offers interactive learning exercises after architectural work (new files, schema changes, refactors). Use when completing features, making design decisions, or when user asks to understand code better. Triggers on "learning exercise", "help me understand", "teach me", "why does this work", or after creating new files/modules. Do NOT use for urgent debugging, quick fixes, or when user says "just ship it".

86

Quality

83%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

SKILL.md
Quality
Evals
Security

Quality

Discovery

89%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is a strong description with excellent trigger term coverage, clear 'when to use' and 'when not to use' guidance, and a distinctive niche. Its main weakness is that the specific capabilities could be more concrete—what exactly does 'interactive learning exercises' entail? The inclusion of negative triggers is a notable strength for disambiguation in a multi-skill environment.

Suggestions

Add more concrete action descriptions, e.g., 'Generates targeted coding exercises, explains architectural patterns, quizzes on design decisions' to replace the vaguer 'offers interactive learning exercises'.

DimensionReasoningScore

Specificity

The description names the domain (skill development during AI-assisted coding) and some actions (offers interactive learning exercises after architectural work), but the concrete actions are somewhat vague—'facilitates deliberate skill development' and 'offers interactive learning exercises' don't specify what those exercises look like or what concrete outputs are produced.

2 / 3

Completeness

Clearly answers both 'what' (facilitates skill development, offers interactive learning exercises after architectural work) and 'when' (explicit 'Use when' clause with multiple trigger scenarios, plus a 'Do NOT use' clause for exclusions). This is a well-structured description covering both dimensions explicitly.

3 / 3

Trigger Term Quality

Excellent coverage of natural trigger terms: 'learning exercise', 'help me understand', 'teach me', 'why does this work', plus contextual triggers like 'after creating new files/modules'. These are phrases users would naturally say. It also includes negative triggers ('just ship it', urgent debugging) which help with disambiguation.

3 / 3

Distinctiveness Conflict Risk

This skill occupies a clear niche—learning/teaching during coding—that is distinct from typical coding, debugging, or documentation skills. The explicit negative triggers ('NOT for urgent debugging, quick fixes') further reduce conflict risk with other coding-related skills.

3 / 3

Total

11

/

12

Passed

Implementation

77%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is a well-crafted skill with excellent actionability and workflow clarity. The concrete dialogue examples and explicit behavioral rules (especially the 'pause for input' principle with [STOP] markers) make it highly usable. Minor weaknesses include some verbosity in repeated patterns across examples and the content being slightly long for a single file without more progressive disclosure.

Suggestions

Consider extracting the detailed exercise type examples into a separate EXERCISES.md file, keeping only brief descriptions and one compact example in the main skill

Consolidate the repeated '[STOP — wait for response]' pattern into a single stated rule rather than repeating it in every example block

DimensionReasoningScore

Conciseness

The content is mostly efficient but has some redundancy. The examples are helpful but slightly verbose — the template-style examples with '[STOP — wait for response]' repeated multiple times could be tightened. The anti-patterns section partially overlaps with the core principle section. However, it generally avoids explaining things Claude already knows.

2 / 3

Actionability

The skill provides highly concrete, actionable guidance with specific example dialogues showing exact phrasing, clear behavioral rules (stop after question mark), specific trigger conditions, and graduated examples for hands-on exploration. The exercise types include complete interaction scripts that are directly usable.

3 / 3

Workflow Clarity

The workflow is clearly sequenced: when to offer → how to offer → core interaction rule (pause for input) → exercise types with step-by-step flows → follow-up techniques. Each exercise type has explicit checkpoints ('[STOP — wait for response]') and clear branching logic (correct vs wrong responses). The when/when-not boundaries are crisp.

3 / 3

Progressive Disclosure

There is one reference to 'references/PRINCIPLES.md' for deeper learning science, which is good. However, the skill is moderately long (~100 lines of content) and could benefit from splitting the exercise type examples into a separate reference file, keeping the main skill leaner. The structure is well-organized with clear headers but everything is inline.

2 / 3

Total

10

/

12

Passed

Validation

100%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation11 / 11 Passed

Validation for skill structure

No warnings or errors.

Repository
tech-leads-club/agent-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.