CtrlK
BlogDocsLog inGet started
Tessl Logo

iterative-development

TDD iteration loops using Claude Code Stop hooks - runs tests after each response, feeds failures back automatically

73

Quality

67%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./skills/iterative-development/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

57%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description identifies a specific and distinctive workflow (TDD loops with Claude Code Stop hooks), which makes it unlikely to conflict with other skills. However, it lacks an explicit 'Use when...' clause and could benefit from more natural trigger terms that users would actually say. The specificity of actions is moderate—it describes a single automated workflow rather than listing multiple concrete capabilities.

Suggestions

Add an explicit 'Use when...' clause, e.g., 'Use when the user wants to set up test-driven development loops, automated test feedback, or red-green-refactor workflows with Claude Code.'

Include more natural trigger term variations such as 'test-driven development', 'red-green-refactor', 'auto-run tests', 'unit tests', 'test automation loop'.

DimensionReasoningScore

Specificity

Names the domain (TDD iteration loops) and mentions specific mechanisms (Claude Code Stop hooks, runs tests, feeds failures back), but doesn't list multiple distinct concrete actions—it's more of a single workflow description than a list of capabilities.

2 / 3

Completeness

The 'what' is reasonably clear (TDD iteration loops using stop hooks that run tests and feed failures back), but there is no explicit 'Use when...' clause or equivalent trigger guidance, which caps this at 2 per the rubric guidelines.

2 / 3

Trigger Term Quality

Includes relevant terms like 'TDD', 'tests', 'failures', and 'Stop hooks', but misses common user variations like 'test-driven development', 'red-green-refactor', 'unit tests', 'test loop', or 'auto-test'. The term 'Stop hooks' is fairly technical and niche.

2 / 3

Distinctiveness Conflict Risk

The combination of 'TDD iteration loops', 'Claude Code Stop hooks', and automatic test feedback creates a very specific niche that is unlikely to conflict with other skills. This is a clearly distinct workflow.

3 / 3

Total

9

/

12

Passed

Implementation

77%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is a strong, highly actionable skill with complete executable scripts and clear workflow documentation. Its main weakness is verbosity — the philosophy section, use-case tables, and some repeated explanations could be trimmed or moved to supplementary files. The core technical content (hook configuration, TDD loop scripts, error classification) is excellent and immediately usable.

Suggestions

Remove the 'Core Philosophy' ASCII box — Claude doesn't need motivational framing, and the concept is already clear from the workflow diagram and scripts.

Move the Python variant and additional hooks (PreToolUse, SessionStart) to separate referenced files to reduce the main skill's token footprint.

DimensionReasoningScore

Conciseness

The skill includes some unnecessary content like the 'Core Philosophy' ASCII box (Claude doesn't need motivational framing), the 'When to Use TDD Loops' tables are somewhat obvious, and the explanation of how Stop hooks work is repeated in both prose and diagram. The Python variant is useful but adds significant length. However, the core technical content is mostly efficient.

2 / 3

Actionability

Provides fully executable bash scripts, complete JSON configuration for hooks, and copy-paste ready setup for both Node.js and Python projects. The scripts handle edge cases (no test files, max iterations, environment vs code errors) and are immediately usable.

3 / 3

Workflow Clarity

The workflow is clearly sequenced with an explicit diagram showing the loop flow. The scripts include validation checkpoints (test → lint → typecheck), error recovery via the exit code 2 feedback loop, iteration counting as a safety valve, and clear distinction between loopable vs non-loopable errors.

3 / 3

Progressive Disclosure

The content is well-structured with clear sections, but it's quite long for a single file. The Python variant, additional hooks (PreToolUse, SessionStart), and the 'When to Use' / 'Core Philosophy' sections could be split into referenced files. The mention of 'Load with: base.md' suggests a multi-file system but no other references are provided.

2 / 3

Total

10

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Repository
alinaqi/claude-bootstrap
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.