Applies proven testing patterns — Arrange-Act-Assert (AAA), Given-When-Then, Test Data Builders, Object Mother, parameterized tests, fixtures, spies, and test doubles — to help write maintainable, reliable, and readable test suites. Use when the user asks about writing unit tests, integration tests, or end-to-end tests; structuring test cases or test suites; applying TDD or BDD practices; working with mocks, stubs, spies, or fakes; improving test coverage or reducing flakiness; or needs guidance on test organization, naming conventions, or assertions in frameworks like Jest, Vitest, pytest, or similar.
85
82%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that hits all the marks. It provides highly specific capabilities (named testing patterns), comprehensive trigger terms covering the full spectrum of testing-related user queries, an explicit 'Use when...' clause with diverse scenarios, and a clear niche that distinguishes it from other development skills. The description uses proper third-person voice throughout.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions and patterns: Arrange-Act-Assert, Given-When-Then, Test Data Builders, Object Mother, parameterized tests, fixtures, spies, test doubles. Clearly describes what it does — applying proven testing patterns to write maintainable test suites. | 3 / 3 |
Completeness | Clearly answers both 'what' (applies testing patterns like AAA, Given-When-Then, etc. to write maintainable test suites) and 'when' with an explicit 'Use when...' clause covering multiple trigger scenarios including writing tests, structuring test cases, applying TDD/BDD, working with test doubles, improving coverage, and test organization. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural terms users would say: 'unit tests', 'integration tests', 'end-to-end tests', 'TDD', 'BDD', 'mocks', 'stubs', 'spies', 'fakes', 'test coverage', 'flakiness', 'test organization', 'naming conventions', 'assertions', and specific frameworks like 'Jest', 'Vitest', 'pytest'. These are all terms users would naturally use. | 3 / 3 |
Distinctiveness Conflict Risk | Clearly occupies a distinct niche around testing patterns and practices. The specific mention of testing methodologies (TDD, BDD), test double types (mocks, stubs, spies, fakes), and testing frameworks (Jest, Vitest, pytest) makes it highly distinguishable from general coding skills or other development skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
64%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a solid, actionable reference for testing patterns with excellent concrete code examples across multiple patterns. Its main weaknesses are moderate verbosity (some sections cover knowledge Claude already has, like naming conventions and test isolation basics) and a monolithic structure that could benefit from splitting advanced content into referenced files. The workflow guidance is present but lacks explicit validation/feedback loops for iterative test development.
Suggestions
Remove or significantly trim the Test Isolation, Naming Conventions, and introductory framing sections — these cover knowledge Claude already possesses.
Add a feedback loop to the workflow section: e.g., 'Run tests → if flaky, check for shared mutable state or async timing → isolate and re-run → verify determinism before committing.'
Consider splitting framework-specific examples (Jest vs pytest vs Vitest) and advanced patterns (Object Mother, Fakes) into separate referenced files to reduce the main file's token footprint.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Generally efficient with good use of tables and code examples, but includes some unnecessary framing ('You are applying proven testing patterns...'), the pattern selection guide and combination workflows add moderate value but also bulk. The Test Isolation and Naming Conventions sections explain things Claude already knows well. | 2 / 3 |
Actionability | Every pattern includes fully executable, copy-paste-ready code examples in realistic contexts (Jest/Vitest syntax). The test doubles table, parameterized test examples, and builder patterns are all concrete and immediately usable. | 3 / 3 |
Workflow Clarity | The 'Pattern Combination Workflows' section provides useful sequencing for how patterns compose, and the verification checklist is helpful. However, there are no explicit validation checkpoints or feedback loops — e.g., no guidance on what to do when tests fail, how to diagnose flaky tests, or iterative refinement steps. | 2 / 3 |
Progressive Disclosure | The content is well-structured with clear headers and a pattern selection guide at the top, but it's a monolithic file with no references to external files for deeper dives. At ~200 lines, some patterns (e.g., Test Data Builders vs Object Mother nuances, framework-specific guidance) could be split into separate reference files. | 2 / 3 |
Total | 9 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
a9e5c83
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.