Pytest-first Python testing with emphasis on fakes over mocks. Covers unit, integration, and async tests; fixture design; coverage setup; and debugging test failures. Use when writing tests, reviewing test quality, designing fixtures, setting up pytest, or debugging failures—e.g., "write unit tests for new feature", "fixture design patterns", "fakes vs mocks comparison", "fix failing tests".
94
93%
Does it follow best practices?
Impact
93%
1.10xAverage score across 3 eval scenarios
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that hits all the marks. It provides specific capabilities (unit/integration/async tests, fixtures, coverage, debugging), uses natural trigger terms developers would actually say, includes an explicit 'Use when...' clause with concrete examples, and carves out a distinct niche with its pytest-first and fakes-over-mocks philosophy.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'unit, integration, and async tests; fixture design; coverage setup; and debugging test failures'. These are clear, actionable capabilities. | 3 / 3 |
Completeness | Clearly answers both what (pytest testing with fakes, unit/integration/async tests, fixtures, coverage, debugging) AND when with explicit 'Use when...' clause containing specific trigger scenarios and example phrases. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural terms users would say: 'write unit tests', 'fixture design patterns', 'fakes vs mocks', 'fix failing tests', 'pytest', 'coverage'. These match how developers naturally phrase testing requests. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive with clear niche: Python + pytest + emphasis on fakes over mocks. The specific framework (pytest) and philosophy (fakes over mocks) differentiate it from generic testing skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
87%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a well-crafted skill that efficiently communicates pytest best practices with strong emphasis on fakes over mocks. The actionable code examples and clear YOU MUST/NEVER sections provide concrete guidance. The main gap is the lack of explicit workflow sequences for common testing tasks like debugging failures or setting up a new test suite.
Suggestions
Add a brief workflow section for common tasks like 'Writing a new test' or 'Debugging a failing test' with numbered steps and validation checkpoints
Consider adding a quick troubleshooting sequence for when tests fail unexpectedly (e.g., 1. Check assertion message, 2. Verify fixture state, 3. Run with -v flag)
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is lean and efficient, assuming Claude's competence with Python and pytest. No unnecessary explanations of basic concepts; every section provides actionable value without padding. | 3 / 3 |
Actionability | Provides fully executable code examples for fakes, factory fixtures, and side effect capture. The patterns are copy-paste ready with clear correct/wrong comparisons. | 3 / 3 |
Workflow Clarity | While the skill provides clear patterns and rules, it lacks explicit multi-step workflows with validation checkpoints. The directory structure and layer references suggest a process but don't provide sequenced steps for test creation or debugging. | 2 / 3 |
Progressive Disclosure | Excellent structure with a concise overview and well-signaled one-level-deep references to detailed materials (test-doubles.md, anti-patterns.md, etc.). Content is appropriately split between quick patterns inline and detailed guidance in reference files. | 3 / 3 |
Total | 11 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
b74de5e
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.