Updates broken pytest tests after intentional code changes — triaging assertion failures from mock-coupling failures from genuine regressions, using Python's introspection to automate where safe. Use when a refactor or API change leaves a pile of failing tests and you need to decide update vs. fix vs. delete.
Install with Tessl CLI
npx tessl i github:santosomar/general-secure-coding-agent-skills --skill python-test-updater97
Quality
96%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that clearly articulates a specific, well-defined use case. It uses third person voice correctly, provides concrete actions (triaging different failure types, automating with introspection), and includes an explicit 'Use when' clause with natural trigger terms. The description effectively distinguishes itself from general testing or Python skills.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'triaging assertion failures from mock-coupling failures from genuine regressions' and 'using Python's introspection to automate where safe'. Describes concrete decision-making (update vs. fix vs. delete). | 3 / 3 |
Completeness | Clearly answers both what ('Updates broken pytest tests...triaging...using Python's introspection') AND when ('Use when a refactor or API change leaves a pile of failing tests and you need to decide update vs. fix vs. delete'). | 3 / 3 |
Trigger Term Quality | Includes natural keywords users would say: 'pytest', 'tests', 'refactor', 'API change', 'failing tests', 'assertion failures', 'mock'. These are terms developers naturally use when dealing with broken tests. | 3 / 3 |
Distinctiveness Conflict Risk | Highly specific niche: pytest test repair after intentional code changes. The combination of 'pytest', 'refactor/API change', and 'triaging failure types' creates a distinct trigger profile unlikely to conflict with general testing or coding skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
92%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a high-quality skill that efficiently teaches Python test triage with concrete, actionable guidance. The failure taxonomy table, executable code examples, and explicit 'Do not' guardrails make it immediately useful. The structured output format ensures consistent, auditable test updates. Minor improvement possible in progressive disclosure for very complex scenarios.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Extremely lean and efficient. No explanation of what pytest is or how Python works. Every section delivers actionable information Claude wouldn't already know. The cross-reference to java-test-updater avoids repetition. | 3 / 3 |
Actionability | Provides executable code examples throughout (conftest.py shim, pytest commands, concrete test rewrites). The failure taxonomy table maps specific error messages to specific actions. Copy-paste ready examples for mock decoupling and assertion fixes. | 3 / 3 |
Workflow Clarity | Clear multi-step workflows with explicit validation: snapshot update workflow has numbered steps with review-before-update checkpoint. The 'Do not' section provides explicit guardrails. Output format creates a structured triage checklist with classification steps. | 3 / 3 |
Progressive Disclosure | Content is well-organized with clear sections and tables, but everything is inline in one file. The reference to java-test-updater is good, but the skill could benefit from splitting detailed examples (like the conftest shim patterns) into a separate reference file for complex migrations. | 2 / 3 |
Total | 11 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.