Debug complex issues using competing hypotheses with parallel investigation, evidence collection, and root cause arbitration. Use this skill when debugging bugs with multiple potential causes, performing root cause analysis, or organizing parallel investigation workflows.
84
81%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a strong skill description that clearly articulates a specific debugging methodology (competing hypotheses with parallel investigation), provides concrete actions, and includes an explicit 'Use this skill when' clause with natural trigger terms. It is well-differentiated from generic debugging skills by emphasizing its structured, multi-hypothesis approach.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'competing hypotheses', 'parallel investigation', 'evidence collection', and 'root cause arbitration'. These describe a clear methodology with distinct steps. | 3 / 3 |
Completeness | Clearly answers both what ('Debug complex issues using competing hypotheses with parallel investigation, evidence collection, and root cause arbitration') and when ('when debugging bugs with multiple potential causes, performing root cause analysis, or organizing parallel investigation workflows') with an explicit 'Use this skill when' clause. | 3 / 3 |
Trigger Term Quality | Includes strong natural trigger terms users would say: 'debugging', 'bugs', 'multiple potential causes', 'root cause analysis', 'parallel investigation'. These cover common ways users describe complex debugging scenarios. | 3 / 3 |
Distinctiveness Conflict Risk | The focus on competing hypotheses, parallel investigation, and root cause arbitration creates a distinct niche that differentiates it from general debugging or simple troubleshooting skills. The methodology-specific language makes it unlikely to conflict with basic debugging skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
62%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill provides a solid conceptual framework for systematic debugging using competing hypotheses, with particularly strong workflow clarity in the arbitration protocol. However, it leans more toward a methodology description than actionable instructions—it lacks concrete tool usage, specific commands for investigation, or executable code examples. The failure mode categories, while comprehensive, largely enumerate things Claude already knows.
Suggestions
Add concrete, executable examples of how to actually investigate hypotheses (e.g., specific grep/search commands, logging insertion patterns, or tool invocations for evidence collection)
Trim the failure mode categories to brief bullet points or move them to a separate reference file—Claude already knows what race conditions and off-by-one errors are
Include a concrete worked example showing the full workflow from bug report through hypothesis generation, evidence collection, and arbitration to root cause identification
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is reasonably well-structured but includes some content that Claude would already know (e.g., basic definitions of logic errors, off-by-one errors, race conditions). The failure mode categories are somewhat verbose lists of things Claude already understands. The evidence table and confidence levels add value but could be tighter. | 2 / 3 |
Actionability | The skill provides a structured methodology with clear categories and tables, but lacks concrete executable examples. There are no actual commands, code snippets for investigation, or specific tool usage patterns. The citation format example is good, but the overall guidance is more of a conceptual framework than step-by-step executable instructions. | 2 / 3 |
Workflow Clarity | The arbitration protocol provides a clear 4-step sequence with explicit categorization criteria, ranking methodology, decision logic for ambiguous cases, and a validation checklist before declaring a fix. The workflow handles edge cases (compound issues, no confirmed hypotheses) and includes a feedback loop for generating new hypotheses. | 3 / 3 |
Progressive Disclosure | The content is well-organized with clear headers and sections, but it's a fairly long monolithic document. The 6 failure mode categories could be referenced from a separate file, and there's no mention of or linking to supplementary materials. For a skill of this length (~100+ lines), some content splitting would improve navigation. | 2 / 3 |
Total | 9 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
27a7ed9
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.