Master systematic debugging techniques, profiling tools, and root cause analysis to efficiently track down bugs across any codebase or technology stack. Use when investigating bugs, performance issues, or unexpected behavior.
67
51%
Does it follow best practices?
Impact
95%
1.01xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./tests/ext_conformance/artifacts/agents-wshobson/developer-essentials/skills/debugging-strategies/SKILL.mdQuality
Discovery
67%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
The description has good structural completeness with an explicit 'Use when' clause and covers the general debugging domain adequately. However, it relies on somewhat abstract category names rather than concrete specific actions, and its trigger terms could be expanded to cover more natural user phrasings. The broad scope ('any codebase or technology stack') weakens distinctiveness.
Suggestions
Replace abstract categories with concrete actions, e.g., 'Analyze stack traces, set breakpoints, profile memory and CPU usage, inspect logs, bisect commits to isolate regressions'.
Expand trigger terms in the 'Use when' clause to include common user phrasings like 'error', 'crash', 'exception', 'not working', 'slow response', 'memory leak'.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Names the domain (debugging) and some actions ('debugging techniques, profiling tools, root cause analysis'), but these are still fairly abstract categories rather than concrete specific actions like 'set breakpoints, analyze stack traces, inspect memory usage'. | 2 / 3 |
Completeness | Clearly answers both 'what' (systematic debugging techniques, profiling tools, root cause analysis to track down bugs) and 'when' with an explicit 'Use when investigating bugs, performance issues, or unexpected behavior' clause. | 3 / 3 |
Trigger Term Quality | Includes some relevant keywords like 'bugs', 'performance issues', 'unexpected behavior', and 'debugging', but misses common natural variations users would say such as 'error', 'crash', 'stack trace', 'slow', 'broken', 'not working', 'exception'. | 2 / 3 |
Distinctiveness Conflict Risk | While debugging is a recognizable niche, the description is broad enough ('any codebase or technology stack') that it could overlap with language-specific debugging skills or performance optimization skills. The phrase 'performance issues' could conflict with dedicated profiling or optimization skills. | 2 / 3 |
Total | 9 / 12 Passed |
Implementation
35%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill reads more like a generic debugging tutorial or textbook chapter than a targeted skill for Claude. It is heavily padded with advice Claude already knows (scientific method, rubber duck debugging, 'take breaks,' 'read error messages'), and much of the content is abstract checklists rather than executable, actionable guidance. The language-specific code examples provide some value but are buried in an overly long document that would benefit significantly from aggressive trimming and better separation of concerns.
Suggestions
Cut at least 60% of the content: remove generic debugging advice (scientific method, mindset tips, 'read error messages,' rubber duck debugging) that Claude already knows, and focus only on specific techniques and executable code.
Convert the abstract markdown checklists (Reproduction Checklist, Information Collection, Hypothesis Formation) into concrete, executable workflows with actual commands and validation steps.
Move language-specific debugging sections (JS/TS, Python, Go) into separate reference files and keep only a brief summary with links in the main skill file.
Add explicit verification/feedback loops to the workflow, e.g., 'After applying fix, re-run the minimal reproduction case to confirm the bug is resolved before proceeding.'
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Extremely verbose and padded with content Claude already knows. Explains basic concepts like the scientific method, rubber duck debugging, 'read error messages,' and 'take breaks.' The debugging mindset section, common mistakes, and best practices are all generic advice that wastes tokens. Much of the content is markdown-within-markdown checklists that restate obvious debugging principles. | 1 / 3 |
Actionability | Contains some executable code examples (pdb, Chrome DevTools, git bisect, VS Code launch.json), but much of the skill is abstract checklists and markdown templates rather than concrete, copy-paste-ready guidance. The 'Debugging Patterns by Issue Type' sections are entirely descriptive bullet points with no executable code. | 2 / 3 |
Workflow Clarity | The four-phase process (Reproduce → Gather Info → Hypothesize → Test) provides a clear sequence, but validation checkpoints and feedback loops are absent. There's no explicit 'verify your fix actually resolved the issue' step integrated into the workflow, and the phases themselves contain only checklists rather than actionable verification steps. | 2 / 3 |
Progressive Disclosure | References to external files are listed at the bottom (references/debugging-tools-guide.md, etc.), which is good, but the main file itself is a monolithic wall of text (~400+ lines) with extensive inline content that could be split into separate reference files. The language-specific debugging sections (JS, Python, Go) would be better as separate references. | 2 / 3 |
Total | 7 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
skill_md_line_count | SKILL.md is long (537 lines); consider splitting into references/ and linking | Warning |
Total | 10 / 11 Passed | |
47823e3
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.