Applies the scientific method to debugging by helping users form specific, testable hypotheses, design targeted experiments, and systematically confirm or reject theories to find root causes. Use when a user says their code isn't working, they're getting an error, something broke, they want to troubleshoot a bug, or they're trying to figure out what's causing an issue. Concrete actions include isolating failing components, forming and testing hypotheses, analyzing error messages, tracing execution paths, and interpreting test results to narrow down root causes.
Install with Tessl CLI
npx tessl i github:rohitg00/skillkit --skill hypothesis-testing90
Does it follow best practices?
If you maintain this skill, you can automatically optimize it using the tessl CLI to improve its score:
npx tessl skill review --optimize ./path/to/skillValidation for skill structure
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that hits all the marks. It clearly articulates the unique methodology (scientific method for debugging), provides comprehensive trigger terms using natural user language, and lists specific concrete actions. The description successfully distinguishes itself from generic debugging help through its hypothesis-driven approach.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'isolating failing components, forming and testing hypotheses, analyzing error messages, tracing execution paths, and interpreting test results to narrow down root causes.' | 3 / 3 |
Completeness | Clearly answers both what (applies scientific method to debugging, specific actions listed) AND when (explicit 'Use when' clause with multiple trigger scenarios). The structure is exemplary. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural user language: 'code isn't working', 'getting an error', 'something broke', 'troubleshoot a bug', 'figure out what's causing an issue' - these are exactly how users naturally describe debugging needs. | 3 / 3 |
Distinctiveness Conflict Risk | The scientific method/hypothesis-driven approach creates a clear niche distinct from general coding help or simple error fixing. The methodology focus (forming hypotheses, designing experiments) differentiates it from basic debugging skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
77%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a well-structured debugging skill with strong actionability and clear workflow guidance. The scientific method framework is effectively translated into practical debugging steps with concrete code examples. The main weakness is moderate verbosity - some explanatory text could be trimmed since Claude understands the scientific method concept, and the lengthy templates could be externalized.
Suggestions
Trim explanatory prose in sections 1-5; the headers and examples convey the method without needing paragraph explanations
Consider moving the detailed 'Hypothesis Tracking Template' to a separate TEMPLATES.md file and referencing it
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The content is mostly efficient but includes some unnecessary verbosity, particularly in the explanatory sections. The templates and examples are useful but could be tightened - Claude doesn't need extensive explanation of the scientific method concept itself. | 2 / 3 |
Actionability | Provides concrete, executable code examples for timing, data, and state hypothesis testing. The templates are copy-paste ready, and the examples show specific patterns with real TypeScript code that can be directly applied. | 3 / 3 |
Workflow Clarity | The 5-step scientific debugging method is clearly sequenced with explicit validation at each stage. The decision tree provides clear checkpoints and feedback loops (results conclusive? → design better test). The hypothesis tracking template enforces systematic progression. | 3 / 3 |
Progressive Disclosure | Content is reasonably structured with clear sections, but the skill is somewhat monolithic at ~150 lines. The 'Integration with Other Skills' section appropriately references related skills, but the main content could benefit from splitting detailed templates into separate reference files. | 2 / 3 |
Total | 10 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.