This skill generates comprehensive test reports with coverage metrics, trends, and stakeholder-friendly formats (HTML, PDF, JSON). It aggregates test results from various frameworks, calculates key metrics (coverage, pass rate, duration), and performs trend analysis. Use this skill when the user requests a test report, coverage analysis, failure analysis, or historical comparisons of test runs. Trigger terms include "test report", "coverage report", "testing trends", "failure analysis", and "historical test data".
Install with Tessl CLI
npx tessl i github:jeremylongshore/claude-code-plugins-plus-skills --skill generating-test-reports80
Quality
53%
Does it follow best practices?
Impact
99%
1.00xAverage score across 6 eval scenarios
Optimize this skill with Tessl
npx tessl skill review --optimize ./backups/skills-batch-20251204-000554/plugins/testing/test-report-generator/skills/test-report-generator/SKILL.mdDiscovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a well-crafted skill description that excels across all dimensions. It provides specific capabilities, explicit trigger terms, clear guidance on when to use it, and occupies a distinct niche in test reporting. The description follows best practices by using third person voice and including both 'what' and 'when' clauses.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'generates comprehensive test reports', 'aggregates test results', 'calculates key metrics (coverage, pass rate, duration)', 'performs trend analysis', and specifies output formats (HTML, PDF, JSON). | 3 / 3 |
Completeness | Clearly answers both what (generates test reports, aggregates results, calculates metrics, performs trend analysis) AND when ('Use this skill when the user requests a test report, coverage analysis, failure analysis, or historical comparisons'). | 3 / 3 |
Trigger Term Quality | Explicitly lists natural trigger terms users would say: 'test report', 'coverage report', 'testing trends', 'failure analysis', 'historical test data'. These are terms users would naturally use when needing this functionality. | 3 / 3 |
Distinctiveness Conflict Risk | Has a clear niche focused specifically on test reporting and coverage metrics. The explicit trigger terms and specific domain (test frameworks, coverage metrics, trend analysis) make it unlikely to conflict with general documentation or code analysis skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
7%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill content describes what test reporting is rather than instructing Claude how to do it. It lacks any executable code, specific commands, tool names, or concrete implementation details. The content reads like marketing copy explaining the concept rather than actionable guidance Claude can follow.
Suggestions
Replace abstract descriptions with concrete, executable code examples showing how to parse test results from specific frameworks (pytest, jest, etc.) and generate reports
Add specific commands or scripts for aggregating coverage data, e.g., 'coverage json -o coverage.json' or parsing pytest output
Include actual output format templates or schemas for the HTML/PDF/JSON reports Claude should generate
Remove the 'How It Works' and 'When to Use This Skill' sections entirely - Claude doesn't need to be told when to use its own skills
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The content is verbose and explains concepts Claude already knows (what test reports are, what coverage means). Sections like 'How It Works' and 'When to Use This Skill' describe rather than instruct, wasting tokens on obvious information. | 1 / 3 |
Actionability | No concrete code, commands, or executable guidance provided. Examples describe what 'the skill will' do abstractly rather than showing actual commands, scripts, or tool invocations Claude should use to generate reports. | 1 / 3 |
Workflow Clarity | Steps are vague descriptions ('Aggregate test results', 'Calculate code coverage') without specifying how to actually perform these actions. No validation checkpoints, no specific tools or commands, no error handling guidance. | 1 / 3 |
Progressive Disclosure | Content has some structure with clear section headers, but everything is inline with no references to external files. The content that exists could be much more concise, and there's no separation of quick-start vs advanced content. | 2 / 3 |
Total | 5 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
allowed_tools_field | 'allowed-tools' contains unusual tool name(s) | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 9 / 11 Passed | |
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.