Verify claims in backlog items, skill documentation, or plugin content against primary sources using web lookups. Spawns parallel verification agents that MUST use WebFetch/WebSearch/gh — training data recall is explicitly rejected as evidence. Produces VERIFIED/REFUTED/INCONCLUSIVE verdicts with citations. Triggers on "fact check", "verify claims", "check against primary sources", or when backlog items are marked UNVERIFIED.
84
81%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Advisory
Suggest reviewing before use
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that clearly communicates what the skill does (verify claims via web lookups, produce verdicts with citations), how it works (parallel verification agents using specific tools), and when to use it (explicit trigger phrases and contextual triggers). It uses third-person voice throughout and is both specific and concise without unnecessary padding.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: verify claims against primary sources, spawn parallel verification agents, use WebFetch/WebSearch/gh tools, produce VERIFIED/REFUTED/INCONCLUSIVE verdicts with citations. Very detailed about the mechanism and outputs. | 3 / 3 |
Completeness | Clearly answers both 'what' (verify claims using web lookups, spawn parallel agents, produce verdicts with citations) and 'when' (explicit triggers: 'fact check', 'verify claims', 'check against primary sources', or when items are marked UNVERIFIED). | 3 / 3 |
Trigger Term Quality | Includes natural trigger terms users would say: 'fact check', 'verify claims', 'check against primary sources', and the contextual trigger 'UNVERIFIED'. These are terms users would naturally use when needing this capability. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive niche — fact-checking and claim verification against primary sources is a very specific function unlikely to overlap with other skills. The explicit mention of verification verdicts (VERIFIED/REFUTED/INCONCLUSIVE) and the rejection of training data recall further distinguish it. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
62%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a well-structured skill with clear workflow sequencing and good use of templates for verdicts and reports. Its main weaknesses are moderate verbosity (mermaid diagrams, concept explanations like CoVe) and incomplete actionability — the core mechanism of spawning '@fact-checker' agents is not concretely specified, and claim extraction lacks worked examples showing input→output. The evidence rules section is a strong differentiator that adds genuine value.
Suggestions
Add a concrete worked example showing a real claim being extracted from a backlog item and the resulting verdict, so Claude can see the full input→output flow.
Specify how '@fact-checker' agents are actually spawned — is this a Task tool invocation, a subagent pattern, or something else? Provide the exact tool call.
Remove or simplify the mermaid diagrams — Claude processes text natively and the flowcharts add token cost without proportional clarity gain; replace with numbered step lists.
Trim the CoVe section to just the 4 numbered steps without the explanatory paragraph — Claude understands bias mitigation without the rationale.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is reasonably well-structured but includes some content that could be tightened — the mermaid diagrams add visual bulk without much value for Claude (who processes text), and some sections like 'When NOT to Use' and the evidence rules could be more compact. The CoVe section explains a concept Claude likely understands. However, most content is instructional rather than explanatory. | 2 / 3 |
Actionability | The skill provides structured templates (verdict format, agent prompt, report format) and specific post-action commands, which is good. However, the core verification process relies on spawning '@fact-checker' agents without explaining how to actually invoke them (what tool/command?), the claim extraction is described abstractly rather than with concrete examples, and there's no executable code — just text templates and flowcharts. The 'backlog list --format json' command is referenced but not fully specified. | 2 / 3 |
Workflow Clarity | The multi-step workflow is clearly sequenced: extract claims → classify → spawn agents in waves → collect verdicts → generate report → post-actions. The CoVe requirement adds an explicit validation/cross-checking loop within each verification. The wave execution pattern handles different claim counts. Post-actions include linting and committing with specific commands. The verdict criteria provide clear decision boundaries. | 3 / 3 |
Progressive Disclosure | The skill references four external skills in the References section with relative paths, providing some progressive disclosure. However, no bundle files exist to support these references, and the skill itself is quite long (~150 lines of substantive content) with sections like the evidence rules and CoVe explanation that could potentially be extracted to reference files. The mermaid diagrams inline add length that could be separate. | 2 / 3 |
Total | 9 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
b9f32ec
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.