Writes rigorous mathematical proofs for ML/AI theory. Use when asked to prove a theorem, lemma, proposition, or corollary, fill in missing proof steps, formalize a proof sketch, 补全证明, 写证明, 证明某个命题, or determine whether a claimed proof can actually be completed under the stated assumptions.
89
88%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that clearly defines its scope (mathematical proofs for ML/AI theory), lists multiple concrete actions, provides explicit trigger guidance via a 'Use when...' clause, and includes multilingual trigger terms. It uses proper third-person voice and is concise yet comprehensive, making it easy for Claude to select appropriately from a large skill set.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'prove a theorem, lemma, proposition, or corollary', 'fill in missing proof steps', 'formalize a proof sketch', and 'determine whether a claimed proof can actually be completed under the stated assumptions'. These are distinct, well-defined tasks. | 3 / 3 |
Completeness | Clearly answers both 'what' (writes rigorous mathematical proofs for ML/AI theory) and 'when' with an explicit 'Use when...' clause listing multiple specific trigger scenarios. Both dimensions are thoroughly addressed. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural trigger terms including 'prove', 'theorem', 'lemma', 'proposition', 'corollary', 'proof steps', 'proof sketch', and even Chinese-language equivalents ('补全证明', '写证明', '证明某个命题'). These are terms users would naturally use when requesting proof assistance. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive — scoped specifically to mathematical proofs in the ML/AI theory domain, with very specific trigger terms like 'theorem', 'lemma', 'proof sketch', and 'formalize'. Unlikely to conflict with general math, coding, or writing skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
77%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a strong, well-structured skill for a complex intellectual task. Its greatest strengths are the clear multi-step workflow with explicit validation gates, the honest handling of unprovable claims, and the concrete output template. The main weakness is moderate verbosity — some sections are redundant and the skill could be tightened by consolidating overlapping extraction/normalization steps.
Suggestions
Consolidate the Inputs section and Step 1/Step 2 to eliminate redundant extraction and normalization checklists — these currently overlap significantly.
Consider moving the Required File Structure template to a separate reference file to reduce the main skill's length.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is reasonably efficient for a complex multi-step workflow, but contains some redundancy (e.g., the extraction checklist in Step 1 overlaps heavily with the Inputs section, and Step 2 partially repeats Step 1's extraction). Some bullet points like 'never write math in plain text' are things Claude already knows. However, most content earns its place given the complexity of the task. | 2 / 3 |
Actionability | The skill provides highly concrete, specific guidance: exact status labels, a complete output file template in markdown, explicit lists of what to check at each step, specific banned phrases ('clearly', 'obviously'), and clear output modes for each scenario. This is an instruction-only skill where the guidance is fully actionable without code. | 3 / 3 |
Workflow Clarity | The 6-step workflow is clearly sequenced with explicit validation checkpoints (Step 3 feasibility triage before writing, Step 6 final verification checklist). It includes feedback loops: if a step can't be justified, downgrade status and write a blockage report. The three output modes provide clear branching logic for different outcomes. | 3 / 3 |
Progressive Disclosure | The content is well-organized with clear sections and headers, but it is quite long and monolithic — all content is inline in a single file with no references to supplementary materials. The Required File Structure template, output modes, and key rules could potentially be split out, though for a skill of this nature the inline approach is defensible. | 2 / 3 |
Total | 10 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
allowed_tools_field | 'allowed-tools' contains unusual tool name(s) | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 9 / 11 Passed | |
700fbe2
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.