Confusion Matrix Generator - Auto-activating skill for ML Training. Triggers on: confusion matrix generator, confusion matrix generator Part of the ML Training skill category.
31
3%
Does it follow best practices?
Impact
72%
0.96xAverage score across 3 eval scenarios
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./planned-skills/generated/07-ml-training/confusion-matrix-generator/SKILL.mdQuality
Discovery
7%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This description is essentially a boilerplate auto-generated stub that repeats the skill name without providing any meaningful detail about capabilities, actions, or usage triggers. It lacks concrete actions, natural trigger terms, and explicit 'when to use' guidance, making it nearly useless for skill selection among multiple options.
Suggestions
Add specific concrete actions the skill performs, e.g., 'Generates confusion matrices from classification model predictions, visualizes true/false positive/negative rates, and computes accuracy, precision, recall, and F1 scores.'
Add an explicit 'Use when...' clause with natural trigger terms, e.g., 'Use when the user asks to evaluate a classifier, visualize prediction errors, generate a confusion matrix, or compute classification metrics.'
Remove the redundant duplicate trigger term and instead include natural keyword variations users might say, such as 'model evaluation', 'classification report', 'prediction accuracy', 'TP/FP/TN/FN'.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | The description names the domain ('ML Training') and mentions 'confusion matrix generator' but does not describe any concrete actions like 'generates confusion matrices from classification results, visualizes prediction accuracy, computes precision/recall metrics.' It merely states the skill name without elaborating on what it actually does. | 1 / 3 |
Completeness | The description fails to clearly answer 'what does this do' beyond naming itself, and the 'when' clause is just a redundant repetition of the skill name rather than explicit trigger guidance. There is no 'Use when...' clause with meaningful context. | 1 / 3 |
Trigger Term Quality | The only trigger term is 'confusion matrix generator' repeated twice. It misses natural variations users would say such as 'confusion matrix', 'classification results', 'prediction accuracy', 'model evaluation', 'true positives', 'false positives', etc. | 1 / 3 |
Distinctiveness Conflict Risk | The term 'confusion matrix' is fairly specific to a particular ML evaluation task, which provides some distinctiveness. However, the vague 'ML Training' category and lack of detailed scope could cause overlap with other ML-related skills. | 2 / 3 |
Total | 5 / 12 Passed |
Implementation
0%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill is an empty template with no actual instructional content. It repeatedly references 'confusion matrix generator' without ever explaining how to generate one, providing no code examples, no library recommendations, no workflow steps, and no concrete guidance of any kind. It fails on every dimension of the rubric.
Suggestions
Add executable code examples showing how to generate a confusion matrix using sklearn (e.g., `from sklearn.metrics import confusion_matrix; cm = confusion_matrix(y_true, y_pred)`) and how to visualize it with seaborn or matplotlib.
Replace the abstract 'Capabilities' and 'When to Use' sections with a concrete workflow: compute predictions, generate the matrix, visualize/interpret results, with specific code for each step.
Include concrete examples with sample input (y_true, y_pred arrays) and expected output (the matrix values and a visualization command).
Remove all meta-description content ('This skill provides automated assistance...') and replace with actionable, token-efficient technical instructions.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The content is entirely filler and meta-description. It explains what the skill does in abstract terms without providing any actual technical content. Every section restates the same vague idea ('confusion matrix generator') without adding substance. | 1 / 3 |
Actionability | There is zero concrete guidance—no code, no commands, no examples of how to actually generate a confusion matrix. The content describes rather than instructs, offering only vague promises like 'provides step-by-step guidance' without delivering any. | 1 / 3 |
Workflow Clarity | No workflow, steps, or process is defined. There are no sequences, no validation checkpoints, and no actual instructions for generating a confusion matrix. | 1 / 3 |
Progressive Disclosure | The content is a monolithic block of placeholder text with no meaningful structure. There are no references to detailed files, no layered organization, and no navigable sections with real content. | 1 / 3 |
Total | 4 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
allowed_tools_field | 'allowed-tools' contains unusual tool name(s) | Warning |
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 9 / 11 Passed | |
3e83543
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.