CtrlK
BlogDocsLog inGet started
Tessl Logo

confusion-matrix-generator

Confusion Matrix Generator - Auto-activating skill for ML Training. Triggers on: confusion matrix generator, confusion matrix generator Part of the ML Training skill category.

31

0.96x
Quality

3%

Does it follow best practices?

Impact

72%

0.96x

Average score across 3 eval scenarios

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./planned-skills/generated/07-ml-training/confusion-matrix-generator/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

7%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This description is essentially a boilerplate auto-generated stub that repeats the skill name without providing any meaningful detail about capabilities, actions, or usage triggers. It lacks concrete actions, natural trigger terms, and explicit 'when to use' guidance, making it nearly useless for skill selection among multiple options.

Suggestions

Add specific concrete actions the skill performs, e.g., 'Generates confusion matrices from classification model predictions, visualizes true/false positive/negative rates, and computes accuracy, precision, recall, and F1 scores.'

Add an explicit 'Use when...' clause with natural trigger terms, e.g., 'Use when the user asks to evaluate a classifier, visualize prediction errors, generate a confusion matrix, or compute classification metrics.'

Remove the redundant duplicate trigger term and instead include natural keyword variations users might say, such as 'model evaluation', 'classification report', 'prediction accuracy', 'TP/FP/TN/FN'.

DimensionReasoningScore

Specificity

The description names the domain ('ML Training') and mentions 'confusion matrix generator' but does not describe any concrete actions like 'generates confusion matrices from classification results, visualizes prediction accuracy, computes precision/recall metrics.' It merely states the skill name without elaborating on what it actually does.

1 / 3

Completeness

The description fails to clearly answer 'what does this do' beyond naming itself, and the 'when' clause is just a redundant repetition of the skill name rather than explicit trigger guidance. There is no 'Use when...' clause with meaningful context.

1 / 3

Trigger Term Quality

The only trigger term is 'confusion matrix generator' repeated twice. It misses natural variations users would say such as 'confusion matrix', 'classification results', 'prediction accuracy', 'model evaluation', 'true positives', 'false positives', etc.

1 / 3

Distinctiveness Conflict Risk

The term 'confusion matrix' is fairly specific to a particular ML evaluation task, which provides some distinctiveness. However, the vague 'ML Training' category and lack of detailed scope could cause overlap with other ML-related skills.

2 / 3

Total

5

/

12

Passed

Implementation

0%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill is an empty template with no actual instructional content. It repeatedly references 'confusion matrix generator' without ever explaining how to generate one, providing no code examples, no library recommendations, no workflow steps, and no concrete guidance of any kind. It fails on every dimension of the rubric.

Suggestions

Add executable code examples showing how to generate a confusion matrix using sklearn (e.g., `from sklearn.metrics import confusion_matrix; cm = confusion_matrix(y_true, y_pred)`) and how to visualize it with seaborn or matplotlib.

Replace the abstract 'Capabilities' and 'When to Use' sections with a concrete workflow: compute predictions, generate the matrix, visualize/interpret results, with specific code for each step.

Include concrete examples with sample input (y_true, y_pred arrays) and expected output (the matrix values and a visualization command).

Remove all meta-description content ('This skill provides automated assistance...') and replace with actionable, token-efficient technical instructions.

DimensionReasoningScore

Conciseness

The content is entirely filler and meta-description. It explains what the skill does in abstract terms without providing any actual technical content. Every section restates the same vague idea ('confusion matrix generator') without adding substance.

1 / 3

Actionability

There is zero concrete guidance—no code, no commands, no examples of how to actually generate a confusion matrix. The content describes rather than instructs, offering only vague promises like 'provides step-by-step guidance' without delivering any.

1 / 3

Workflow Clarity

No workflow, steps, or process is defined. There are no sequences, no validation checkpoints, and no actual instructions for generating a confusion matrix.

1 / 3

Progressive Disclosure

The content is a monolithic block of placeholder text with no meaningful structure. There are no references to detailed files, no layered organization, and no navigable sections with real content.

1 / 3

Total

4

/

12

Passed

Validation

81%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation9 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

allowed_tools_field

'allowed-tools' contains unusual tool name(s)

Warning

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

9

/

11

Passed

Repository
jeremylongshore/claude-code-plugins-plus-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.