CtrlK
BlogDocsLog inGet started
Tessl Logo

tracking-model-versions

Build this skill enables AI assistant to track and manage ai/ml model versions using the model-versioning-tracker plugin. it should be used when the user asks to manage model versions, track model lineage, log model performance, or implement version control f... Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.

57

Quality

50%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/ai-ml/model-versioning-tracker/skills/tracking-model-versions/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

50%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description identifies a clear domain (AI/ML model versioning) and lists some relevant actions, but it is undermined by a truncated capability list and entirely boilerplate 'when' guidance that provides no real trigger information. The generic closing sentence ('Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.') is filler that adds zero discriminative value for skill selection. Additionally, the description uses second/third person inconsistently and begins with 'Build this skill enables AI assistant', which is awkward and not a proper third-person action description.

Suggestions

Replace the boilerplate 'Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.' with explicit trigger conditions, e.g., 'Use when the user asks about model versioning, experiment tracking, model registry, ML model lineage, or comparing model performance across versions.'

Complete the truncated text ('version control f...') and list all concrete actions the skill performs, such as 'register new model versions, compare metrics across versions, roll back to previous model checkpoints, visualize model lineage graphs.'

Remove the 'Build this skill enables AI assistant to' preamble and rewrite in direct third-person voice, e.g., 'Tracks and manages AI/ML model versions using the model-versioning-tracker plugin.'

DimensionReasoningScore

Specificity

The description names the domain (AI/ML model versioning) and some actions (track model versions, log model performance, track model lineage, implement version control), but the list appears truncated ('version control f...') and the actions are not fully comprehensive or concrete.

2 / 3

Completeness

The 'what' is partially addressed (manage model versions, track lineage, log performance), but the 'when' clause is entirely boilerplate ('Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.') and provides no explicit, actionable trigger guidance.

2 / 3

Trigger Term Quality

It includes some relevant keywords like 'model versions', 'model lineage', 'model performance', and 'version control', but the generic trailing phrase 'Trigger with relevant phrases based on skill purpose' adds no value and common user variations (e.g., 'ML model registry', 'model tracking', 'experiment tracking') are missing.

2 / 3

Distinctiveness Conflict Risk

The AI/ML model versioning domain is somewhat specific, but the vague trailing text and overlap with general version control or MLOps skills could cause confusion. It's not generic enough to conflict with many skills, but not sharp enough to be clearly distinct.

2 / 3

Total

8

/

12

Passed

Implementation

50%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill provides a solid structural overview of MLflow-based model versioning with clear sections and a useful error handling table. However, it falls short on actionability by describing workflows in prose rather than providing executable code examples, and it lacks validation checkpoints in its multi-step process. The referenced bundle files don't exist, undermining the progressive disclosure strategy.

Suggestions

Add at least one complete, executable Python code block showing the core workflow (start run, log params/metrics, log model, register, transition stage) instead of just naming API calls in prose.

Add explicit validation checkpoints in the workflow, e.g., 'Verify the run was logged: `mlflow.search_runs(experiment_ids=[exp_id])` should show your run' after step 3, and a verification step after stage transitions.

Convert the narrative examples into concrete input/output pairs with actual code snippets and expected MLflow output.

Provide the referenced bundle files (model_card_template.md and example_mlflow_workflow.yaml) or remove the references to avoid dead links.

DimensionReasoningScore

Conciseness

The skill is moderately verbose. The overview restates what the description already covers, the examples section describes scenarios in prose rather than showing executable code, and some explanations (like what model cards contain) are things Claude already knows. However, the error handling table and instructions are reasonably efficient.

2 / 3

Actionability

Instructions reference specific CLI commands and API calls (e.g., `mlflow.register_model()`, `client.transition_model_version_stage()`), but no complete executable code blocks are provided. The examples are narrative descriptions rather than copy-paste-ready code snippets. Key details like how to actually set up the tracking URI in code or a complete Python workflow are missing.

2 / 3

Workflow Clarity

The 8-step workflow is clearly sequenced and covers the full lifecycle from connection to model cards. However, there are no explicit validation checkpoints or feedback loops — step 5 mentions stage transitions but doesn't include verification that the transition succeeded, and there's no 'validate before proceeding' pattern for critical steps like model registration or artifact upload.

2 / 3

Progressive Disclosure

The skill references external files (`assets/model_card_template.md` and `assets/example_mlflow_workflow.yaml`) which is good progressive disclosure design, but no bundle files are provided, making these references dead links. The main file itself is fairly long with inline content (error table, examples, resources) that could be split out, though the structure with clear sections is reasonable.

2 / 3

Total

8

/

12

Passed

Validation

81%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation9 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

allowed_tools_field

'allowed-tools' contains unusual tool name(s)

Warning

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

9

/

11

Passed

Repository
jeremylongshore/claude-code-plugins-plus-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.