CtrlK
BlogDocsLog inGet started
Tessl Logo

ml-pipeline-workflow

Build end-to-end MLOps pipelines from data preparation through model training, validation, and production deployment. Use when creating ML pipelines, implementing MLOps practices, or automating model training and deployment workflows.

55

Quality

43%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/machine-learning-ops/skills/ml-pipeline-workflow/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

67%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description has a solid structure with an explicit 'Use when' clause and covers the MLOps domain adequately. However, it stays at a relatively high level of abstraction, listing pipeline stages rather than specific concrete actions, and the trigger terms could be expanded to cover more natural user language variations. It would benefit from more specific capabilities and additional keywords to improve both specificity and distinctiveness.

Suggestions

Add more specific concrete actions such as 'configure experiment tracking, set up model registries, implement feature stores, automate retraining schedules, manage model versioning'.

Expand trigger terms to include natural variations like 'machine learning pipeline', 'model serving', 'CI/CD for ML', 'experiment tracking', 'model monitoring', and common tool names users might reference.

DimensionReasoningScore

Specificity

Names the domain (MLOps) and lists some actions ('data preparation', 'model training', 'validation', 'production deployment'), but these are high-level pipeline stages rather than multiple specific concrete actions like 'create feature stores, configure hyperparameter tuning, set up model registries, implement A/B testing'.

2 / 3

Completeness

Clearly answers both 'what' (build end-to-end MLOps pipelines from data preparation through model training, validation, and production deployment) and 'when' (explicit 'Use when' clause covering creating ML pipelines, implementing MLOps practices, or automating model training and deployment workflows).

3 / 3

Trigger Term Quality

Includes relevant keywords like 'ML pipelines', 'MLOps', 'model training', 'deployment workflows', but misses common natural variations users might say such as 'machine learning', 'CI/CD for models', 'model serving', 'experiment tracking', 'feature engineering', 'model registry', or specific tools like 'Kubeflow', 'MLflow'.

2 / 3

Distinctiveness Conflict Risk

The MLOps focus provides some distinctiveness, but terms like 'model training' and 'data preparation' could overlap with general data science or machine learning skills. The description doesn't clearly delineate boundaries from adjacent skills like a general ML skill or a deployment/DevOps skill.

2 / 3

Total

9

/

12

Passed

Implementation

20%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill reads more like a high-level overview document or table of contents than actionable guidance. It enumerates well-known MLOps concepts, tools, and best practices that Claude already knows, without providing the concrete, executable code, specific configurations, or novel insights that would make it useful. The content is significantly bloated relative to the actionable value it delivers.

Suggestions

Replace the abstract descriptions with concrete, executable code examples—e.g., a complete minimal Airflow DAG or Dagster pipeline definition that actually runs, rather than comments saying 'See template'.

Remove sections that merely list concepts Claude already knows (e.g., 'When to Use This Skill', the enumeration of orchestration tools and deployment platforms) and focus on project-specific conventions or non-obvious patterns.

Add explicit validation checkpoints with real commands in the workflow (e.g., 'Run `great_expectations checkpoint run my_checkpoint` and only proceed if all expectations pass').

Provide at least one complete end-to-end example pipeline (even if minimal) that can be copied and adapted, rather than fragmentary pseudocode snippets.

DimensionReasoningScore

Conciseness

Extremely verbose and padded with high-level descriptions Claude already knows. Sections like 'When to Use This Skill', 'What This Skill Provides', 'Integration Points', and 'Common Patterns' are largely enumerations of well-known concepts (canary deployments, A/B testing, DAG orchestration) without adding novel, specific guidance. The 'Progressive Disclosure' section ironically describes progressive disclosure as a concept rather than implementing it.

1 / 3

Actionability

Almost no executable code or concrete commands. The Python and YAML snippets are either pseudocode, comments pointing to other files ('See references/...'), or trivially obvious list definitions. There is nothing copy-paste ready—no real pipeline code, no actual CLI commands, no concrete configuration examples.

1 / 3

Workflow Clarity

The Production Workflow section lists four phases with sub-steps in a clear sequence, which provides some structure. However, there are no explicit validation checkpoints with commands, no feedback loops for error recovery, and the troubleshooting section is generic advice rather than actionable debugging workflows.

2 / 3

Progressive Disclosure

References to external files (references/ and assets/ directories) are present and clearly signaled, which is good. However, the main file itself is a monolithic wall of text with extensive inline content that could be split out, and the referenced files may not exist. The skill talks about progressive disclosure as a concept in a dedicated section rather than actually implementing it structurally.

2 / 3

Total

6

/

12

Passed

Validation

100%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation11 / 11 Passed

Validation for skill structure

No warnings or errors.

Repository
wshobson/agents
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.