Azure Machine Learning SDK v2 for Python. Use for ML workspaces, jobs, models, datasets, compute, and pipelines.
72
66%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./skills/antigravity-azure-ai-ml-py/SKILL.mdQuality
Discovery
75%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
The description is concise and well-structured with a clear 'Use for...' clause and strong distinctiveness through the Azure ML SDK v2 branding. Its main weaknesses are that it lists resource nouns rather than concrete actions (e.g., 'submit training jobs', 'deploy models') and could benefit from additional natural trigger terms users might use like 'AzureML', 'training', 'deployment', or 'endpoint'.
Suggestions
Replace or augment the noun list with concrete actions: e.g., 'Create and manage ML workspaces, submit training jobs, register models, deploy endpoints, build pipelines'.
Add common user-facing trigger variations such as 'AzureML', 'training', 'deployment', 'experiment', 'endpoint', and 'inference' to improve keyword coverage.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Names the domain (Azure ML SDK v2) and lists several resource types (workspaces, jobs, models, datasets, compute, pipelines), but these are categories rather than concrete actions like 'create pipelines', 'deploy models', or 'submit training jobs'. | 2 / 3 |
Completeness | Clearly answers both 'what' (Azure Machine Learning SDK v2 for Python) and 'when' ('Use for ML workspaces, jobs, models, datasets, compute, and pipelines') with an explicit 'Use for...' clause that provides trigger guidance. | 3 / 3 |
Trigger Term Quality | Includes relevant keywords like 'Azure Machine Learning', 'ML', 'workspaces', 'jobs', 'models', 'datasets', 'compute', 'pipelines', and 'SDK v2', but misses common user variations like 'AzureML', 'training', 'deployment', 'experiment', 'endpoint', or '.py'. | 2 / 3 |
Distinctiveness Conflict Risk | Highly distinctive due to the specific mention of 'Azure Machine Learning SDK v2 for Python', which clearly differentiates it from general ML skills, other cloud provider skills, or generic Python skills. | 3 / 3 |
Total | 10 / 12 Passed |
Implementation
57%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill provides strong, actionable code examples covering the breadth of Azure ML SDK v2 operations, making it a solid API reference. However, it suffers from being a monolithic document with no progressive disclosure, lacks workflow sequencing with validation checkpoints for multi-step operations like training pipelines, and includes some boilerplate content that doesn't add value.
Suggestions
Split detailed sections (Pipelines, Environments, Datastores) into separate reference files and link to them from a concise overview in SKILL.md
Add an explicit end-to-end workflow (e.g., 'Train and Deploy') with numbered steps and validation checkpoints (e.g., verify compute exists before submitting job, check job status before registering model)
Remove the generic 'Best Practices', 'When to Use', and 'Limitations' boilerplate sections — these don't add actionable value for Claude
Add error handling patterns for common failures (e.g., compute not found, authentication errors, job failures) with retry/recovery guidance
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The content is mostly efficient with executable code examples, but includes some unnecessary sections like the generic 'Best Practices' list (tips Claude already knows), the boilerplate 'When to Use' and 'Limitations' sections, and the operations table could be more compact. The overall length is substantial for what is essentially an API reference. | 2 / 3 |
Actionability | Every section provides fully executable, copy-paste ready Python code with correct imports, proper class instantiation, and realistic parameter values. The code examples cover all major operations (auth, data, models, compute, jobs, pipelines, environments, datastores) with concrete, runnable snippets. | 3 / 3 |
Workflow Clarity | The skill presents individual operations clearly but lacks workflow sequencing for multi-step processes. For example, creating a job involves compute, environment, data, and code setup, but there's no explicit workflow tying these together with validation checkpoints. The pipeline section shows component composition but doesn't include error handling or validation steps. | 2 / 3 |
Progressive Disclosure | The content is a monolithic wall of code examples with no references to external files for detailed topics. At ~200 lines, the pipeline details, environment setup, and datastore management could be split into separate reference files. There's no navigation structure or links to deeper documentation. | 1 / 3 |
Total | 8 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
431bfad
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.