Deploy and query Databricks Model Serving endpoints. Use when (1) deploying MLflow models or AI agents to endpoints, (2) creating ChatAgent/ResponsesAgent agents, (3) integrating UC Functions or Vector Search tools, (4) querying deployed endpoints, (5) checking endpoint status. Covers classical ML models, custom pyfunc, and GenAI agents.
89
86%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Quality
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is an excellent skill description that clearly defines its scope, provides explicit trigger conditions, and uses domain-specific terminology that makes it highly distinguishable. It follows the recommended pattern with a concise opening statement followed by a numbered 'Use when' list, and adds a closing sentence that broadens the scope appropriately. The description uses proper third-person voice throughout.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: deploying MLflow models, creating ChatAgent/ResponsesAgent agents, integrating UC Functions or Vector Search tools, querying deployed endpoints, and checking endpoint status. Also specifies coverage of classical ML models, custom pyfunc, and GenAI agents. | 3 / 3 |
Completeness | Clearly answers both 'what' (deploy and query Databricks Model Serving endpoints, covering classical ML, custom pyfunc, and GenAI agents) and 'when' with an explicit 'Use when' clause listing five specific trigger scenarios. | 3 / 3 |
Trigger Term Quality | Includes strong natural keywords users would say: 'Databricks', 'Model Serving', 'endpoints', 'MLflow', 'ChatAgent', 'ResponsesAgent', 'UC Functions', 'Vector Search', 'deploy', 'query', 'pyfunc', 'GenAI agents'. These cover the domain-specific terms a user working with Databricks would naturally use. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive with a clear niche around Databricks Model Serving endpoints specifically. The combination of Databricks-specific terminology (UC Functions, Vector Search, MLflow, ChatAgent/ResponsesAgent) makes it very unlikely to conflict with other skills. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
72%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a well-structured skill that excels at progressive disclosure and actionability, with clear decision tables, concrete code examples, and well-organized references to detailed sub-files. The main weaknesses are some redundancy in the content (reference tables appearing twice, MCP tools potentially duplicating sub-file content) and missing validation checkpoints in the multi-step deployment workflows. The Foundation Model API endpoint table is valuable domain-specific knowledge that Claude wouldn't have.
Suggestions
Add explicit validation checkpoints to the GenAI agent deployment workflow (e.g., 'Verify model logged: check output for model URI' after Step 5, 'If endpoint NOT_READY after 20 min, check logs' after Step 6)
Remove the duplicate reference—the 'Quick Decision' table and 'Reference Files' table overlap significantly; consolidate into one comprehensive table
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The large Foundation Model API endpoint table is useful reference data Claude wouldn't know, but the file is quite long (~200+ lines) with some redundancy—the reference files table appears twice (decision table + reference files section), and the MCP tools section repeats information likely covered in referenced files. Some sections like Prerequisites state obvious things. | 2 / 3 |
Actionability | Provides concrete, executable code examples for both GenAI agent deployment and classical ML workflows, specific MCP tool invocations with exact parameter names, and copy-paste ready commands. The ResponsesAgent output format section with wrong/correct examples is particularly actionable. | 3 / 3 |
Workflow Clarity | The GenAI agent quick start has a clear 7-step sequence, but lacks explicit validation checkpoints—there's no 'verify model was logged successfully' step between logging and deployment, no feedback loop for handling failures at each step, and Step 6 just defers to another file. The classical ML workflow is even less structured. | 2 / 3 |
Progressive Disclosure | Excellent progressive disclosure with a clear decision table upfront, concise overview content in the main file, and well-signaled one-level-deep references to 9 topic-specific files. The reference files table clearly indicates when to read each file. | 3 / 3 |
Total | 10 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
b4071a0
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.