Develop Lakeflow Spark Declarative Pipelines (formerly Delta Live Tables) on Databricks. Use when building batch or streaming data pipelines with Python or SQL. Invoke BEFORE starting implementation.
83
80%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./examples/content-moderator/template/.agents/skills/databricks-pipelines/SKILL.mdQuality
Discovery
89%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a solid skill description with excellent trigger term coverage (including the former product name), clear when-to-use guidance, and strong distinctiveness through specific product naming. The main weakness is that it could be more specific about the concrete actions it enables (e.g., creating streaming tables, defining data quality expectations, configuring pipeline settings) rather than staying at the high level of 'develop pipelines'.
Suggestions
Add specific concrete actions like 'create materialized views, define streaming tables, configure data quality expectations, set up pipeline clusters' to improve specificity.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | The description names the domain (Lakeflow Spark Declarative Pipelines / Delta Live Tables on Databricks) and mentions batch/streaming data pipelines with Python or SQL, but does not list specific concrete actions like 'create materialized views', 'define streaming tables', 'configure expectations', etc. | 2 / 3 |
Completeness | Clearly answers both 'what' (develop Lakeflow Spark Declarative Pipelines on Databricks) and 'when' (when building batch or streaming data pipelines with Python or SQL), plus includes an explicit invocation instruction ('Invoke BEFORE starting implementation'). | 3 / 3 |
Trigger Term Quality | Includes strong natural keywords users would say: 'Lakeflow', 'Spark', 'Declarative Pipelines', 'Delta Live Tables', 'Databricks', 'batch', 'streaming', 'data pipelines', 'Python', 'SQL'. The inclusion of the former name 'Delta Live Tables' is especially helpful for users who know the old terminology. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive with specific product names (Lakeflow Spark Declarative Pipelines, Delta Live Tables, Databricks) that create a clear niche unlikely to conflict with generic data engineering or other pipeline skills. | 3 / 3 |
Total | 11 / 12 Passed |
Implementation
70%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a well-structured skill that serves effectively as a routing/decision document for a complex domain. Its greatest strengths are the comprehensive decision tree, the thorough common traps section, and excellent progressive disclosure to sub-skills. Its main weaknesses are the verbosity of the API reference tables (especially deprecated columns) and the lack of inline executable code examples for core pipeline definitions, relying entirely on referenced files for implementation details.
Suggestions
Add 1-2 minimal executable code examples inline (e.g., a simple streaming table and materialized view definition in both Python and SQL) so the skill is actionable without requiring sub-skill reads for basic cases.
Consider condensing the API reference tables by removing or collapsing the deprecated columns into a separate 'Migration from deprecated APIs' section, reducing token usage significantly.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is mostly efficient and contains genuinely useful decision trees, trap lists, and API tables that Claude wouldn't inherently know. However, the comprehensive API reference tables are extremely verbose with deprecated columns that could be condensed, and some explanatory text (e.g., 'Lakeflow Spark Declarative Pipelines (formerly Delta Live Tables / DLT) is a framework for building batch and streaming data pipelines') is unnecessary. The CLAUDE.md/AGENTS.md template content also adds bulk. | 2 / 3 |
Actionability | The skill provides concrete CLI commands for scaffolding, deploying, and running pipelines, plus specific API signatures in tables. However, it lacks executable code examples for the actual pipeline definitions (no Python or SQL code showing how to define a streaming table, materialized view, etc.), instead deferring everything to reference files. The decision tree and common traps are highly actionable as guidance but the skill itself doesn't contain copy-paste-ready pipeline code. | 2 / 3 |
Workflow Clarity | The development workflow section provides a clear 4-step sequence (validate → deploy → run → check status) with explicit commands. The decision tree provides excellent sequencing for choosing dataset types. The skill explicitly warns about deploying before running, warns about full refresh dangers, and provides clear guidance on when to use selective refresh — these serve as validation checkpoints for destructive operations. | 3 / 3 |
Progressive Disclosure | The skill excels at progressive disclosure with a clear overview structure, well-organized API reference tables that link to specific sub-skills for each feature (Python and SQL variants), and a dedicated 'Pipeline API Reference' section with clearly signaled one-level-deep references. The MANDATORY instruction to read reference files before implementing reinforces the navigation pattern. | 3 / 3 |
Total | 10 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
6338825
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.