CtrlK
BlogDocsLog inGet started
Tessl Logo

databricks-performance-tuning

Optimize Databricks cluster and query performance. Use when jobs are running slowly, optimizing Spark configurations, or improving Delta Lake query performance. Trigger with phrases like "databricks performance", "spark tuning", "databricks slow", "optimize databricks", "cluster performance".

83

Quality

81%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

SKILL.md
Quality
Evals
Security

Quality

Discovery

89%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is a solid skill description with excellent trigger terms and completeness. Its main weakness is that the 'what' portion is somewhat high-level—it says 'optimize performance' without listing specific concrete actions like adjusting shuffle partitions, configuring autoscaling, or analyzing query plans. Adding 2-3 specific actions would elevate it further.

Suggestions

Add specific concrete actions to the first sentence, e.g., 'Optimize Databricks cluster and query performance by tuning Spark configurations, adjusting shuffle partitions, configuring autoscaling, and improving Delta Lake file compaction.'

DimensionReasoningScore

Specificity

Names the domain (Databricks cluster and query performance) and mentions some areas like Spark configurations and Delta Lake query performance, but doesn't list specific concrete actions (e.g., 'tune shuffle partitions', 'configure autoscaling', 'optimize file compaction').

2 / 3

Completeness

Clearly answers both 'what' (optimize Databricks cluster and query performance) and 'when' (jobs running slowly, optimizing Spark configurations, improving Delta Lake query performance) with explicit trigger phrases.

3 / 3

Trigger Term Quality

Includes a strong set of natural trigger terms users would actually say: 'databricks performance', 'spark tuning', 'databricks slow', 'optimize databricks', 'cluster performance'. These cover common user phrasings well.

3 / 3

Distinctiveness Conflict Risk

Highly distinctive with a clear niche around Databricks/Spark/Delta Lake performance tuning. The specific platform and technology references make it very unlikely to conflict with other skills.

3 / 3

Total

11

/

12

Passed

Implementation

72%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is a strong, highly actionable skill with excellent concrete guidance for Databricks performance tuning across multiple workload types. Its main weaknesses are moderate verbosity (could trim ~20% without losing value) and the lack of explicit validation checkpoints in the workflow, which is important given that operations like OPTIMIZE and VACUUM affect production tables. The error handling table and before/after comparison example are standout features.

Suggestions

Add explicit validation gates after key steps — e.g., 'Run DESCRIBE DETAIL and verify numFiles decreased before proceeding to VACUUM' and 'Check spark.sql(query).explain() confirms broadcast join before running at scale'.

Trim the Prerequisites and Output sections, and reduce inline comments that explain obvious things (e.g., '# 128MB' next to the byte value is fine, but comments like '# triggers expensive shuffle' could be cut since the BAD/GOOD labels already convey this).

DimensionReasoningScore

Conciseness

The skill is mostly efficient with good use of tables and code blocks, but includes some unnecessary elements like the 'Prerequisites' section (Claude knows it needs access), the 'Output' section restating what was covered, and minor verbosity in comments. The recommend_cluster function, while useful, adds bulk that could be trimmed.

2 / 3

Actionability

Excellent actionability throughout — every step includes executable SQL or Python code with specific commands, concrete instance types, real config keys, and copy-paste ready examples. The before/after comparison script and query history analysis are particularly practical.

3 / 3

Workflow Clarity

Steps are clearly numbered and sequenced, but there are no explicit validation checkpoints or feedback loops. For operations like OPTIMIZE and VACUUM (which can be destructive/long-running on large tables), there's no 'verify before proceeding' step. The DESCRIBE DETAIL check after OPTIMIZE is mentioned but not enforced as a gate.

2 / 3

Progressive Disclosure

Content is well-structured with clear sections, tables for quick reference, and external resource links for deeper dives. The skill appropriately keeps actionable content inline while pointing to official docs for comprehensive references. The 'Next Steps' section provides clear navigation to related skills.

3 / 3

Total

10

/

12

Passed

Validation

81%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation9 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

allowed_tools_field

'allowed-tools' contains unusual tool name(s)

Warning

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

9

/

11

Passed

Repository
jeremylongshore/claude-code-plugins-plus-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.