CtrlK
BlogDocsLog inGet started
Tessl Logo

cost-comparison

Use when comparing costs between time periods, environments, accounts, regions, or teams to understand spending differences and identify inefficiencies

39

Quality

37%

Does it follow best practices?

Impact

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/cost-analyst/skills/cost-comparison/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

40%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description provides reasonable 'when to use' guidance with relevant trigger contexts but fails to specify concrete actions or capabilities, making it unclear what the skill actually does. It reads more like a use-case statement than a skill description, lacking the 'what it does' half of a complete description. The trigger terms are decent but could be expanded with more natural user language variations.

Suggestions

Add explicit capability statements before the 'Use when' clause, e.g., 'Generates cost comparison reports, visualizes spending trends, and highlights anomalies across billing data.'

Expand trigger terms to include common variations like 'billing', 'budget', 'cloud costs', 'expense tracking', 'cost optimization', or specific platforms if applicable.

Specify the input/output formats or tools involved (e.g., 'Analyzes CSV billing exports' or 'Queries cost APIs') to improve distinctiveness and help Claude differentiate this from other financial or analytics skills.

DimensionReasoningScore

Specificity

The description lacks concrete actions. It mentions 'comparing costs' and 'identify inefficiencies' but doesn't specify what concrete operations are performed (e.g., generate reports, create charts, query billing APIs, parse CSV files). The language is abstract and domain-adjacent rather than action-oriented.

1 / 3

Completeness

The description starts with 'Use when' which addresses the 'when' aspect well, but the 'what does this do' part is weak — it only vaguely says 'comparing costs' and 'identify inefficiencies' without specifying concrete capabilities or outputs. The 'what' is implied rather than explicitly stated.

2 / 3

Trigger Term Quality

Contains some useful trigger terms like 'costs', 'spending', 'accounts', 'regions', 'teams', and 'time periods' that users might naturally mention. However, it misses common variations like 'billing', 'budget', 'cloud costs', 'AWS/GCP/Azure', 'cost optimization', 'invoice', or 'expense'.

2 / 3

Distinctiveness Conflict Risk

The focus on cost comparison across dimensions (time periods, environments, accounts, regions, teams) provides some specificity, but 'costs' and 'spending' are broad enough to overlap with general financial analysis, budgeting, or cloud management skills. Without specifying the tool, platform, or output format, it could conflict with adjacent skills.

2 / 3

Total

7

/

12

Passed

Implementation

35%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill is excessively verbose, spending most of its token budget on generic analytical concepts (percentage differences, ratios, variance analysis) and extensive output templates that Claude doesn't need spelled out. While it provides a reasonable structural framework for cost comparisons and references external files appropriately, the core content could be reduced by 60-70% without losing actionable information. The API call examples are helpful but pseudocode-level rather than executable.

Suggestions

Reduce the output format section to a brief description of required sections (executive summary, dimensional breakdown, recommendations) rather than full placeholder tables — Claude can generate appropriate tables without templates.

Remove explanations of basic math concepts (absolute difference, percentage difference, ratios) and statistical concepts (variance, standard deviation) that Claude already knows — just specify which metrics to calculate.

Move the 'Common Comparison Scenarios' and 'Advanced Techniques' sections to a separate reference file to keep SKILL.md focused on the core workflow.

Add validation checkpoints: verify API responses return data before calculating, confirm period lengths match before comparing, and validate that dimension filters return non-empty results.

DimensionReasoningScore

Conciseness

Extremely verbose at ~300+ lines. Extensive template tables, scenario walkthroughs, and advanced techniques sections explain things Claude already knows (what a percentage difference is, how to calculate ratios, what variance analysis means). The output format section alone is massive with placeholder tables that could be summarized in a few lines. Much of this is generic analytical guidance, not domain-specific knowledge.

1 / 3

Actionability

Provides pseudocode-style API call examples (get_cost_data) that show parameter patterns, which is useful. However, none of the code is truly executable — the function signatures appear illustrative rather than copy-paste ready, and the math formulas are trivially obvious (Cost_A - Cost_B). The Python security constraints are concrete and actionable, but most guidance is descriptive templates rather than executable code.

2 / 3

Workflow Clarity

Steps 1-7 provide a clear sequence for the comparison process, but there are no validation checkpoints or feedback loops. For a skill involving data queries and calculations, there's no guidance on verifying API responses, handling missing data, or confirming calculation correctness before presenting results. The workflow reads more like a checklist of considerations than a validated process.

2 / 3

Progressive Disclosure

References to external files (best-practices.md, cloudzero-tools-reference.md, etc.) are well-signaled in the 'See Also' section. However, the SKILL.md itself is monolithic — the output format templates, common scenarios, and advanced techniques sections could easily be split into separate reference files. The massive inline content undermines the progressive disclosure pattern.

2 / 3

Total

7

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Repository
Cloudzero/cloudzero-claude-marketplace
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.