CtrlK
BlogDocsLog inGet started
Tessl Logo

model-selection

Automatically applies when choosing LLM models and providers. Ensures proper model comparison, provider selection, cost optimization, fallback patterns, and multi-model strategies.

77

1.25x
Quality

54%

Does it follow best practices?

Impact

93%

1.25x

Average score across 6 eval scenarios

SecuritybySnyk

Passed

No known issues

Optimize this skill with Tessl

npx tessl skill review --optimize ./skills/ai-llm/model-selection/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

67%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description adequately communicates its domain and includes an explicit 'when' clause, which is a strength. However, the listed capabilities read more like category headings than concrete actions, and the trigger terms could be expanded to include natural user language variations (e.g., 'which model to use', 'API pricing', specific provider names). The description is functional but could be more specific and keyword-rich.

Suggestions

Replace abstract category labels with concrete actions, e.g., 'Compares token pricing across providers, recommends models based on task requirements, configures fallback chains when primary models are unavailable'.

Add natural trigger terms users would actually say, such as 'OpenAI vs Anthropic', 'which LLM to use', 'API costs', 'model selection', 'GPT', 'Claude', 'token pricing', 'rate limits'.

DimensionReasoningScore

Specificity

Names the domain (LLM models and providers) and lists some actions (model comparison, provider selection, cost optimization, fallback patterns, multi-model strategies), but these are more like category labels than concrete specific actions. For example, 'cost optimization' is vague compared to something like 'compare token pricing across providers' or 'calculate cost per request'.

2 / 3

Completeness

Clearly answers both 'what' (model comparison, provider selection, cost optimization, fallback patterns, multi-model strategies) and 'when' ('Automatically applies when choosing LLM models and providers'). The trigger condition is explicitly stated upfront.

3 / 3

Trigger Term Quality

Includes some relevant keywords like 'LLM models', 'providers', 'cost optimization', 'fallback patterns', and 'multi-model strategies'. However, it misses many natural user terms like 'OpenAI', 'Anthropic', 'GPT', 'API', 'token pricing', 'rate limits', 'which model should I use', 'cheapest model', etc.

2 / 3

Distinctiveness Conflict Risk

The domain of LLM model/provider selection is reasonably specific, but terms like 'cost optimization' and 'multi-model strategies' could overlap with general architecture or infrastructure skills. The niche is identifiable but not sharply delineated with unique trigger terms.

2 / 3

Total

9

/

12

Passed

Implementation

42%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill provides highly actionable, executable Python code for model selection and provider management, but is severely over-engineered for a SKILL.md file. The ~500+ lines of complete class implementations with full docstrings, type hints, and hardcoded pricing data make it extremely verbose and likely to become stale. The content would benefit enormously from splitting into a concise overview with references to detailed implementation files.

Suggestions

Reduce the SKILL.md to a concise overview (~50-80 lines) showing key patterns and interfaces, moving full class implementations to separate reference files (e.g., MODEL_REGISTRY.md, ROUTING.md, FALLBACK.md, COST_OPTIMIZATION.md)

Remove hardcoded pricing data (which will become stale) and instead show the schema/pattern for maintaining pricing, or note that pricing should be fetched/configured externally

Add validation checkpoints to the Auto-Apply workflow, e.g., 'Test fallback chain with simulated failures before deploying' and 'Verify routing rules match expected models for sample prompts'

Trim docstrings and inline comments that explain obvious Python patterns Claude already understands (e.g., 'Register a model', 'Get model by ID')

DimensionReasoningScore

Conciseness

Extremely verbose at ~500+ lines. The ModelRegistry, CostOptimizer, and ModelEnsemble classes are fully spelled out with extensive docstrings, type hints, and inline comments that Claude already knows how to write. The pricing data is time-sensitive and will become stale. Much of this could be condensed to patterns and key interfaces rather than complete implementations.

1 / 3

Actionability

All code is fully executable Python with complete class definitions, type annotations, and usage examples. The code is copy-paste ready with Pydantic models, async patterns, and concrete routing/fallback implementations.

3 / 3

Workflow Clarity

The Auto-Apply section provides a 7-step sequence, but there are no validation checkpoints or feedback loops. For operations involving model selection and cost optimization in production, there's no guidance on verifying that routing rules work correctly or that fallback chains are functioning before deployment.

2 / 3

Progressive Disclosure

The entire skill is a monolithic wall of code with no content split into separate files. The ModelRegistry, ModelRouter, FallbackChain, CostOptimizer, and ModelEnsemble could each be separate reference files, with the SKILL.md providing a concise overview and links. Related Skills are listed but the main content is all inline.

1 / 3

Total

7

/

12

Passed

Validation

81%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation9 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

skill_md_line_count

SKILL.md is long (713 lines); consider splitting into references/ and linking

Warning

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

9

/

11

Passed

Repository
majiayu000/claude-skill-registry
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.