CtrlK
BlogDocsLog inGet started
Tessl Logo

agent-challenges

Agent skill for challenges - invoke with $agent-challenges

Install with Tessl CLI

npx tessl i github:ruvnet/claude-flow --skill agent-challenges
What are skills?

40

1.59x

Does it follow best practices?

Evaluation99%

1.59x

Agent success when using this skill

Validation for skill structure

SKILL.md
Review
Evals

Discovery

0%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This description is critically deficient across all dimensions. It provides no information about what the skill does, when to use it, or what domain it operates in. The word 'challenges' is too generic to serve as a meaningful trigger or differentiator.

Suggestions

Define what 'challenges' means in this context - specify the concrete actions the skill performs (e.g., 'Creates coding challenges', 'Solves puzzle problems', 'Manages competitive programming tasks')

Add an explicit 'Use when...' clause with natural trigger terms users would actually say when needing this skill

Include domain-specific keywords and file types/formats if applicable to distinguish this skill from others

DimensionReasoningScore

Specificity

The description provides no concrete actions whatsoever. 'Agent skill for challenges' is completely abstract with no indication of what the skill actually does.

1 / 3

Completeness

The description fails to answer both 'what does this do' and 'when should Claude use it'. There is no 'Use when...' clause and no explanation of capabilities.

1 / 3

Trigger Term Quality

The only potential trigger term is 'challenges' which is extremely generic and could apply to countless contexts. '$agent-challenges' is a command syntax, not a natural user phrase.

1 / 3

Distinctiveness Conflict Risk

'Challenges' is extremely vague and could conflict with any skill involving problem-solving, debugging, troubleshooting, or competitive tasks. No clear niche is established.

1 / 3

Total

4

/

12

Passed

Implementation

14%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This skill reads more like a marketing document or persona description than actionable technical guidance. It spends most of its tokens explaining what the agent does conceptually rather than how to actually accomplish tasks. The MCP tool examples are a good start but lack the concrete workflows, validation steps, and error handling needed for reliable execution.

Suggestions

Remove explanatory sections about what challenges, leaderboards, and gamification are - Claude already knows these concepts. Focus only on Flow Nexus-specific implementation details.

Add concrete workflows with validation steps, e.g., 'To submit a solution: 1. Validate syntax locally, 2. Call challenge_submit, 3. Check response.status, 4. If failed, parse response.errors for feedback'

Include example responses from MCP tools showing expected data structures and how to handle success/failure cases

Split detailed category descriptions and gamification features into separate reference files, keeping SKILL.md as a concise quick-start guide

DimensionReasoningScore

Conciseness

Extremely verbose with extensive explanations of concepts Claude already knows (what gamification is, what algorithms are, what leaderboards do). Lists like 'challenge categories you manage' and 'gamification features you leverage' explain obvious concepts rather than providing actionable guidance.

1 / 3

Actionability

Provides some concrete MCP tool calls with example parameters, which is useful. However, the code examples are incomplete (placeholder IDs, pseudocode-style solution_code), and there's no guidance on actual implementation, error handling, or expected responses.

2 / 3

Workflow Clarity

The 'challenge curation approach' lists abstract steps like 'Skill Assessment' and 'Challenge Selection' without concrete sequences, validation checkpoints, or feedback loops. No guidance on what to do when submissions fail or how to handle edge cases.

1 / 3

Progressive Disclosure

Monolithic wall of text with no references to external documentation. All content is inline with no structure for discovery. Categories, features, and standards are all dumped in one file without clear navigation or separation of concerns.

1 / 3

Total

5

/

12

Passed

Validation

100%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation11 / 11 Passed

Validation for skill structure

No warnings or errors.

Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.