Agent skill for challenges - invoke with $agent-challenges
Install with Tessl CLI
npx tessl i github:ruvnet/claude-flow --skill agent-challenges40
Does it follow best practices?
If you maintain this skill, you can automatically optimize it using the tessl CLI to improve its score:
npx tessl skill review --optimize ./path/to/skillEvaluation — 99%
↑ 1.59xAgent success when using this skill
Validation for skill structure
Discovery
0%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This description is critically deficient across all dimensions. It provides no information about what the skill does, when to use it, or what domain it operates in. The word 'challenges' is too generic to serve as a meaningful trigger or differentiator.
Suggestions
Define what 'challenges' means in this context - specify the concrete actions the skill performs (e.g., 'Creates coding challenges', 'Solves puzzle problems', 'Manages competitive programming tasks')
Add an explicit 'Use when...' clause with natural trigger terms users would actually say when needing this skill
Include domain-specific keywords and file types/formats if applicable to distinguish this skill from others
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | The description provides no concrete actions whatsoever. 'Agent skill for challenges' is completely abstract with no indication of what the skill actually does. | 1 / 3 |
Completeness | The description fails to answer both 'what does this do' and 'when should Claude use it'. There is no 'Use when...' clause and no explanation of capabilities. | 1 / 3 |
Trigger Term Quality | The only potential trigger term is 'challenges' which is extremely generic and could apply to countless contexts. '$agent-challenges' is a command syntax, not a natural user phrase. | 1 / 3 |
Distinctiveness Conflict Risk | 'Challenges' is extremely vague and could conflict with any skill involving problem-solving, debugging, troubleshooting, or competitive tasks. No clear niche is established. | 1 / 3 |
Total | 4 / 12 Passed |
Implementation
14%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill reads more like a marketing document or persona description than actionable technical guidance. It spends most of its tokens explaining what the agent does conceptually rather than how to actually accomplish tasks. The MCP tool examples are a good start but lack the concrete workflows, validation steps, and error handling needed for reliable execution.
Suggestions
Remove explanatory sections about what challenges, leaderboards, and gamification are - Claude already knows these concepts. Focus only on Flow Nexus-specific implementation details.
Add concrete workflows with validation steps, e.g., 'To submit a solution: 1. Validate syntax locally, 2. Call challenge_submit, 3. Check response.status, 4. If failed, parse response.errors for feedback'
Include example responses from MCP tools showing expected data structures and how to handle success/failure cases
Split detailed category descriptions and gamification features into separate reference files, keeping SKILL.md as a concise quick-start guide
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Extremely verbose with extensive explanations of concepts Claude already knows (what gamification is, what algorithms are, what leaderboards do). Lists like 'challenge categories you manage' and 'gamification features you leverage' explain obvious concepts rather than providing actionable guidance. | 1 / 3 |
Actionability | Provides some concrete MCP tool calls with example parameters, which is useful. However, the code examples are incomplete (placeholder IDs, pseudocode-style solution_code), and there's no guidance on actual implementation, error handling, or expected responses. | 2 / 3 |
Workflow Clarity | The 'challenge curation approach' lists abstract steps like 'Skill Assessment' and 'Challenge Selection' without concrete sequences, validation checkpoints, or feedback loops. No guidance on what to do when submissions fail or how to handle edge cases. | 1 / 3 |
Progressive Disclosure | Monolithic wall of text with no references to external documentation. All content is inline with no structure for discovery. Categories, features, and standards are all dumped in one file without clear navigation or separation of concerns. | 1 / 3 |
Total | 5 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.