Test grab system (distance grab, one-hand grab, two-hand grab) against the grab example using the iwsdk CLI.
56
63%
Does it follow best practices?
Impact
—
No eval scenarios have been run
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./.claude/skills/test-grab/SKILL.mdQuality
Discovery
50%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
The description is specific about the concrete actions (testing three types of grabs) and the tool used (iwsdk CLI), giving it a clear niche. However, it completely lacks a 'Use when...' clause, making it unclear when Claude should select this skill. The trigger terms are domain-specific and may not capture all natural user phrasings.
Suggestions
Add an explicit 'Use when...' clause, e.g., 'Use when the user asks to test or validate grab interactions, run grab system tests, or use the iwsdk CLI for grab examples.'
Include natural trigger term variations such as 'test grabs', 'grab interaction testing', 'run grab tests', or 'iwsdk grab example' to improve discoverability.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'distance grab, one-hand grab, two-hand grab' testing against a grab example using a named CLI tool ('iwsdk CLI'). | 3 / 3 |
Completeness | Describes what the skill does (test grab system using iwsdk CLI) but has no explicit 'Use when...' clause or equivalent trigger guidance, which per the rubric should cap completeness at 2, and the 'when' is entirely missing, placing it at 1. | 1 / 3 |
Trigger Term Quality | Includes relevant domain-specific terms like 'grab system', 'distance grab', 'one-hand grab', 'two-hand grab', 'iwsdk CLI', but these are fairly technical and may not cover all natural variations a user might say (e.g., 'test grabs', 'run grab tests', 'VR grab'). | 2 / 3 |
Distinctiveness Conflict Risk | Highly specific niche — testing a grab system with the iwsdk CLI is very distinct and unlikely to conflict with other skills. | 3 / 3 |
Total | 9 / 12 Passed |
Implementation
77%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is a highly actionable and well-structured test procedure with excellent workflow clarity, explicit validation checkpoints, and comprehensive error recovery. Its main weakness is length — the document is quite long for a single SKILL.md and could benefit from splitting reference material into separate files. The MCPCALL helper adds necessary but verbose infrastructure.
Suggestions
Consider extracting the MCPCALL shell function, Component Reference table, and Known Issues into separate referenced files to improve progressive disclosure and reduce the main file's token footprint.
The repeated 'IMPORTANT' notes about running commands one at a time and using sleep could be consolidated into a single 'Execution Rules' section at the top to reduce redundancy.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The skill is quite long (~300+ lines) but most content is necessary test procedures. The MCPCALL helper function is verbose but justified since it's executable infrastructure. Some sections like the 'Known Issues' and 'Component Reference' table are efficient. However, the introductory explanation of the MCPCALL pattern and some repeated instructions ('IMPORTANT' notes) add moderate bloat. | 2 / 3 |
Actionability | Every step provides exact, copy-paste-ready bash commands with specific tool names, JSON arguments, and clear assertions. The placeholder syntax (<distance>, <onehand-pos.x>) is well-defined through prior discovery steps. The component reference table and trigger vs squeeze distinction provide critical operational details. | 3 / 3 |
Workflow Clarity | The workflow is meticulously sequenced across 5 steps with explicit validation at each stage: connectivity verification before tests, snapshot/diff patterns for state validation, sleep commands for timing, and a dedicated recovery section with retry logic. The pre-test setup, entity discovery, and cleanup are all clearly ordered with explicit checkpoints and failure handling (e.g., 'If server fails within 60s, report FAIL for all suites'). | 3 / 3 |
Progressive Disclosure | The content is a monolithic document with no references to external files, which is understandable given no bundle files exist. However, at this length (~300+ lines), the component reference, known issues, and recovery sections could benefit from being split into separate files. The internal organization with clear headers and suites is good but the sheer volume in one file is suboptimal. | 2 / 3 |
Total | 10 / 12 Passed |
Validation
90%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 10 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
frontmatter_unknown_keys | Unknown frontmatter key(s) found; consider removing or moving to metadata | Warning |
Total | 10 / 11 Passed | |
b3d1162
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.