CtrlK
BlogDocsLog inGet started
Tessl Logo

arn-spark-clickable-prototype

This skill should be used when the user says "clickable prototype", "arn clickable prototype", "interactive prototype", "test interactions", "validate UX", "user journeys", "test navigation", "make it clickable", "prototype interactions", "test the prototype", "build the screens", "create the UI", "screen mockups", or wants to generate a clickable interactive prototype with linked screens and validate it through iterative build-review cycles with Playwright-based interaction testing, per-criterion scoring, an independent judge verdict, and versioned output.

74

Quality

68%

Does it follow best practices?

Impact

No eval scenarios have been run

SecuritybySnyk

Advisory

Suggest reviewing before use

Optimize this skill with Tessl

npx tessl skill review --optimize ./plugins/arn-spark/skills/arn-spark-clickable-prototype/SKILL.md
SKILL.md
Quality
Evals
Security

Quality

Discovery

89%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

The description excels at trigger term coverage and completeness, providing an extensive list of natural user phrases and clearly stating both what the skill does and when to use it. However, the description reads as a single long sentence that front-loads trigger terms in a list format rather than naturally describing capabilities, and the 'what it does' portion is somewhat compressed into a dense clause. The structure prioritizes keyword matching over readability.

Suggestions

Restructure to lead with concrete capabilities (e.g., 'Generates clickable interactive prototypes with linked screens. Validates UX through iterative build-review cycles using Playwright-based interaction testing, per-criterion scoring, and an independent judge verdict.') followed by a 'Use when...' clause with the trigger terms.

Use third-person active voice for capability statements (e.g., 'Builds multi-screen clickable prototypes, tests navigation and interactions via Playwright, scores results per criterion') rather than embedding everything in a 'when' clause.

DimensionReasoningScore

Specificity

The description mentions some concrete actions like 'generate a clickable interactive prototype with linked screens', 'Playwright-based interaction testing', 'per-criterion scoring', 'independent judge verdict', and 'versioned output', but these are crammed into a single run-on clause rather than clearly listed as distinct capabilities. The specific actions are present but not well-organized.

2 / 3

Completeness

The description explicitly answers both 'what' (generate clickable interactive prototypes with linked screens, validate through iterative build-review cycles with Playwright testing, per-criterion scoring, judge verdict, versioned output) and 'when' (with an explicit list of trigger phrases prefaced by 'should be used when'). Both dimensions are clearly addressed.

3 / 3

Trigger Term Quality

The description includes an extensive list of natural trigger phrases users would say: 'clickable prototype', 'interactive prototype', 'test interactions', 'validate UX', 'user journeys', 'make it clickable', 'screen mockups', 'create the UI', 'build the screens'. These cover many natural variations a user might use.

3 / 3

Distinctiveness Conflict Risk

The combination of clickable prototypes, Playwright-based interaction testing, per-criterion scoring, and independent judge verdict creates a very distinct niche. This is unlikely to conflict with generic UI design skills or generic testing skills due to its specific workflow description.

3 / 3

Total

11

/

12

Passed

Implementation

47%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is a comprehensive orchestration skill with excellent workflow clarity and thorough error handling, but it suffers significantly from verbosity. The ~400+ line body contains substantial redundancy (agent invocation details repeated in workflow and summary table, error handling duplicating inline fallback descriptions) and over-explains concepts Claude can infer. Actionability is moderate — the procedural steps are clear but lack concrete executable examples like actual CLI commands or script snippets.

Suggestions

Reduce the skill by at least 40% by eliminating redundancy: remove the Agent Invocation Guide table (it restates what's already in the workflow), consolidate error handling into the workflow steps where they occur, and cut explanatory text Claude doesn't need (e.g., explaining what a hub screen is, why screens should be state-specific).

Extract the Prerequisites section into a separate `prerequisites.md` reference file and summarize it in 5-10 lines in the main skill, since the prerequisite detection logic alone is ~80 lines.

Add concrete executable examples: show an actual dev server start command, a sample Playwright script snippet for screen capture, or a real agent invocation call rather than abstract 'invoke agent with...' descriptions.

Move the Error Handling section to a separate reference file (e.g., `error-handling.md`) since most entries duplicate fallback logic already described inline in the workflow steps.

DimensionReasoningScore

Conciseness

This skill is extremely verbose at ~400+ lines. It explains many things Claude already knows (how to kill background processes, how to poll URLs, what a hub screen is), repeats agent invocation details in both the workflow and the Agent Invocation Guide table, and includes extensive conditional logic for every edge case that could be more concisely expressed. The error handling section largely duplicates information already stated inline in the workflow steps.

1 / 3

Actionability

The skill provides a detailed multi-step workflow with specific file paths, agent names, and output locations, which is good. However, it lacks executable code examples — there are no actual commands for starting prototypes, no Playwright script snippets, no concrete agent invocation syntax. The guidance is procedural but relies on abstract descriptions ('invoke the agent with...') rather than copy-paste ready instructions.

2 / 3

Workflow Clarity

The workflow is exceptionally well-sequenced with clear step numbering (1 through 8), explicit validation checkpoints (4c expert review, 4d threshold evaluation, Step 5 judge review), feedback loops (failing criteria feed back into next build cycle), and clear decision points with user confirmation gates. Resume detection and cycle budget management are well-defined.

3 / 3

Progressive Disclosure

The skill references external templates (journey-template.md, review-report-template.md, clickable-prototype-criteria.md, showcase-capture-guide.md) which is good progressive disclosure design, but no bundle files were provided to verify these exist. The main SKILL.md itself is monolithic — the Agent Invocation Guide and Error Handling sections could be separate reference files. The prerequisites section alone is ~80 lines that could be extracted.

2 / 3

Total

8

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Repository
AppsVortex/arness
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.