CtrlK
BlogDocsLog inGet started
Tessl Logo

arn-code-review-implementation

This skill should be used when the user says "review implementation", "review the project", "check implementation", "quality review", "validate implementation", "implementation review", or wants a post-execution quality gate to verify that the implementation follows the project's stored code and testing patterns and matches the plan. Reports issues as ERRORS, WARNINGS, INFO with a verdict. Do NOT use this for reviewing PRs (use arn-code-review-pr) or validating plans (use arn-code-review-plan).

86

Quality

83%

Does it follow best practices?

Impact

No eval scenarios have been run

SecuritybySnyk

Risky

Do not use without reviewing

SKILL.md
Quality
Evals
Security

Quality

Discovery

89%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is a strong skill description with excellent trigger term coverage and explicit 'when to use' guidance including negative boundaries that distinguish it from related skills. The main weakness is that the 'what it does' portion could be more specific about the concrete actions performed beyond verifying and reporting. The description effectively uses third person voice and avoids vague language.

DimensionReasoningScore

Specificity

The description mentions verifying implementation against stored code/testing patterns and matching the plan, and reporting issues as ERRORS/WARNINGS/INFO with a verdict. It names the domain and some actions but doesn't list multiple concrete specific actions beyond 'verify' and 'report'.

2 / 3

Completeness

Clearly answers both 'what' (verifies implementation follows stored code/testing patterns, reports issues with severity levels and verdict) and 'when' (explicit trigger phrases listed, plus negative boundaries distinguishing from PR review and plan validation skills).

3 / 3

Trigger Term Quality

Excellent coverage of natural trigger terms: 'review implementation', 'review the project', 'check implementation', 'quality review', 'validate implementation', 'implementation review', and 'post-execution quality gate'. These are phrases users would naturally say.

3 / 3

Distinctiveness Conflict Risk

Highly distinctive with explicit negative boundaries ('Do NOT use this for reviewing PRs... or validating plans...') and specific trigger phrases. The distinction from related skills (arn-code-review-pr, arn-code-review-plan) is clearly drawn, minimizing conflict risk.

3 / 3

Total

11

/

12

Passed

Implementation

77%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is a well-structured, highly actionable implementation review skill with clear workflows, explicit check IDs, severity classifications, and a well-defined report format. Its main weakness is length — the document packs a lot of conditional workflows (visual regression layers, cross-layer comparison, sketch promotion) inline that could benefit from being split into reference files. The content is slightly verbose in framing but the operational substance is strong.

Suggestions

Extract Steps 3b (Visual Regression), 3c (Cross-Layer Comparison), and 3d (Sketch Promotion) into separate reference files to reduce the main SKILL.md length and improve progressive disclosure.

Remove explanatory framing sentences like 'This is a self-contained review -- it does not invoke sub-agents' and 'This is NOT a hardcoded checklist' that describe the skill's nature rather than instruct Claude on what to do.

DimensionReasoningScore

Conciseness

The skill is thorough but verbose in places. The pipeline position diagram and some explanatory text (e.g., 'This is a self-contained review -- it does not invoke sub-agents') add tokens without much value for Claude. The check tables are efficient, but the overall document is quite long with some redundant framing. The dynamic pattern compliance section repeats the same structure (read, identify, check, generate findings) five times with minor variations that could be templated.

2 / 3

Actionability

The skill provides highly concrete, structured guidance: specific check IDs with severity levels, exact file paths to read, precise report format with template, clear verdict logic with thresholds, and specific error handling for each failure mode. Every step tells Claude exactly what to do and what to produce.

3 / 3

Workflow Clarity

The workflow is clearly sequenced (Steps 1-5) with explicit conditional logic (skip conditions for visual regression, cross-layer, sketch promotion, single-phase projects), validation checkpoints embedded in the check tables, and a clear feedback loop at the end (NEEDS FIXES -> offer to help fix). Error handling section covers all failure modes with specific recovery actions.

3 / 3

Progressive Disclosure

The skill references a bundle file (`cross-layer-comparison-guide.md`) which is good progressive disclosure, but the main SKILL.md is very long (~250+ lines of detailed content) with sections like the full visual regression multi-layer workflow and cross-layer comparison that could be split into reference files. The check tables are well-structured inline content, but Steps 3b-3d add significant bulk that could be externalized.

2 / 3

Total

10

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

frontmatter_unknown_keys

Unknown frontmatter key(s) found; consider removing or moving to metadata

Warning

Total

10

/

11

Passed

Repository
AppsVortex/arness
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.