Enforce universal standards for high-performance development. Use when profiling bottlenecks, reducing latency, fixing memory leaks, improving throughput, or optimizing algorithm complexity in any language.
68
62%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./.github/skills/common/common-performance-engineering/SKILL.mdQuality
Discovery
82%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
The description is reasonably well-structured with a clear 'Use when...' clause and good trigger terms that users would naturally use when seeking performance help. Its main weaknesses are that the 'what' portion is somewhat vague ('enforce universal standards' is abstract) and the scope is extremely broad ('any language'), which could create overlap with more specific optimization or code quality skills.
Suggestions
Replace 'Enforce universal standards for high-performance development' with more concrete actions, e.g., 'Identifies performance bottlenecks, recommends optimization strategies, and refactors code for reduced latency and memory usage.'
Consider narrowing scope or adding more distinctive framing to reduce potential conflict with language-specific optimization skills or general code review skills.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | The description names a domain (performance optimization) and lists several actions like 'profiling bottlenecks, reducing latency, fixing memory leaks, improving throughput, optimizing algorithm complexity,' but these read more like categories of concern than concrete, specific actions (e.g., no mention of specific techniques, tools, or outputs). | 2 / 3 |
Completeness | Clearly answers both 'what' (enforce universal standards for high-performance development) and 'when' with an explicit 'Use when...' clause listing five specific trigger scenarios. | 3 / 3 |
Trigger Term Quality | Includes strong natural trigger terms users would actually say: 'profiling', 'bottlenecks', 'latency', 'memory leaks', 'throughput', 'algorithm complexity', and 'optimizing'. These cover a good range of how users naturally describe performance problems. | 3 / 3 |
Distinctiveness Conflict Risk | While performance optimization is a recognizable niche, the phrase 'universal standards' and 'in any language' makes it very broad. It could overlap with language-specific optimization skills or general code review skills that also touch on performance. | 2 / 3 |
Total | 10 / 12 Passed |
Implementation
42%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
The skill provides a well-organized overview of performance engineering principles with good progressive disclosure to reference files, but critically lacks actionable, executable content. It reads more like a checklist of general best practices than a skill that teaches Claude something new — most of these concepts (avoid O(n²), use async I/O, memoize pure functions) are things Claude already knows. The absence of any concrete code examples, specific profiling tool commands, or language-specific patterns significantly limits its utility.
Suggestions
Add concrete, executable code examples for key patterns — e.g., a profiling command (`py-spy record -o profile.svg -- python app.py`), a memoization implementation, or a batching pattern in at least one language.
Replace generic advice Claude already knows (e.g., 'Aim for O(1) or O(n)') with specific, novel guidance such as particular tool configurations, project-specific thresholds, or decision trees for choosing between optimization strategies.
Add explicit validation criteria to the workflow — e.g., 'Confirm latency improved by ≥10%' or 'Run benchmark suite and compare against baseline numbers before merging'.
Include at least one complete before/after example showing a real bottleneck identification and fix cycle to make the workflow concrete.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Mostly efficient but includes some unnecessary explanations Claude already knows (e.g., explaining what SLIs/SLOs stand for, what tree shaking is). Some bullet points are generic advice rather than novel, actionable guidance specific to this skill. | 2 / 3 |
Actionability | The skill is almost entirely abstract guidance with no concrete code, commands, or executable examples. Statements like 'Memoize pure, expensive functions' and 'Use efficient serialization' describe rather than instruct — there are no copy-paste ready snippets, profiling commands, or specific tool invocations. | 1 / 3 |
Workflow Clarity | The 4-step workflow (Baseline → Identify → Fix → Verify) provides a clear sequence with an implicit feedback loop, but lacks explicit validation checkpoints, specific profiling commands/tools, and concrete criteria for what constitutes a regression or sufficient improvement. | 2 / 3 |
Progressive Disclosure | The skill is well-structured as an overview with clear sections and appropriately references a separate implementation patterns file for detailed examples. Navigation is one level deep and clearly signaled. | 3 / 3 |
Total | 8 / 12 Passed |
Validation
81%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 9 / 11 Passed
Validation for skill structure
| Criteria | Description | Result |
|---|---|---|
metadata_version | 'metadata.version' is missing | Warning |
metadata_field | 'metadata' should map string keys to string values | Warning |
Total | 9 / 11 Passed | |
4c72e76
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.