CtrlK
CommunityDocumentationLog inGet started
Tessl Logo

agent-orchestration-improve-agent

tessl i github:sickn33/antigravity-awesome-skills --skill agent-orchestration-improve-agent

Systematic improvement of existing agents through performance analysis, prompt engineering, and continuous iteration.

55%

Overall

SKILL.md
Review
Evals

Validation

81%
CriteriaDescriptionResult

description_trigger_hint

Description may be missing an explicit 'when to use' trigger hint (e.g., 'Use when...')

Warning

metadata_version

'metadata' field is not a dictionary

Warning

license_field

'license' field is missing

Warning

Total

13

/

16

Passed

Implementation

55%

This skill provides a comprehensive framework for agent optimization with excellent workflow structure and clear phases, but suffers from verbosity and lack of progressive disclosure. The content explains many concepts Claude already understands and presents everything inline rather than using appropriate file references. Actionability is moderate—guidance is structured but lacks truly executable examples.

Suggestions

Split detailed content (evaluation metrics, test categories, rollback procedures) into separate reference files and link from the main skill with clear navigation

Replace pseudocode tool references like 'Use: context-manager' with actual executable commands or clarify these are conceptual placeholders

Remove explanatory content about well-known concepts (A/B testing basics, semantic versioning) and assume Claude's competence

Add concrete code examples for at least one phase, such as a Python script for metrics collection or a real prompt template

DimensionReasoningScore

Conciseness

The skill contains useful information but is verbose in places, explaining concepts Claude likely knows (e.g., what A/B testing is, basic versioning semantics). The extensive lists and explanatory text could be significantly condensed while preserving actionability.

2 / 3

Actionability

Provides structured guidance with some concrete elements like metric templates and rollback triggers, but lacks executable code. The 'Use: context-manager' and 'Use: prompt-engineer' commands are pseudocode-like references without actual implementation details or real tool invocations.

2 / 3

Workflow Clarity

Clear four-phase workflow with explicit sequencing, validation checkpoints (A/B testing, staged rollout), and feedback loops (rollback procedures with specific triggers). The progression from analysis to deployment is well-structured with clear decision points.

3 / 3

Progressive Disclosure

Monolithic wall of text with no references to external files. All content is inline despite being over 200 lines. Detailed sections like evaluation metrics, test categories, and constitutional principles could be split into separate reference documents.

1 / 3

Total

8

/

12

Passed

Activation

33%

The description identifies a clear domain (agent improvement) but relies on abstract terminology rather than concrete actions. The complete absence of a 'Use when...' clause significantly weakens its utility for skill selection, and the trigger terms used are more technical than what users would naturally say when seeking help with their agents.

Suggestions

Add an explicit 'Use when...' clause with natural trigger terms like 'agent not working', 'improve my agent', 'agent performance issues', 'fix agent behavior', or 'optimize agent responses'.

Replace abstract actions with concrete specifics: instead of 'performance analysis', specify 'analyze agent error patterns, review conversation logs, identify failure modes'.

Include file type or artifact triggers if applicable, such as 'agent configuration files', 'system prompts', or specific agent frameworks the skill supports.

DimensionReasoningScore

Specificity

Names the domain (agent improvement) and some actions (performance analysis, prompt engineering, continuous iteration), but these are somewhat abstract rather than concrete specific actions like 'analyze error logs' or 'rewrite system prompts'.

2 / 3

Completeness

Describes what the skill does but completely lacks a 'Use when...' clause or any explicit trigger guidance. Per rubric guidelines, missing explicit trigger guidance caps completeness at 2, and this is weaker than that threshold.

1 / 3

Trigger Term Quality

Includes some relevant terms like 'agents', 'prompt engineering', and 'performance analysis', but missing common variations users might say like 'fix my agent', 'agent not working', 'improve prompts', 'debug agent', or 'optimize agent'.

2 / 3

Distinctiveness Conflict Risk

Somewhat specific to agent improvement, but 'prompt engineering' could overlap with general prompt writing skills, and 'performance analysis' is generic enough to conflict with debugging or analytics skills.

2 / 3

Total

7

/

12

Passed

Reviewed

Table of Contents

ValidationImplementationActivation

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.