This skill enables Claude to design and execute chaos engineering experiments to test system resilience. It is used when the user requests help with failure injection, latency simulation, resource exhaustion testing, or resilience validation. The skill is triggered by discussions of chaos experiments (GameDays), failure injection strategies, resilience testing, and validation of recovery mechanisms like circuit breakers and retry logic. It leverages tools like Chaos Mesh, Gremlin, Toxiproxy, and AWS FIS to simulate real-world failures and assess system behavior.
Install with Tessl CLI
npx tessl i github:jeremylongshore/claude-code-plugins-plus-skills --skill conducting-chaos-engineering68
Does it follow best practices?
If you maintain this skill, you can automatically optimize it using the tessl CLI to improve its score:
npx tessl skill review --optimize ./path/to/skillValidation for skill structure
Discovery
100%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a strong skill description that excels across all dimensions. It provides specific concrete actions, comprehensive trigger terms that users in this domain would naturally use, explicit 'when to use' guidance, and occupies a clearly distinct niche in chaos engineering. The description uses proper third-person voice throughout.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Lists multiple specific concrete actions: 'design and execute chaos engineering experiments', 'failure injection', 'latency simulation', 'resource exhaustion testing', 'resilience validation', and mentions specific tools like Chaos Mesh, Gremlin, Toxiproxy, and AWS FIS. | 3 / 3 |
Completeness | Clearly answers both what ('design and execute chaos engineering experiments to test system resilience') and when ('when the user requests help with failure injection, latency simulation...', 'triggered by discussions of chaos experiments'). Explicit trigger guidance is provided. | 3 / 3 |
Trigger Term Quality | Excellent coverage of natural terms users would say: 'chaos engineering', 'failure injection', 'latency simulation', 'resource exhaustion', 'resilience testing', 'GameDays', 'circuit breakers', 'retry logic', plus specific tool names that users familiar with the domain would mention. | 3 / 3 |
Distinctiveness Conflict Risk | Highly distinctive niche with specific domain terminology (chaos engineering, GameDays, Chaos Mesh, Gremlin) that would not overlap with general testing or infrastructure skills. The combination of chaos-specific tools and failure injection concepts creates a clear, unique trigger profile. | 3 / 3 |
Total | 12 / 12 Passed |
Implementation
20%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This skill content reads more like a marketing description than actionable guidance. It explains what chaos engineering is and what Claude 'will do' rather than providing concrete configurations, commands, or executable examples. The lack of actual Chaos Mesh YAML, Toxiproxy configs, or AWS FIS templates makes this skill nearly unusable for practical chaos engineering work.
Suggestions
Replace abstract descriptions with executable examples: provide actual Chaos Mesh manifests, Toxiproxy configurations, or AWS FIS experiment JSON templates
Remove explanatory content about what chaos engineering is and what the skill 'empowers' - Claude already knows this
Add validation checkpoints to the workflow: how to verify experiments are running correctly, how to abort safely, how to confirm system recovery
Include specific monitoring commands or queries to track during experiments (e.g., Prometheus queries, CloudWatch metrics to watch)
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | The content is verbose and explains concepts Claude already knows (what chaos engineering is, how tools work). Phrases like 'empowers Claude to act as a chaos engineering specialist' and extensive explanations of the process add no value. | 1 / 3 |
Actionability | No concrete code, commands, or executable examples are provided. The examples describe what 'the skill will' do abstractly rather than providing actual Chaos Mesh manifests, Toxiproxy configs, or AWS FIS experiment templates. | 1 / 3 |
Workflow Clarity | The 'How It Works' section provides a high-level sequence (design → tool selection → execution → analysis), but lacks validation checkpoints, specific commands, or feedback loops for error recovery during chaos experiments. | 2 / 3 |
Progressive Disclosure | Content is organized into sections but everything is inline with no references to detailed documentation. Tool-specific configurations and advanced scenarios could be split into separate files but aren't. | 2 / 3 |
Total | 6 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.