github.com/ArabelaTso/Skills-4-SE
Skill | Added | Review |
|---|---|---|
java-regression-test-generator Automatically generate regression tests for Java codebases by analyzing changes between old and new code versions. Use when users need to: (1) Generate tests after refactoring or code changes, (2) Ensure previously tested behavior still works in new versions, (3) Cover modified or newly added code paths, (4) Migrate existing tests to work with updated APIs or signatures, (5) Maintain test coverage during code evolution. Supports JUnit and TestNG frameworks with unit tests, parameterized tests, and exception testing patterns. | 85 1.00x No change in agent success vs baseline Impact 76% 1.00xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
deprecated-api-updater Identify and replace deprecated API usage in source code with modern alternatives. Use when: (1) Modernizing legacy codebases, (2) Upgrading framework versions (React, Django, Spring, etc.), (3) Fixing deprecation warnings in build output, (4) Preparing for major version upgrades, (5) Ensuring code uses current best practices. Supports Python, JavaScript/TypeScript, Java, and other major languages with both AST-based detection and pattern matching for accurate identification and automated replacement with validation. | 93 1.60x Agent success vs baseline Impact 90% 1.60xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
java-test-updater Update Java test classes and methods to work with new code versions after refactoring or modifications. Use when code changes break existing tests due to signature changes, refactoring, or behavior modifications. Takes old and new code versions plus old tests as input, and outputs updated tests that compile and pass against the new code. Handles method signature changes, class refactoring, assertion updates, and mock modifications. | 92 1.00x No change in agent success vs baseline Impact 100% 1.00xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
interval-difference-analyzer Analyze differences in program intervals between two versions of a program (old and new) to identify added, removed, or modified intervals. Use when comparing program versions, analyzing variable ranges, detecting behavioral changes in numeric computations, validating refactorings, or assessing migration impacts. Supports optional test suite integration to validate interval changes. Generates comprehensive reports highlighting intervals requiring further testing or verification. | 80 1.85x Agent success vs baseline Impact 100% 1.85xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
invariant-inference Automatically infer loop invariants for code verification and correctness proofs. Use when analyzing loops to identify properties that hold throughout execution, generating assertions for verification, proving loop correctness, or documenting loop behavior. Supports Python, Java, C/C++, and language-agnostic analysis. Generates invariants as code assertions (assert statements). Triggers when users ask to infer invariants, find loop properties, generate loop assertions, prove loop correctness, or verify loop behavior. | 94 1.12x Agent success vs baseline Impact 97% 1.12xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
library-advisor Recommend relevant Isabelle/HOL or Coq standard library theories, lemmas, and tactics based on proof goals. Use when: (1) Users need library lemmas for their proof, (2) Proof goals match standard library patterns, (3) Users ask what libraries to import, (4) Specific lemmas are needed for list/set/arithmetic operations, (5) Users are stuck and need to know what library support exists, or (6) Guidance on find_theorems/Search commands is needed. Supports both Isabelle/HOL and Coq standard libraries. | 90 1.03x Agent success vs baseline Impact 96% 1.03xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
metamorphic-property-extractor Automatically identify metamorphic properties (symmetry, linearity, additivity, input invariances) from programs or functions. Use when generating metamorphic tests, discovering program properties, validating transformations, or creating test oracles without explicit specifications. Analyzes control flow, data flow, and sample executions to output structured properties for metamorphic test generation and verification. | 85 1.78x Agent success vs baseline Impact 93% 1.78xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
proof-failure-explainer Analyze and explain why Isabelle or Coq proofs fail, identifying the root cause such as type mismatches, missing assumptions, incorrect goals, unification failures, or inapplicable tactics. Use when the user encounters proof failures, error messages in formal verification, stuck proof states, or asks why their Isabelle/Coq proof doesn't work. | 92 1.01x Agent success vs baseline Impact 88% 1.01xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
markdown-document-structurer Reorganizes markdown documents into well-structured, consistent format while preserving content and improving readability. Use when Claude needs to: (1) Fix heading hierarchy issues (skipped levels, multiple h1s), (2) Generate or update table of contents, (3) Standardize formatting (lists, code blocks, emphasis, links), (4) Improve grammar and spelling, (5) Add missing standard sections (installation, usage, etc.), (6) Remove redundant or duplicate content, (7) Restructure technical docs, READMEs, or long-form content for better organization and flow. | 90 1.10x Agent success vs baseline Impact 98% 1.10xAverage score across 3 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0f00a4f | |
incremental-python-programmer Takes a Python repository and natural language feature description as input, implements the feature with proper code placement, generates comprehensive tests, and ensures all tests pass. Use when Claude needs to: (1) Add new features to existing Python projects, (2) Implement functions, classes, or modules based on requirements, (3) Modify existing code to add functionality, (4) Generate unit and integration tests for new code, (5) Fix failing tests after implementation, (6) Ensure code follows existing patterns and conventions. | 69 1.05x Agent success vs baseline Impact 90% 1.05xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
issue-report-generator Automatically generate clear, actionable issue reports from failing tests and repository analysis. Analyze test failures to understand expected vs. actual behavior, identify affected code components, and produce well-structured Markdown reports suitable for GitHub Issues or similar trackers. Use when a test fails, when debugging issues, or when the user asks to create an issue report, generate a bug report, or document a test failure. | 80 1.05x Agent success vs baseline Impact 98% 1.05xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
model-guided-code-repair Automatically repair code violations of temporal properties using model-checking counterexamples as guidance. Use when working with formal verification results, temporal logic property violations (LTL, CTL), model checking counterexamples, or when asked to repair property violations, fix counterexamples, repair temporal properties, fix model checking violations, or repair code based on counterexamples. Applicable to concurrent systems, state machines, synchronization issues, safety/liveness properties, and resource management violations. | 87 1.26x Agent success vs baseline Impact 96% 1.26xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
mutation-test-suite-optimizer Optimize test suites using mutation testing to maximize mutation kill rate with minimal tests. Use when you need to reduce test suite size while maintaining quality, identify redundant tests, improve mutation coverage, or validate test effectiveness. The skill analyzes test coverage, execution intervals, and redundancy using mutation operators, selects or generates a minimal subset of tests that maximizes mutation kill rate, and outputs an optimized test suite with detailed reports showing killed and surviving mutants. Supports multiple programming languages and mutation testing frameworks. | 86 1.96x Agent success vs baseline Impact 100% 1.96xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
bug-reproduction-test-generator Automatically generates executable tests that reproduce reported bugs from issue reports and code repositories. Use when users need to: (1) Create a test that reproduces a bug described in an issue report, (2) Generate failing tests from bug descriptions, stack traces, or error messages, (3) Validate bug reports by creating reproducible test cases, (4) Convert issue reports into executable regression tests. Takes a repository and issue report as input and produces test code that reliably triggers the reported bug. | 90 1.41x Agent success vs baseline Impact 92% 1.41xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
java-api-consistency-validator Validate API consistency between two versions of Java libraries. Use when you need to compare API behavior, signatures, and exceptions between Java library versions to identify breaking changes, incompatible modifications, and behavior differences. The skill performs static analysis of Java code, compares method signatures, class definitions, interface changes, parameter types, return types, exception declarations, and generates a detailed JSON report with breaking changes, warnings, and migration guidance. Supports Java libraries and packages. | 88 1.78x Agent success vs baseline Impact 100% 1.78xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
multi-version-behavior-comparator Compare behavior across multiple versions of programs or repositories. Use when you need to analyze how functionality changes between versions, identify regressions, compare outputs and exceptions, or validate upgrades. The skill compares execution behavior, test results, outputs, exceptions, and observable states across versions, generating detailed reports showing behavioral divergences, potential regressions, added/removed functionality, and areas requiring validation. Supports multiple programming languages and can work with test suites or execution traces. | 84 1.58x Agent success vs baseline Impact 100% 1.58xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
failure-oriented-instrumentation Selectively instruments code to capture runtime data for debugging failures and bugs. Use when investigating crashes, exceptions, unexpected behavior, test failures, or performance issues. Analyzes stack traces and error messages to identify suspicious code regions, then adds targeted logging, tracing, and assertions to capture variable values, execution paths, timing, and conditional branches. Supports Python, JavaScript/TypeScript, Java, and C/C++. | 82 1.10x Agent success vs baseline Impact 66% 1.10xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
flaky-test-detector Identifies non-deterministic or unreliable tests through static code analysis and test result analysis. Use when Claude needs to find flaky tests, analyze test reliability, or investigate intermittent test failures. Supports Python (pytest, unittest) and Java (JUnit, TestNG) test frameworks. Trigger when users mention "flaky tests", "intermittent failures", "non-deterministic tests", "unreliable tests", or ask to "find flaky tests", "analyze test stability", or "why tests fail randomly". | 94 1.04x Agent success vs baseline Impact 97% 1.04xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
environment-setup-assistant Generate setup scripts and instructions for development environments across platforms. Use when: (1) Setting up new development machines (Python, Node.js, Docker, databases), (2) Creating automated setup scripts for team onboarding, (3) Need cross-platform setup instructions (macOS, Linux, Windows), (4) Installing development tools and dependencies, (5) Configuring version managers and package managers. Provides executable setup scripts, platform-specific guides, and tool installation instructions. | 93 1.15x Agent success vs baseline Impact 90% 1.15xAverage score across 3 eval scenarios Securityby Advisory Suggest reviewing before use Reviewed: Version: 0f00a4f | |
fuzzing-input-generator Generate randomized and edge-case inputs to detect unexpected failures, bugs, and security vulnerabilities through fuzz testing. Use when creating test cases for robustness testing, generating adversarial inputs, testing error handling, finding edge cases, or security testing. Produces Python test code with fuzzing inputs for strings, numbers, and structured data focusing on edge cases, invalid inputs, and random valid inputs. Triggers when users ask to generate fuzz tests, create randomized test inputs, test edge cases, find bugs through fuzzing, or generate adversarial test cases. | 89 1.37x Agent success vs baseline Impact 77% 1.37xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
module-component-generator Generates complete modules and components from interface contracts and specifications. Use when Claude needs to build larger software components, implement interfaces, create service layers, or generate complete modules with multiple related classes. Supports Python (with type hints and protocols) and Java (with interfaces and abstract classes). Trigger when users request "implement this interface", "generate a module for", "create a service layer", "build a component that implements", or provide interface definitions that need implementation. | 87 1.31x Agent success vs baseline Impact 95% 1.31xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
integration-test-generator Generate integration tests for multiple interacting components in Python. Use when testing interactions between: (1) Multiple services or APIs (REST/GraphQL endpoints, microservices), (2) Database operations with repositories/ORMs (SQLAlchemy, Django ORM), (3) External services (payment gateways, email services, third-party APIs), (4) Message queues and event-driven systems, (5) Full stack workflows (API + database + business logic). Provides test structure templates, fixtures, test data builders, and patterns for pytest-based integration testing. | 87 1.02x Agent success vs baseline Impact 91% 1.02xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
design-pattern-suggestor Recommends appropriate software design patterns based on problem descriptions, requirements, or code scenarios. Use when designing software architecture, refactoring code, solving common design problems, or choosing between design approaches. Analyzes the problem context and suggests suitable creational, structural, behavioral, architectural, or concurrency patterns with implementation guidance and trade-off analysis. | 90 1.14x Agent success vs baseline Impact 92% 1.14xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-change-summarizer Generates clear and structured pull request descriptions from code changes. Use when Claude needs to: (1) Create PR descriptions from git diffs or code changes, (2) Summarize what changed and why, (3) Document breaking changes with migration guides, (4) Add technical details and design decisions, (5) Provide testing instructions, (6) Enhance descriptions with security, performance, and architecture notes, (7) Document dependency changes. Takes code changes as input, outputs comprehensive PR description in Markdown. | 82 1.15x Agent success vs baseline Impact 95% 1.15xAverage score across 3 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0f00a4f | |
coverage-enhancer Analyze existing test suites and source code to suggest additional unit tests that improve test coverage. Use this skill when working with test files and source code to identify untested code paths, missing edge cases, uncovered branches, untested error conditions, and gaps in test coverage. Supports major testing frameworks (pytest, Jest, JUnit, Go testing, etc.) and generates targeted test suggestions based on coverage analysis. | 84 1.15x Agent success vs baseline Impact 96% 1.15xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f |