github.com/ArabelaTso/Skills-4-SE
Skill | Added | Review |
|---|---|---|
edge-case-generator Automatically identify potential boundary and exception cases from requirements, specifications, or existing code, and generate comprehensive test cases targeting boundary conditions, edge cases, and uncommon scenarios. Use this skill when analyzing programs, code repositories, functions, or APIs to discover and test corner cases, null handling, overflow conditions, empty inputs, concurrent access patterns, and other exceptional scenarios that are often missed in standard testing. | 93 1.03x Agent success vs baseline Impact 96% 1.03xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
cpp-to-dafny-translator Translate C/C++ programs to equivalent Dafny code while preserving semantics and ensuring verification. Use when users ask to convert, translate, or port C/C++ code to Dafny, or when they need to formally verify C/C++ algorithms using Dafny's verification capabilities. Handles functions, structs, pointers, arrays, memory management, and ensures the generated Dafny code is well-typed, executable, verifiable, and can successfully run. | 90 1.00x No change in agent success vs baseline Impact 96% 1.00xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
design-smell-detector Identify design quality issues in code including high coupling, low cohesion, God classes, long methods, and other code smells. Use when: (1) Reviewing code architecture and design quality, (2) Identifying refactoring opportunities, (3) Detecting God classes or classes with too many responsibilities, (4) Finding high coupling or low cohesion issues, (5) Analyzing code maintainability and technical debt. Detects coupling smells, cohesion problems, complexity issues, size violations, and encapsulation problems with actionable refactoring suggestions. | 90 1.05x Agent success vs baseline Impact 79% 1.05xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
critical-interval-security-checker Analyzes code to identify security-critical time intervals and timing vulnerabilities in authentication, authorization, and time-sensitive security operations. Use this skill when reviewing code for proper timeout enforcement, token expiration, session management, rate limiting, password reset validity, or any time-sensitive security mechanism. Detects missing expiration checks, excessive timeout values, lack of rate limiting, client-side only validation, hardcoded timeouts, and timing attack vulnerabilities. Triggers when users ask to check security timeouts, verify token expiration handling, audit session timeout implementation, review rate limiting, or analyze time-based security controls. | 95 1.17x Agent success vs baseline Impact 99% 1.17xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
error-explanation-generator Explains test failures and provides actionable debugging guidance. Use when tests fail (unit, integration, E2E), builds fail, or code throws errors. Analyzes error messages, stack traces, and test output to identify root causes and suggest concrete fixes. Handles pytest, jest, junit, mocha, vitest, selenium, cypress, playwright, and other testing frameworks across Python, JavaScript/TypeScript, Java, Go, and other languages. | 91 1.25x Agent success vs baseline Impact 88% 1.25xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
dead-code-eliminator Identify and analyze unused or redundant code including unused functions/methods, unused variables/imports, unreachable code, and redundant conditions. Use when cleaning up codebases, improving maintainability, reducing technical debt, or conducting code quality audits. Analyzes Python code using AST analysis and produces markdown reports listing dead code locations with line numbers, severity ratings, and recommendations. Triggers when users ask to find dead code, remove unused code, identify unused imports, find unreachable code, or clean up redundant logic. | 81 2.06x Agent success vs baseline Impact 93% 2.06xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
counterexample-to-test-generator Automatically generates executable test cases from model checking counterexample traces. Translates abstract counterexample states and transitions into concrete test inputs, execution steps, and assertions that reproduce property violations. Use when working with model checker outputs (SPIN, CBMC, NuSMV, TLA+, Java PathFinder, etc.) and needing to create regression tests, validate bug fixes, or reproduce verification failures in executable test suites. | 90 1.16x Agent success vs baseline Impact 99% 1.16xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
directed-test-input-generator Generate targeted test inputs to reach specific code paths and hard-to-reach behaviors in Python code. Use when: (1) Targeting uncovered branches or specific execution paths, (2) Need coverage-guided test generation, (3) Want to leverage LLM understanding of code semantics for meaningful test inputs, (4) Testing boundary conditions and edge cases systematically, (5) Combining symbolic reasoning with fuzzing. Provides path analysis, constraint solving, coverage-guided strategies, and LLM-driven semantic generation for comprehensive test input creation. | 91 1.28x Agent success vs baseline Impact 82% 1.28xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
imperative-to-coq-model-extractor Extract abstract mathematical models from imperative code (C, C++, Python, Java, etc.) suitable for formal reasoning in Coq. Use when the user asks to model imperative code in Coq, create Coq specifications from imperative programs, extract mathematical models for verification, or translate imperative algorithms to Coq for formal reasoning and proof. | 88 1.03x Agent success vs baseline Impact 99% 1.03xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
counterexample-generator Generate concrete counterexamples when formal verification, assertions, or specifications fail. Use this skill when debugging failed proofs, understanding why verification fails, creating minimal reproducing examples, analyzing assertion violations, investigating invariant breaks, or diagnosing specification mismatches. Produces concrete input values, execution traces, and state information that demonstrate the failure. | 78 1.10x Agent success vs baseline Impact 85% 1.10xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-completion-semantic-constraints Automatically complete partial code snippets while satisfying semantic constraints including variable types, invariants, pre/post-conditions, interface contracts, and expected input/output behavior. Use when users provide incomplete code with specific requirements like "complete this function that takes a list and returns sorted unique elements" or "fill in this method body that must maintain the invariant that x stays positive" or "implement this interface method with these type constraints." Produces compilable, executable code with tests and a constraint satisfaction report. | 87 1.02x Agent success vs baseline Impact 95% 1.02xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
ci-pipeline-synthesizer Generate GitHub Actions CI/CD pipeline configurations for automated building and testing of library and package projects. Use when creating or updating CI workflows for npm packages, Python packages, Go modules, Rust crates, or other library projects that need automated build and test pipelines. Includes templates for common package ecosystems with best practices for dependency caching, matrix testing, and artifact publishing. | 87 1.11x Agent success vs baseline Impact 98% 1.11xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-summarizer Generate concise summaries of source code at multiple scales. Use when users ask to summarize, explain, or understand code - whether it's a single function, a class, a module, or an entire codebase. Handles function-level code by explaining intention and core logic, and large codebases by providing high-level overviews with drill-down capabilities for specific modules. | 84 1.11x Agent success vs baseline Impact 94% 1.11xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-review-assistant Conduct comprehensive code reviews identifying bugs, security issues, performance problems, code quality concerns, and best practice violations. Use when reviewing pull requests, examining code changes, evaluating new code, assessing code quality, or providing feedback on implementations. Analyzes code for correctness, security vulnerabilities, performance bottlenecks, maintainability issues, test coverage, documentation quality, and adherence to coding standards. Produces structured markdown reviews with categorized findings, severity ratings, specific examples, and actionable recommendations. Triggers when users ask to review code, check pull requests, evaluate implementations, find bugs, or assess code quality. | 84 1.25x Agent success vs baseline Impact 83% 1.25xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-pattern-extractor Analyze codebases to identify reusable code patterns, duplications, and implementation patterns for future development. Use when refactoring code, identifying technical debt, finding opportunities for abstraction, or documenting common patterns in a directory or module. Outputs pattern catalogs, refactoring suggestions, and reusable template code. | 76 1.40x Agent success vs baseline Impact 76% 1.40xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
containerization-assistant Generate Dockerfiles, Docker Compose configurations, and Kubernetes manifests for containerizing applications. Use when: (1) Creating Dockerfiles for Node.js, Python, Java, Go, or other applications, (2) Setting up multi-service environments with Docker Compose, (3) Generating Kubernetes deployments, services, and ingress configurations, (4) Optimizing container images for production, (5) Implementing containerization best practices. Provides both ready-to-use templates and custom-generated configurations based on project requirements. | 88 1.06x Agent success vs baseline Impact 92% 1.06xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
bug-to-patch-generator Generate code fixes and patches from bug reports, failing test cases, error messages, and stack traces. Use this skill when debugging code, fixing test failures, addressing GitHub issues, resolving runtime errors, or patching security vulnerabilities. Analyzes the bug context, identifies root causes, and generates precise code patches with explanations and validation steps. | 80 1.33x Agent success vs baseline Impact 100% 1.33xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
configuration-generator Generate configuration files for applications, services, and infrastructure. Use when: (1) Setting up new projects (package.json, requirements.txt, tsconfig.json), (2) Creating Docker or Kubernetes configurations, (3) Configuring CI/CD pipelines (GitHub Actions, GitLab CI, CircleCI), (4) Setting up web servers (Nginx, Apache), (5) Defining infrastructure as code (Terraform, CloudFormation), (6) Generating linter/formatter configs (ESLint, Prettier, Black). Provides templates and custom-generated configs for diverse tech stacks. | 93 1.24x Agent success vs baseline Impact 97% 1.24xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-smell-detector Identify and report code smells indicating poor design or maintainability issues in Python code, including duplicate code, magic numbers, hardcoded values, God classes, feature envy, inappropriate intimacy, data clumps, primitive obsession, and long parameter lists. Use when conducting code quality audits, preparing for refactoring, improving codebase maintainability, or performing design reviews. Produces markdown reports with severity ratings, locations, descriptions, and specific refactoring recommendations with before/after examples. Triggers when users ask to find code smells, identify design issues, suggest refactorings, improve code quality, or detect maintainability problems. | 89 1.43x Agent success vs baseline Impact 93% 1.43xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-repair-generation-combo Automatically repair buggy code and generate comprehensive tests for Python, Java, and C++ programs. Use when users need to fix logic errors or runtime errors in functions, modules, or repositories. Accepts specifications via natural language descriptions, existing test cases, or input/output examples. Generates corrected code, creates or updates tests to verify correctness and prevent regressions, and produces a detailed report explaining the bug, fix, and testing strategy. Triggers on requests like "fix this bug", "repair this code", "debug this function", or "this code is broken". | 89 1.14x Agent success vs baseline Impact 95% 1.14xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
counterexample-explainer Explain why counterexamples violate specifications by analyzing formal specifications (temporal logic, invariants, pre/postconditions, code contracts), informal requirements (user stories, acceptance criteria), test specifications (assertions, property-based tests), and providing step-by-step traces showing state changes, comparing expected vs actual behavior, identifying root causes, and assessing violation impact. Use when debugging test failures, understanding model checker output, explaining runtime assertion violations, analyzing static analysis warnings, or teaching specification concepts. Produces structured markdown explanations with traces, comparisons, state diagrams, and cause chains. Triggers when users ask why something failed, explain a violation, understand a counterexample, debug a specification, or analyze why a test fails. | 93 1.45x Agent success vs baseline Impact 96% 1.45xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
build-ci-migration-assistant Automatically migrates build systems and CI/CD configurations to target platforms. Use when modernizing build infrastructure, switching CI/CD providers, or standardizing across projects. Supports common migration paths including Maven↔Gradle, npm↔Yarn, Travis CI→GitHub Actions, CircleCI→GitHub Actions, Jenkins→GitLab CI, and GitLab CI→GitHub Actions. Analyzes existing configuration, generates equivalent target configuration, maps dependencies and commands, and provides validation and migration documentation. | 95 1.00x No change in agent success vs baseline Impact 100% 1.00xAverage score across 3 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0f00a4f | |
component-boundary-identifier Identifies boundaries between modules or components in software systems through static code analysis and dependency detection. Use when Claude needs to analyze software architecture, identify module boundaries, detect boundary violations, find circular dependencies, or assess component coupling. Supports Python (packages and imports) and Java (packages and dependencies). Trigger when users ask to "identify boundaries", "find component boundaries", "detect boundary violations", "analyze module structure", "check architecture", or "find circular dependencies". | 90 1.20x Agent success vs baseline Impact 96% 1.20xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
bisect-aware-instrumentation Instrument code to support efficient git bisect by producing deterministic pass/fail signals and concise runtime summaries for each tested commit. Use when debugging regressions with git bisect, automating bisect workflows, creating bisect test scripts, handling flaky tests during bisection, or needing clear exit codes and logging for automated bisect runs. Helps identify the exact commit that introduced a bug through automated testing. | 92 1.30x Agent success vs baseline Impact 86% 1.30xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f | |
code-refactoring-assistant Suggest and apply code refactorings to improve readability, maintainability, and code quality. Use this skill when improving existing code structure, eliminating code smells, applying design patterns, simplifying complex logic, extracting duplicated code, renaming for clarity, or preparing code for new features. Provides specific before/after examples, explains benefits, identifies risks, and ensures behavior preservation through tests. | 76 1.26x Agent success vs baseline Impact 91% 1.26xAverage score across 3 eval scenarios Securityby Passed No known issues Reviewed: Version: 0f00a4f |