Detects logical and semantic bugs by understanding program intent — catches issues that syntax-only tools miss. Use when static analysis has already run and found nothing, when the user reports incorrect behavior but no crash, or when reviewing algorithmic code for correctness.
Install with Tessl CLI
npx tessl i github:santosomar/general-secure-coding-agent-skills --skill semantic-bug-detector96
Quality
95%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Discovery
89%Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.
This is a strong description that clearly differentiates itself from other code analysis tools by focusing on semantic/logical bugs that syntax tools miss. It excels at completeness with explicit 'Use when' triggers and has distinctive positioning. The main weakness is that the capabilities could be more concrete with specific actions rather than general bug detection.
| Dimension | Reasoning | Score |
|---|---|---|
Specificity | Names the domain (logical/semantic bugs, program intent) and describes the general action (detects bugs, catches issues), but doesn't list multiple specific concrete actions like 'trace variable states', 'verify loop invariants', or 'check boundary conditions'. | 2 / 3 |
Completeness | Clearly answers both what ('Detects logical and semantic bugs by understanding program intent') AND when ('Use when static analysis has already run and found nothing, when the user reports incorrect behavior but no crash, or when reviewing algorithmic code for correctness'). | 3 / 3 |
Trigger Term Quality | Includes natural keywords users would say: 'logical bugs', 'semantic bugs', 'incorrect behavior', 'no crash', 'algorithmic code', 'correctness', 'static analysis'. These cover common variations of how users describe this problem. | 3 / 3 |
Distinctiveness Conflict Risk | Carves out a clear niche distinct from syntax checkers, linters, and static analysis tools by explicitly positioning itself for cases where those tools fail. The triggers ('static analysis found nothing', 'incorrect behavior but no crash') are highly specific. | 3 / 3 |
Total | 11 / 12 Passed |
Implementation
100%Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.
This is an excellent skill that teaches semantic bug detection through concrete patterns, a detailed signal catalog, and a thorough worked example that demonstrates the full detection-to-confirmation workflow. It respects Claude's intelligence by focusing on domain-specific knowledge (the smell patterns, intent hierarchy, FP suppression heuristics) rather than explaining basic concepts. The ranking system and confidence levels show sophisticated understanding of the uncertainty inherent in semantic analysis.
| Dimension | Reasoning | Score |
|---|---|---|
Conciseness | Every section earns its place—no explanation of what bugs are or how Python works. The signal catalog table is dense with actionable patterns, and the worked example demonstrates rather than explains. | 3 / 3 |
Actionability | Provides a concrete signal catalog with specific patterns to look for, a complete worked example with real code and test confirmation, and a clear output format. The intent hierarchy and FP suppression rules are immediately applicable. | 3 / 3 |
Workflow Clarity | Clear sequence: establish intent → detect smells → rank by confirmability → suppress FPs → output findings. The worked example demonstrates the full workflow including the critical validation step of writing a confirming test. | 3 / 3 |
Progressive Disclosure | Well-organized with clear sections (intent question, signal catalog, ranking, FP suppression, worked example, edge cases, do not, output format). Appropriate length for a single file with no need for external references. | 3 / 3 |
Total | 12 / 12 Passed |
Validation
100%Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.
Validation — 11 / 11 Passed
Validation for skill structure
No warnings or errors.
Table of Contents
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.