CtrlK
BlogDocsLog inGet started
Tessl Logo

prompt-injection

Audit applications for AI prompt injection, agent security, and LLM permission boundary vulnerabilities. Use when the user mentions 'prompt injection,' 'LLM security,' 'AI security,' 'jailbreak,' 'indirect prompt injection,' 'prompt leaking,' 'AI red team,' 'LLM vulnerabilities,' 'AI input validation,' 'system prompt extraction,' 'agent security,' 'MCP security,' 'AI permissions,' 'AI privilege escalation,' or needs to secure any application with AI features, AI agents, or LLM integrations.

68

Quality

83%

Does it follow best practices?

Impact

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

SKILL.md
Quality
Evals
Security

Quality

Discovery

89%

Based on the skill's description, can an agent find and select it at the right time? Clear, specific descriptions lead to better discovery.

This is a strong skill description with excellent trigger term coverage and completeness, clearly answering both what the skill does and when to use it. The main weakness is that the 'what' portion could be more specific by listing multiple concrete audit actions rather than a single high-level 'audit applications for...' statement. Overall, it would perform very well in a multi-skill selection scenario due to its distinctive domain and comprehensive trigger terms.

Suggestions

Expand the capability description with more specific concrete actions, e.g., 'Audit applications for AI prompt injection, test system prompt isolation, review input validation and output filtering, assess agent permission boundaries, and evaluate LLM integration security.'

DimensionReasoningScore

Specificity

The description names the domain (AI/LLM security) and a general action ('Audit applications for...vulnerabilities'), but doesn't list multiple specific concrete actions like 'test for prompt injection vectors, review system prompt isolation, validate input sanitization, assess agent permission boundaries.' The action is singular and somewhat high-level.

2 / 3

Completeness

Clearly answers both 'what' (audit applications for AI prompt injection, agent security, and LLM permission boundary vulnerabilities) and 'when' (explicit 'Use when...' clause with extensive trigger terms and broader conditions like 'needs to secure any application with AI features, AI agents, or LLM integrations').

3 / 3

Trigger Term Quality

Excellent coverage of natural trigger terms users would say: 'prompt injection,' 'LLM security,' 'AI security,' 'jailbreak,' 'indirect prompt injection,' 'prompt leaking,' 'AI red team,' 'LLM vulnerabilities,' 'system prompt extraction,' 'agent security,' 'MCP security,' 'AI permissions,' 'AI privilege escalation.' These are comprehensive and cover many natural variations.

3 / 3

Distinctiveness Conflict Risk

Highly distinctive niche focused specifically on AI/LLM security auditing. The trigger terms are very specific to this domain (prompt injection, jailbreak, system prompt extraction, MCP security) and unlikely to conflict with general security or general AI skills.

3 / 3

Total

11

/

12

Passed

Implementation

77%

Reviews the quality of instructions and guidance provided to agents. Good implementation is clear, handles edge cases, and produces reliable results.

This is a strong, highly actionable security audit skill with excellent workflow structure and concrete code examples covering a wide range of AI/LLM vulnerability classes. Its main weakness is length — it's a monolithic document that could benefit from splitting detailed checklists and agent-specific patterns into separate reference files. Some background explanation is unnecessary for Claude but the domain-specific content largely earns its place.

Suggestions

Split the detailed checklists (permission check table, defense assessment table) and agent-specific audit patterns (Step 4) into separate reference files to improve progressive disclosure and reduce the main file's token footprint.

Remove or significantly condense the Background section — Claude already understands prompt injection concepts; a single sentence linking to OWASP LLM01 would suffice.

DimensionReasoningScore

Conciseness

The skill is comprehensive but includes some explanatory content Claude already knows (e.g., defining what prompt injection is, explaining the three attack classes at a conceptual level). The background section and some inline comments could be trimmed. However, the checklists and tables are efficient formats, and most content is domain-specific enough to justify inclusion.

2 / 3

Actionability

The skill provides concrete, executable code examples for both vulnerable and remediated patterns across Python and JSX. Grep patterns for discovery, specific file/line-level audit steps, validation checklists, and a detailed output report template make this highly actionable and copy-paste ready.

3 / 3

Workflow Clarity

The 7-step methodology is clearly sequenced from attack surface mapping through defense assessment. Each step builds on the previous one logically. Validation checkpoints are embedded (e.g., tool validation gates, permission check checklists, defense assessment tables). The output format enforces structured reporting with severity and remediation prioritization.

3 / 3

Progressive Disclosure

The skill is a monolithic document at ~300+ lines with no bundle files or references to supplementary materials. Content like the full permission check checklist, defense assessment table, and detailed agent security patterns could be split into separate reference files. The references section lists external resources but doesn't link to any bundled supporting files.

2 / 3

Total

10

/

12

Passed

Validation

90%

Checks the skill against the spec for correct structure and formatting. All validation checks must pass before discovery and implementation can be scored.

Validation10 / 11 Passed

Validation for skill structure

CriteriaDescriptionResult

allowed_tools_field

'allowed-tools' contains unusual tool name(s)

Warning

Total

10

/

11

Passed

Repository
briiirussell/cybersecurity-skills
Reviewed

Table of Contents

Is this your skill?

If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.