Discover and install skills, docs, and rules to enhance your AI agent's capabilities.
| Name | Contains | Score |
|---|---|---|
sandbox-agent rivet-dev/skills Deploy, configure, and integrate Sandbox Agent - a universal API for orchestrating AI coding agents (Claude Code, Codex, OpenCode, Amp) in sandboxed environments. Use when setting up sandbox-agent server locally or in cloud sandboxes (E2B, Daytona, Docker), creating and managing agent sessions via SDK or API, streaming agent events and handling human-in-the-loop interactions, building chat UIs for coding agents, or understanding the universal schema for agent responses. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Advisory Suggest reviewing before use Reviewed: Version: 48d2a96 |
google-search-browser-use grasseed/google-search-browser-use Use browser-use to perform Google searches, open results, and extract key information from live pages. Use when the user asks to "search Google", "look this up on Google", or needs current web results via a real browser session (often to avoid bot blocks). | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Advisory Suggest reviewing before use Reviewed: Version: d456459 |
strategy-frameworks menkesu/awesome-pm-skills Creates product strategies using Crossing the Chasm, Playing to Win, and strategic canvas frameworks. Use when defining where to play and how to win, choosing beachhead markets, or connecting tactics to strategy. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 53530ef |
metrics-frameworks menkesu/awesome-pm-skills Defines right metrics using North Star framework, AARRR, and leading vs lagging indicators. Use when choosing metrics, instrumenting products, creating dashboards, or distinguishing vanity metrics from actionable ones. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 53530ef |
promptfoo-evaluation fernandezbaptiste/claude-code-skills Configures and runs LLM evaluation using Promptfoo framework. Use when setting up prompt testing, creating evaluation configs (promptfooconfig.yaml), writing Python custom assertions, implementing llm-rubric for LLM-as-judge, or managing few-shot examples in prompts. Triggers on keywords like "promptfoo", "eval", "LLM evaluation", "prompt testing", or "model comparison". | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 4f0eae8 |
video-comparer fernandezbaptiste/claude-code-skills This skill should be used when comparing two videos to analyze compression results or quality differences. Generates interactive HTML reports with quality metrics (PSNR, SSIM) and frame-by-frame visual comparisons. Triggers when users mention "compare videos", "video quality", "compression analysis", "before/after compression", or request quality assessment of compressed videos. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 4f0eae8 |
tdg-personal/crosspost v0.1.0 Multi-platform content distribution across X, LinkedIn, Threads, and Bluesky. Adapts content per platform using content-engine patterns. Never posts identical content cross-platform. Use when the user wants to distribute content across social platforms. Contains: crosspost Multi-platform content distribution across X, LinkedIn, Threads, and Bluesky. Adapts content per platform using content-engine patterns. Never posts identical content cross-platform. Use when the user wants to distribute content across social platforms. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 0.1.0 |
pantheon-ai/pin v0.1.0 Pin session decisions, questions, constraints, and corrections to a persistent board that survives context compaction. Contains: pin Pin session decisions, questions, objections, scope constraints, and corrections to a persistent board that survives context compaction. Use PROACTIVELY when: (1) user approves/rejects a recommendation, (2) user asks a clarifying question about a proposal, (3) user states a scope constraint, (4) user corrects a misunderstanding. Also use when user says pin, track this, mark as approved, board, show pins. Auto-invoke without user asking whenever a decision, question, or constraint is detected. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0.1.0 |
Classify a problem using Cynefin triangulation before acting — routes to the right skill chain (investigate, brainstorm, probe, troubleshoot). Contains: frame-problem Sense-making before action. Classify problem using Cynefin triangulation (3 tests + decomposition) to route to the right skill chain. Use when: frame, what approach, how should I start, which skill, where to begin, unsure what to do. NOT for known tasks — just do them. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 0.1.0 |
jbvc/lint-and-validate v0.1.0 Automatic quality control, linting, and static analysis procedures. Use after every code modification to ensure syntax correctness and project standards. Triggers onKeywords: lint, format, check, validate, types, static analysis. Contains: lint-and-validate Automatic quality control, linting, and static analysis procedures. Use after every code modification to ensure syntax correctness and project standards. Triggers onKeywords: lint, format, check, validate, types, static analysis. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 0.1.0 |
Calibrate research done on socially noisy web sources so agents do not mistake crowd mood for truth. Includes source-specific skills for Moltbook, Hacker News, Reddit, and Product Hunt. Contains: social-source-calibration Route and calibrate research drawn from socially noisy web communities so agents do not mistake crowd mood for truth. Use when research notes, summaries, or quoted material come from Moltbook, Hacker News, Reddit, or Product Hunt and the job is to decide how much weight that material deserves, separate concrete weak signals from vibe/noise, or choose the right source-specific calibration skill before carrying findings forward. hacker-news-source-calibration Calibrate research done on Hacker News so agents do not mistake experienced technical cynicism, anti-hype sentiment, or comment-thread confidence for balanced evidence. Use when summarizing Hacker News reactions, extracting concerns from HN threads, citing HN as part of research, or deciding how much weight to give repeated negative or skeptical Hacker News comments. moltbook-source-calibration Interpret and calibrate already-collected research material derived from Moltbook so agents do not mistake noise, spam, novelty, or social heat for reliable evidence. Use when weighing notes, summaries, or quoted material from Moltbook as part of research, deciding whether a Moltbook claim is worth following up, checking source reliability or evidence quality, or separating concrete weak signals from social-performance noise. Prefer when Moltbook is being used as a weak-signal discovery source rather than as authoritative proof. This skill classifies gathered Moltbook material as concrete report, vibe signal, or noise; flags evidence strength and uncertainty; and suggests whether the claim is follow-up-worthy. This is an informational calibration skill, not a browsing or execution workflow. product-hunt-source-calibration Interpret and calibrate already-collected research material derived from Product Hunt so agents do not mistake launch-day momentum, supportive comments, or leaderboard position for durable product quality or market truth. Use when weighing notes, summaries, reviews, rankings, or quoted material from Product Hunt as part of research, or when deciding how much weight to give Product Hunt launch traction, comments, and maker feedback. This is an informational calibration skill, not a browsing or execution workflow. reddit-source-calibration Calibrate research done on Reddit so agents do not mistake subreddit culture, anecdotal intensity, or comment popularity for reliable evidence. Use when summarizing Reddit threads, extracting recurring user pain from Reddit discussions, citing Reddit as part of research, or deciding how much weight to give repeated subreddit sentiment. | SkillsDocs | 89 1.07x Agent success vs baseline Impact 100% 1.07xAverage score across 1 eval scenario Securityby Passed No known issues Reviewed: Version: 0.1.7 |
Assess your AI fluency using Anthropic's 4D framework (Dakan, Feller & Anthropic, 2025). Scans Claude Code sessions, runs LLM-based behavior classification on all messages, asks a self-assessment questionnaire for 6 unobservable behaviors, and generates a visual HTML report with scores and actionable feedback. Use when "assess fluency", "AI fluency", "fluency report", "fluency assessment", "4D framework", or "how AI fluent am I". Contains: ai-fluency-assessment Assess your AI fluency using Anthropic's 4D framework (Dakan, Feller & Anthropic, 2025). Scans Claude Code sessions, runs LLM-based behavior classification on all messages, asks a self-assessment questionnaire for 6 unobservable behaviors, and generates a visual HTML report with scores and actionable feedback. Use when "assess fluency", "AI fluency", "fluency report", "fluency assessment", "4D framework", or "how AI fluent am I". | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 0.1.3 |
Use when facing 2+ independent tasks that can be worked on without shared state or sequential dependencies Contains: dispatching-parallel-agents Runs multiple independent tasks concurrently by dispatching one focused agent per problem domain, completing batch workloads faster than sequential processing. Use when facing 2+ parallel tasks such as multiple failing test files, broken subsystems, or independent bugs that can be investigated and fixed simultaneously without shared state or sequential dependencies. Also useful when users ask to "run things in parallel", "work on multiple issues at the same time", handle "batch processing", or "multi-task" across independent problem domains. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 0.3.0 |
Query and display available GitHub Copilot AI models with their capabilities, context limits, and features. Use when: "what models are available", "show copilot models", "list github models", "check model capabilities", "switch models". Examples: - user: "What models can I use with GitHub Copilot?" → fetch and display available models - user: "Show me models with vision support" → filter models by capability - user: "Which model has the largest context window?" → compare model specifications - user: "List all GPT-5 models" → filter by model family Contains: github-copilot-models Query and display available GitHub Copilot AI models with their capabilities, context limits, and features. Use when: "what models are available", "show copilot models", "list github models", "check model capabilities", "switch models". Examples: - user: "What models can I use with GitHub Copilot?" → fetch and display available models - user: "Show me models with vision support" → filter models by capability - user: "Which model has the largest context window?" → compare model specifications - user: "List all GPT-5 models" → filter by model family | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Passed No known issues Reviewed: Version: 0.2.0 |
Create, update, and organize Google Calendar events and schedules. Check availability, book time, and manage calendars. Use when asked to schedule a meeting, set up an appointment, book a call, check gcal, or manage calendar events. Contains: google-calendar Create, update, and organize Google Calendar events and schedules. Check availability, book time, and manage calendars. Use when asked to schedule a meeting, set up an appointment, book a call, check gcal, or manage calendar events. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0.3.1 |
odyssey4me/google-sheets v0.1.3 Read, write, and format Google Sheets spreadsheets. Manage cell values, ranges, formulas, pivot tables, and charts. Use when asked to update a gsheet, edit a Google spreadsheet, add formulas, or work with spreadsheet data. Contains: google-sheets Read, write, and format Google Sheets spreadsheets. Manage cell values, ranges, formulas, pivot tables, and charts. Use when asked to update a gsheet, edit a Google spreadsheet, add formulas, or work with spreadsheet data. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0.1.3 |
odyssey4me/google-docs v0.2.1 Create and modify Google Docs documents. Read content, insert tables, apply heading styles, and manage formatting. Use when asked to edit a gdoc, write a Google document, update a doc, or format document content. Contains: google-docs Create and modify Google Docs documents. Read content, insert tables, apply heading styles, and manage formatting. Use when asked to edit a gdoc, write a Google document, update a doc, or format document content. | Skills | 89 Impact — Average score across 0 eval scenarios Securityby Risky Do not use without reviewing Reviewed: Version: 0.2.1 |
A curated collection of Agent Skills for working with dbt, to help AI agents understand and execute dbt workflows more effectively. Contains: creating-mermaid-dbt-dag Generates a Mermaid flowchart diagram of dbt model lineage using MCP tools, manifest.json, or direct code parsing as fallbacks. Use when visualizing dbt model lineage and dependencies as a Mermaid diagram in markdown format. migrating-dbt-core-to-fusion Use when a user needs help triaging dbt-core to Fusion migration errors. Runs dbt-autofix first, then classifies remaining errors into actionable categories (auto-fixable, guided fixes, needs input, blocked). migrating-dbt-project-across-platforms Use when migrating a dbt project from one data platform or data warehouse to another (e.g., Snowflake to Databricks, Databricks to Snowflake) using dbt Fusion's real-time compilation to identify and fix SQL dialect differences. adding-dbt-unit-test Creates unit test YAML definitions that mock upstream model inputs and validate expected outputs. Use when adding unit tests for a dbt model or practicing test-driven development (TDD) in dbt. answering-natural-language-questions-with-dbt Writes and executes SQL queries against the data warehouse using dbt's Semantic Layer or ad-hoc SQL to answer business questions. Use when a user asks about analytics, metrics, KPIs, or data (e.g., "What were total sales last quarter?", "Show me top customers by revenue"). NOT for validating, testing, or building dbt models during development. building-dbt-semantic-layer Use when creating or modifying dbt Semantic Layer components — semantic models, metrics, dimensions, entities, measures, or time spines. Covers MetricFlow configuration, metric types (simple, derived, cumulative, ratio, conversion), and validation for both latest and legacy YAML specs. configuring-dbt-mcp-server Generates MCP server configuration JSON, resolves authentication setup, and validates server connectivity for dbt. Use when setting up, configuring, or troubleshooting the dbt MCP server for AI tools like Claude Desktop, Claude Code, Cursor, or VS Code. fetching-dbt-docs Retrieves and searches dbt documentation pages in LLM-friendly markdown format. Use when fetching dbt documentation, looking up dbt features, or answering questions about dbt Cloud, dbt Core, or the dbt Semantic Layer. running-dbt-commands Formats and executes dbt CLI commands, selects the correct dbt executable, and structures command parameters. Use when running models, tests, builds, compiles, or show queries via dbt CLI. Use when unsure which dbt executable to use or how to format command parameters. troubleshooting-dbt-job-errors Diagnoses dbt Cloud/platform job failures by analyzing run logs, querying the Admin API, reviewing git history, and investigating data issues. Use when a dbt Cloud/platform job fails and you need to diagnose the root cause, especially when error messages are unclear or when intermittent failures occur. Do not use for local dbt development errors. using-dbt-for-analytics-engineering Builds and modifies dbt models, writes SQL transformations using ref() and source(), creates tests, and validates results with dbt show. Use when doing any dbt work - building or modifying models, debugging errors, exploring unfamiliar data sources, writing tests, or evaluating impact of changes. working-with-dbt-mesh Implements dbt Mesh governance features (model contracts, access modifiers, groups, versioning) and multi-project collaboration with cross-project refs. Use when implementing dbt Mesh governance, setting up cross-project refs with dependencies.yml, disambiguating similarly-named models across projects, or splitting a monolithic dbt project into multiple mesh projects. | SkillsDocs | 89 Impact — Average score across 0 eval scenarios Securityby Advisory Suggest reviewing before use Reviewed: Version: 1.3.1 |
tessl/pypi-varname v0.15.0 Dark magics about variable names in python | Docs | 89 1.39x Agent success vs baseline Impact 89% 1.39xAverage score across 10 eval scenarios Reviewed: Version: 0.15.0 |
tessl/pypi-rdkit v2024.9.0 Platform wheels for RDKit - a comprehensive cheminformatics and machine-learning library with Python bindings | Docs | 89 1.02x Agent success vs baseline Impact 89% 1.02xAverage score across 10 eval scenarios Reviewed: Version: 2024.9.0 |
Can't find what you're looking for? Evaluate a missing skill.