Advanced AI agent benchmark scenarios that push Vercel's cutting-edge platform features — Workflow DevKit, AI Gateway, MCP, Chat SDK, Queues, Flags, Sandbox, and multi-agent orchestration. Designed to stress-test skill injection for complex, multi-system builds.
44
43%
Does it follow best practices?
Impact
—
No eval scenarios have been run
Advisory
Suggest reviewing before use
Optimize this skill with Tessl
npx tessl skill review --optimize ./.claude/skills/benchmark-agents/SKILL.mdSecurity
1 medium severity finding. This skill can be installed but you should review these findings before use.
The skill fetches instructions or code from an external URL at runtime, and the fetched content directly controls the agent’s prompts or executes code. This dynamic dependency allows the external source to modify the agent’s behavior without any changes to the skill itself.
Potentially malicious external URL detected (high risk: 1.00). The skill explicitly runs "npx add-plugin https://github.com/vercel/vercel-plugin -s project -y" during setup, which fetches and executes code from the GitHub repository at runtime and is required for the eval flow (https://github.com/vercel/vercel-plugin), so it is a runtime external dependency that can execute remote code.
61f1903
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.