Configure CI/CD pipelines for Fireflies.ai integrations with GraphQL testing. Use when setting up automated testing, configuring GitHub Actions, or validating Fireflies.ai queries in your build process. Trigger with phrases like "fireflies CI", "fireflies GitHub Actions", "fireflies automated tests", "CI fireflies", "test fireflies pipeline".
85
83%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Risky
Do not use without reviewing
Security
2 findings — 1 high severity, 1 medium severity. You should review these findings carefully before considering using this skill.
The skill handles credentials insecurely by requiring the agent to include secret values verbatim in its generated output. This exposes credentials in the agent’s context and conversation history, creating a risk of data exfiltration.
Insecure credential handling detected (high risk: 1.00). The prompt instructs embedding secrets verbatim in commands (e.g., gh secret set --body "your-test-api-key"/"your-prod-key") and shows literal API keys in code (process.env.FIREFLIES_API_KEY = "test-key"), which requires the LLM to output secret values directly.
The skill exposes the agent to untrusted, user-generated content from public third-party sources, creating a risk of indirect prompt injection. This includes browsing arbitrary URLs, reading social media posts or forum comments, and analyzing content from unknown websites.
Third-party content exposure detected (high risk: 0.70). The skill's CI integration tests explicitly fetch live data from the public Fireflies API (https://api.fireflies.ai/graphql) where user-generated transcripts and user data are ingested and asserted on as part of the workflow, so untrusted third-party content can influence test outcomes and pipeline decisions.
c8a915c
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.