Evaluate scientific claims and evidence quality. Use for assessing experimental design validity, identifying biases and confounders, applying evidence grading frameworks (GRADE, Cochrane Risk of Bias), or teaching critical analysis. Best for understanding evidence quality, identifying flaws. For formal peer review writing use peer-review.
76
67%
Does it follow best practices?
Impact
96%
1.02xAverage score across 3 eval scenarios
Advisory
Suggest reviewing before use
Optimize this skill with Tessl
npx tessl skill review --optimize ./scientific-skills/scientific-critical-thinking/SKILL.mdSecurity
1 medium severity finding. This skill can be installed but you should review these findings before use.
The skill fetches instructions or code from an external URL at runtime, and the fetched content directly controls the agent’s prompts or executes code. This dynamic dependency allows the external source to modify the agent’s behavior without any changes to the skill itself.
Potentially malicious external URL detected (high risk: 0.90). The skill includes runtime calls to the OpenRouter API (https://openrouter.ai/api/v1) — the scripts require an OPENROUTER_API_KEY and use model responses (Gemini image generation and Gemini 3.1 quality review) to drive iterative diagram generation and decision-making, so external content directly controls the agent's behavior.
25e1c0f
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.