CtrlK
BlogDocsLog inGet started
Tessl Logo

pantheon-ai/challenge

Challenge AI output with structured devil's-advocate protocols: anchor, verify, framing, and deep sub-commands.

86

Quality

86%

Does it follow best practices?

Impact

Pending

No eval scenarios have been run

SecuritybySnyk

Passed

No known issues

Overview
Quality
Evals
Security
Files

reference.mdreferences/

Challenge — Reference

Pattern Catalog (v1: 8 patterns)

#PatternSubcommandError TypeSourceMechanism
1GatekeeperanchorPremature commitmentFlorian WP01 §4.5Demand explicit pass/fail criteria for the decision BEFORE accepting it. Ask: "What must be true for this to be correct?" then check which criteria remain unverified. Blocks premature acceptance.
2ResetanchorPremature commitmentFlorian WP01 §4.5Discard the current solution entirely. Re-read ONLY the original problem. Generate a fresh answer from first principles. Compare with original — divergences reveal anchoring.
3Alternative ApproachesanchorPremature commitmentWhite et al. 2023 (Vanderbilt)Force generation of ≥2 genuine alternatives (different mechanisms, not strawmen). For each: "Why was this NOT chosen?" — if no reason exists, the original wasn't properly evaluated.
4Pre-mortemanchorPremature commitmentKlein 2007; EMNLP 2024 (Wang)Assume the decision was implemented and FAILED. Generate 3 failure scenarios (most likely, second-order, black swan). For each: identify early warning signs and mitigation actions available NOW.
5Proof DemandverifyFactual errorsFlorian WP01 §4.5Extract every factual claim. For each, classify: ✅ Sourced (citation exists), ⚠️ Unsourced (stated as fact without evidence), ❌ Contradicted (evidence disagrees). Flag ⚠️/❌ as untrustworthy until verified.
6CoVeverifyFactual errorsDhuliawala et al. 2023 (Meta)4-step Chain-of-Verification: (1) state the claim, (2) write specific verification questions, (3) answer each question INDEPENDENTLY without looking at the original claim, (4) compare independent answers to original — discrepancies = errors.
7Fact Check ListverifyFactual errorsWhite et al. 2023 (Vanderbilt)Decompose response into atomic checkable assertions (one fact per line). Rate confidence: High/Medium/Low/Unknown. For Low/Unknown: write a concrete verification action (search query, doc to read, test to run). Priority-order by impact × uncertainty.
8SocraticframingFraming errorsChang 20236-stage questioning: (1) Definition — what do key terms actually mean? (2) Elenchus — is the definition consistent with usage? (3) Dialectic — what's the opposite position? (4) Maieutics — what's the real goal, stripped of framing? (5) Generalization — local issue or systemic pattern? (6) Counterfactual — if the problem didn't exist, what changes?
9SteelmanframingFraming errorsRationalist traditionBuild the STRONGEST possible counter-argument to the current framing (opposite of strawman). Use best available evidence for the counter-position. Assume it's correct — what would that imply? Stress-test: "What must be true for the steelman to be wrong?" If no clear answer, the framing is weak.

When-To-Use Guide

SymptomError TypeSubcommand
AI picked one solution without exploring optionsAnchoring bias/challenge anchor
AI committed to approach too quicklyPremature commitment/challenge anchor
Specific facts, numbers, or claims feel uncertainFactual error/challenge verify
AI cited something that can't be verifiedHallucination/challenge verify
The answer is correct but might be solving the wrong problemFraming error/challenge framing
The question itself seems wrongWrong problem/challenge framing
High-stakes decision, want genuine debiasing in fresh contextAll error types/challenge deep

Challenge Report Format

All subcommand outputs MUST follow this structure:

## Challenge Report: [Pattern(s) Applied]

**Target**: [what was challenged]
**Error type**: [anchoring | factual | framing | high-stakes]

### Technique Selection

- **Family**: [anchor (premature commitment) | verify (factual errors) | framing (wrong problem)]
- **Patterns applied**: [named patterns, e.g., Gatekeeper, Pre-mortem, CoVe]
- **Why these patterns**: [what about the target triggered this selection — specific observations]
- **Patterns considered but skipped**: [and why, or "none — full protocol applied"]

### Findings

[Pattern-specific structured output — see protocol files.
For each finding include: Observation → Technique (family + pattern) → Reasoning → Confidence]

### Verdict

- **Assessment**: [Decision holds / Needs revision / Needs rejection]
- **Confidence**: [High / Medium / Low]
- **What would flip this**: [specific evidence or condition that would change the verdict]
- **Strongest counter to this verdict**: [steelman the opposite conclusion]

### Recommended Action

[Proceed as-is | Proceed with modifications: X | Reconsider: Y]

Sources

  • Florian WP01 §4.5 — Gatekeeper, Proof Demand, Reset (practitioner patterns)
  • White et al. 2023 — Vanderbilt Prompt Pattern Catalog (arxiv.org/abs/2302.11382)
  • Dhuliawala et al. 2023 — Chain-of-Verification (CoVe), Meta AI (arxiv.org/abs/2309.11495)
  • Wang et al. 2024 — Devil's Advocate, EMNLP (arxiv.org/abs/2405.16334)
  • Chang 2023 — Socratic Method prompting (arxiv.org/abs/2303.08769)
  • Klein 2007 — Pre-mortem analysis (Harvard Business Review)

references

reference.md

SKILL.md

tile.json