cognitive-bias-detection
Cognitive Bias Detection
Core: Reasoning is systematically distorted by predictable errors operating below conscious awareness. The most dangerous analyses are the ones that feel most certain. This skill audits the reasoning process, not just conclusions.
Highest-Impact Biases
Evaluation & Decision
Confirmation Bias — Seeking/interpreting/remembering info that confirms existing beliefs; disconfirming evidence is dismissed.
- Signal: "The data confirms what we suspected." Counter-evidence underweighted.
- Fix: Seek the strongest case against. Assign someone to argue opposite.
Anchoring — Overweighting the first number/estimate/framing.
- Signal: Estimates cluster around an initial figure; comparisons against an unvalidated reference.
- Fix: Generate estimates independently first. "What would this look like without the anchor?"
Availability Heuristic — Overweighting recent/vivid events.
- Signal: "We just had this incident" → overestimating its probability. Quiet failures ignored.
- Fix: Use base rates. "How often does this happen over a long period?"
Sunk Cost Fallacy — Continuing because of past investment, not future value.
- Signal: "We've put 6 months into this." Reluctance to abandon despite evidence.
- Fix: "If we hadn't invested anything, would we start this today?"
Planning Fallacy — Systematic underestimation of time/cost/risk.
- Signal: Optimistic estimates, no buffer, past overruns treated as exceptions.
- Fix: Reference class forecasting — how long did similar projects actually take?
Social & Group
Groupthink — Group harmony overrides realistic appraisal; dissent suppressed.
- Signal: Quick agreement, no devil's advocate, contrarian views dismissed socially.
- Fix: Assign formal devil's advocate. Independent written opinions before discussion.
Authority Bias — Overweighting authority's opinion independent of actual expertise.
- Signal: "The CTO thinks X, so it must be right." Analysis stops when authority speaks.
- Fix: Evaluate argument on merits. "What's the evidence, separate from who said it?"
In-group Bias — Favoring own group's people/ideas/solutions.
- Signal: Internal solutions evaluated more generously than identical external ones.
- Fix: Blind evaluation. "Would we accept this from a competitor?"
Framing & Perception
Framing Effect — Same info → different decisions based on presentation (gain vs. loss).
- Signal: "90% success" vs. "10% failure" trigger different reactions.
- Fix: Reframe every option multiple ways. Does the decision change?
Survivorship Bias — Conclusions from visible successes, ignoring invisible failures.
- Signal: "Company X did Y and succeeded" — but how many tried Y and failed?
- Fix: Seek failure cases. "What don't we see because they didn't survive?"
Dunning-Kruger — Low competence → overconfidence; high competence → underconfidence.
- Signal: Extreme certainty in novel/complex domain, or excessive expert hedging.
- Fix: Calibrate confidence against demonstrated track record in this domain.
Recency Bias — Overweighting recent data, underweighting long-term patterns.
- Signal: Last quarter dominates analysis; historical base rates ignored.
- Fix: Extend the time window. Multi-year trends, not just recent.
Output Format
🔍 Bias Scan
| Bias | Present? | Signal Observed | Severity |
|---|---|---|---|
| Confirmation | Y/Possible/N | [Evidence] | L/M/H |
| Sunk Cost | Y/Possible/N | [Evidence] | L/M/H |
| ... |
⚠️ High-Risk Findings
For each high-severity bias:
- Bias: Name + brief description
- How it shows up: Specific evidence
- What it distorts: Conclusion/estimate skewed, in which direction
- Debiasing move: Concrete corrective action
🧹 Debiased Re-evaluation
- What changes if the bias is removed?
- What evidence is actually strong vs. inflated by bias?
- Does the conclusion still hold?
🎯 Confidence Calibration
- Actual warranted confidence absent bias?
- What needs to be true for higher confidence?
- Most important thing to validate before committing?
Meta-Check: Is Claude Biased Here?
When generating evaluations, check:
- Confirming what the user wants to hear? (sycophancy / confirmation)
- Anchoring to the user's first framing?
- Overweighting most vivid/recent example?
- Assuming user's group/team/approach is better without evidence?
If yes — flag and correct.
Triggers
- "How would this look if we'd concluded the opposite from the start?"
- "What's the strongest evidence against the current conclusion?"
- "Are we continuing because it's right, or because we've invested too much to stop?"
- "Who benefits from this conclusion, and are they evaluating it?"
- "If a stranger reviewed this reasoning, what would they say we're missing?"
More from andurilcode/skills
causal-inference
Apply causal inference whenever the user is interpreting metrics, debugging system behavior, reading A/B test results, or trying to understand whether an observed change was caused by an action or by something else. Triggers on phrases like "X caused Y", "since we deployed this, metrics changed", "the A/B test showed a lift", "why did this metric move?", "is this correlation or causation?", "we changed X and Y improved", "how do we know this worked?", "the data shows…", or any situation where conclusions are being drawn from observational data. Also trigger before any decision based on metric interpretation — confusing correlation with causation leads to interventions that don't work and misattribution of credit. Never assume causation without applying this skill.
30probabilistic-thinking
Apply probabilistic and Bayesian thinking whenever the user needs to reason under uncertainty, compare risks, prioritize between options, update beliefs based on new evidence, or make decisions without complete information. Triggers on phrases like "what are the odds?", "how likely is this?", "should I be worried about X?", "which risk is bigger?", "does this data change anything?", "is this a signal or noise?", "what's the probability?", "how confident are we?", or any situation where decisions are being made based on incomplete or ambiguous evidence. Also trigger when someone is treating uncertain outcomes as certainties, or when probability language is being used loosely ("probably", "unlikely", "very likely") without quantification. Don't leave uncertainty unexamined.
27inversion-premortem
Apply inversion and pre-mortem thinking whenever the user asks to evaluate a plan, strategy, architecture, feature, or decision before execution — or when they want to stress-test something that already exists. Triggers on phrases like "is this a good idea?", "what could go wrong?", "review this plan", "should we do this?", "are we missing anything?", "stress-test this", "what are the risks?", or any request to validate a decision or design. Use this skill proactively — if the user is about to commit to something, this skill should be consulted even if they don't ask for it explicitly.
23analogical-thinking
Apply analogical thinking whenever the user is designing a system, architecture, or process and would benefit from structural patterns that already exist in other domains — or when a problem feels novel but may have been solved elsewhere under a different name. Triggers on phrases like "how should we structure this?", "has anyone solved this before?", "we're designing from scratch", "what's a good model for this?", "I keep feeling like this resembles something", "what patterns apply here?", or when facing architecture, organizational, or process design decisions. Also trigger when a problem has been analyzed thoroughly but no good solution has emerged — the answer may exist in an adjacent domain. Don't reinvent what's been solved. Recognize the shape of the problem first.
22first-principles-thinking
Apply first principles thinking whenever the user is questioning whether a design, strategy, or solution is fundamentally right — not just well-executed. Triggers on phrases like "are we solving the right problem?", "why do we do it this way?", "is this the best approach?", "everyone does X but should we?", "we've always done it this way", "challenge our assumptions", "start from scratch", "is there a better way?", or when the user seems to be iterating on a flawed premise rather than questioning the premise itself. Also trigger when a proposed solution feels like an incremental improvement on something that may be fundamentally broken. Don't optimize a flawed foundation — question it first.
21scenario-planning
Apply scenario planning whenever the user is making long-term decisions, building roadmaps, evaluating strategies, or operating in an environment with significant uncertainty about how the future will unfold. Triggers on phrases like "what should our roadmap look like?", "how do we plan for the future?", "what if things change?", "we're not sure which direction the market will go", "how do we make this strategy resilient?", "what's our plan B?", "what are the different futures we could face?", or when a plan assumes a single future state. Also trigger when someone is over-committed to one expected outcome and hasn't stress-tested the strategy against alternative futures. Don't plan for one future — plan for multiple.
21