--- name: scout-mindset-bias-check description: Use to detect and remove cognitive biases from reasoning. Invoke when prediction feels emotional, stuck at 50/50, or when you want to validate forecasting process. Use when user mentions scout mindset, soldier mindset, bias check, reversal test, scope sensitivity, or cognitive distortions. --- # Scout Mindset & Bias Check ## Table of Contents - [What is Scout Mindset?](#what-is-scout-mindset) - [When to Use This Skill](#when-to-use-this-skill) - [Interactive Menu](#interactive-menu) - [Quick Reference](#quick-reference) - [Resource Files](#resource-files) --- ## What is Scout Mindset? **Scout Mindset** (Julia Galef) is the motivation to see things as they are, not as you wish them to be. Contrast with **Soldier Mindset**, which defends a position regardless of evidence. **Core Principle:** Your goal is to map the territory accurately, not win an argument. **Why It Matters:** - Forecasting requires intellectual honesty - Biases systemically distort probabilities - Emotional attachment clouds judgment - Motivated reasoning leads to overconfidence --- ## When to Use This Skill Use this skill when: - **Prediction feels emotional** - You want a certain outcome - **Stuck at 50/50** - Indecisive, can't commit to probability - **Defending a position** - Arguing for your forecast, not questioning it - **After inside view analysis** - Used specific details, need bias check - **Disagreement with others** - Different people, different probabilities - **Before finalizing** - Last sanity check Do NOT skip this when stakes are high, you have strong priors, or forecast affects you personally. --- ## Interactive Menu **What would you like to do?** ### Core Workflows **1. [Run the Reversal Test](#1-run-the-reversal-test)** - Check if you'd accept opposite evidence - Detect motivated reasoning - Validate evidence standards - Expose special pleading **2. [Check Scope Sensitivity](#2-check-scope-sensitivity)** - Ensure probabilities scale with inputs - Linear scaling test - Reference point calibration - Magnitude assessment **3. [Test Status Quo Bias](#3-test-status-quo-bias)** - Challenge "no change" assumptions - Entropy principle - Change vs stability energy - Default state inversion **4. [Audit Confidence Intervals](#4-audit-confidence-intervals)** - Validate CI width - Surprise test - Historical calibration - Overconfidence check **5. [Run Full Bias Audit](#5-run-full-bias-audit)** - Comprehensive bias scan - All major cognitive biases - Systematic checklist - Prioritized remediation **6. [Learn the Framework](#6-learn-the-framework)** - Deep dive into methodology - Read [Scout vs Soldier Mindset](resources/scout-vs-soldier.md) - Read [Cognitive Bias Catalog](resources/cognitive-bias-catalog.md) - Read [Debiasing Techniques](resources/debiasing-techniques.md) **7. Exit** - Return to main forecasting workflow --- ## 1. Run the Reversal Test **Check if you'd accept evidence pointing the opposite direction.** ``` Reversal Test Progress: - [ ] Step 1: State your current conclusion - [ ] Step 2: Identify supporting evidence - [ ] Step 3: Reverse the evidence - [ ] Step 4: Ask "Would I still accept it?" - [ ] Step 5: Adjust for double standards ``` ### Step 1: State your current conclusion **What are you predicting?** - Prediction: [Event] - Probability: [X]% - Direction: [High/Low confidence] ### Step 2: Identify supporting evidence **List the evidence that supports your conclusion.** **Example:** Candidate A will win (75%) 1. Polls show A ahead by 5% 2. A has more campaign funding 3. Expert pundits favor A 4. A has better debate ratings ### Step 3: Reverse the evidence **Imagine the same evidence pointed the OTHER way.** **Reversed:** What if polls showed B ahead, B had more funding, experts favored B, and B had better ratings? ### Step 4: Ask "Would I still accept it?" **The Critical Question:** > If this reversed evidence existed, would I accept it as valid and change my prediction? **Three possible answers:** **A) YES - I would accept reversed evidence** ✓ No bias detected, continue with current reasoning **B) NO - I would dismiss reversed evidence** ⚠ **Warning:** Motivated reasoning - you're accepting evidence when it supports you, dismissing equivalent evidence when it doesn't (special pleading) **C) UNSURE - I'd need to think about it** ⚠ **Warning:** Asymmetric evidence standards suggest rationalizing, not reasoning ### Step 5: Adjust for double standards **If you answered B or C:** **Ask:** Why do I dismiss this evidence in one direction but accept it in the other? Is there an objective reason, or am I motivated by preference? **Common rationalizations:** - "This source is biased" (only when it disagrees) - "Sample size too small" (only for unfavorable polls) - "Outlier data" (only for data you dislike) - "Context matters" (invoked selectively) **The Fix:** - **Option 1:** Reject the evidence entirely (if you wouldn't trust it reversed, don't trust it now) - **Option 2:** Accept it in both directions (trust evidence regardless of direction) - **Option 3:** Weight it appropriately (maybe it's weak evidence both ways) **Probability adjustment:** If you detected double standards, move probability 10-15% toward 50% **Next:** Return to [menu](#interactive-menu) --- ## 2. Check Scope Sensitivity **Ensure your probabilities scale appropriately with magnitude.** ``` Scope Sensitivity Progress: - [ ] Step 1: Identify the variable scale - [ ] Step 2: Test linear scaling - [ ] Step 3: Check reference point calibration - [ ] Step 4: Validate magnitude assessment - [ ] Step 5: Adjust for scope insensitivity ``` ### Step 1: Identify the variable scale **What dimension has magnitude?** - Number of people (100 vs 10,000 vs 1,000,000) - Dollar amounts ($1K vs $100K vs $10M) - Time duration (1 month vs 1 year vs 10 years) ### Step 2: Test linear scaling **The Linearity Test:** Double the input, check if impact doubles. **Example: Startup funding** - If raised $1M: ___% - If raised $10M: ___% - If raised $100M: ___% **Scope sensitivity check:** Did probabilities scale reasonably? If they barely changed → Scope insensitive ### Step 3: Check reference point calibration **The Anchoring Test:** Did you start with a number (base rate, someone else's forecast, round number) and insufficiently adjust? **The fix:** - Generate probability from scratch without looking at others - Then compare and reconcile differences - Don't just "split the difference" - reason about why estimates differ ### Step 4: Validate magnitude assessment **The "1 vs 10 vs 100" Test:** For your forecast, vary the scale by 10×. **Example: Project timeline** - 1 month: P(success) = ___% - 10 months: P(success) = ___% - 100 months: P(success) = ___% **Expected:** Probability should change significantly. If all three estimates are within 10 percentage points → Scope insensitivity ### Step 5: Adjust for scope insensitivity **The problem:** Your emotional system responds to the category, not the magnitude. **The fix:** **Method 1: Logarithmic scaling** - Use log scale for intuition **Method 2: Reference class by scale** - Don't use "startups" as reference class. Use "Startups that raised $1M" (10% success) vs "Startups that raised $100M" (60% success) **Method 3: Explicit calibration** - Use a formula: P(success) = base_rate + k × log(amount) **Next:** Return to [menu](#interactive-menu) --- ## 3. Test Status Quo Bias **Challenge the assumption that "no change" is the default.** ``` Status Quo Bias Progress: - [ ] Step 1: Identify status quo prediction - [ ] Step 2: Calculate energy to maintain status quo - [ ] Step 3: Invert the default - [ ] Step 4: Apply entropy principle - [ ] Step 5: Adjust probabilities ``` ### Step 1: Identify status quo prediction **Are you predicting "no change"?** Examples: "This trend will continue," "Market share will stay the same," "Policy won't change" Status quo predictions often get inflated probabilities because change feels risky. ### Step 2: Calculate energy to maintain status quo **The Entropy Principle:** In the absence of active energy input, systems decay toward disorder. **Question:** "What effort is required to keep things the same?" **Examples:** - **Market share:** To maintain requires matching competitor innovation → Energy required: High → Status quo is HARD - **Policy:** To maintain requires no proposals for change → Energy required: Low → Status quo is easier ### Step 3: Invert the default **Mental Exercise:** - **Normal framing:** "Will X change?" (Default = no) - **Inverted framing:** "Will X stay the same?" (Default = no) **Bias check:** If P(change) + P(same) ≠ 100%, you have status quo bias. ### Step 4: Apply entropy principle **Second Law of Thermodynamics (applied to forecasting):** **Ask:** 1. Is this system open or closed? 2. Is energy being input to maintain/improve? 3. Is that energy sufficient? ### Step 5: Adjust probabilities **If you detected status quo bias:** **For "no change" predictions that require high energy:** - Reduce P(status quo) by 10-20% - Increase P(change) correspondingly **For predictions where inertia truly helps:** No adjustment needed **The heuristic:** If maintaining status quo requires active effort, decay is more likely than you think. **Next:** Return to [menu](#interactive-menu) --- ## 4. Audit Confidence Intervals **Validate that your CI width reflects true uncertainty.** ``` Confidence Interval Audit Progress: - [ ] Step 1: State current CI - [ ] Step 2: Run surprise test - [ ] Step 3: Check historical calibration - [ ] Step 4: Compare to reference class variance - [ ] Step 5: Adjust CI width ``` ### Step 1: State current CI **Current confidence interval:** - Point estimate: ___% - Lower bound: ___% - Upper bound: ___% - Width: ___ percentage points - Confidence level: ___ (usually 80% or 90%) ### Step 2: Run surprise test **The Surprise Test:** "Would I be **genuinely shocked** if the true value fell outside my confidence interval?" **Calibration:** - 80% CI → Should be shocked 20% of the time - 90% CI → Should be shocked 10% of the time **Test:** Imagine the outcome lands just below your lower bound or just above your upper bound. **Three possible answers:** - **A) "Yes, I'd be very surprised"** - ✓ CI appropriately calibrated - **B) "No, not that surprised"** - ⚠ CI too narrow (overconfident) → Widen interval - **C) "I'd be amazed if it landed in the range"** - ⚠ CI too wide → Narrow interval ### Step 3: Check historical calibration **Look at your past forecasts:** 1. Collect last 20-50 forecasts with CIs 2. Count how many actual outcomes fell outside your CIs 3. Compare to theoretical expectation | CI Level | Expected Outside | Your Actual | |----------|------------------|-------------| | 80% | 20% | ___% | | 90% | 10% | ___% | **Diagnosis:** Actual > Expected → CIs too narrow (overconfident) - Most common ### Step 4: Compare to reference class variance **If you have reference class data:** 1. Calculate standard deviation of reference class outcomes 2. Your CI should roughly match that variance **Example:** Reference class SD = 12%, your 80% CI ≈ Point estimate ± 15% If your CI is narrower than reference class variance, you're claiming to know more than average. Justify why, or widen CI. ### Step 5: Adjust CI width **Adjustment rules:** - **If overconfident:** Multiply current width by 1.5× to 2× - **If underconfident:** Reduce width by 0.5× to 0.75× **Next:** Return to [menu](#interactive-menu) --- ## 5. Run Full Bias Audit **Comprehensive scan of major cognitive biases.** ``` Full Bias Audit Progress: - [ ] Step 1: Confirmation bias check - [ ] Step 2: Availability bias check - [ ] Step 3: Anchoring bias check - [ ] Step 4: Affect heuristic check - [ ] Step 5: Overconfidence check - [ ] Step 6: Attribution error check - [ ] Step 7: Prioritize and remediate ``` See [Cognitive Bias Catalog](resources/cognitive-bias-catalog.md) for detailed descriptions. **Quick audit questions:** ### 1. Confirmation Bias - [ ] Did I seek out disconfirming evidence? - [ ] Did I give equal weight to evidence against my position? - [ ] Did I actively try to prove myself wrong? **If NO to any → Confirmation bias detected** ### 2. Availability Bias - [ ] Did I rely on recent/memorable examples? - [ ] Did I use systematic data vs "what comes to mind"? - [ ] Did I check if my examples are representative? **If NO to any → Availability bias detected** ### 3. Anchoring Bias - [ ] Did I generate my estimate independently first? - [ ] Did I avoid being influenced by others' numbers? - [ ] Did I adjust sufficiently from initial anchor? **If NO to any → Anchoring bias detected** ### 4. Affect Heuristic - [ ] Do I have an emotional preference for the outcome? - [ ] Did I separate "what I want" from "what will happen"? - [ ] Would I make the same forecast if incentives were reversed? **If NO to any → Affect heuristic detected** ### 5. Overconfidence - [ ] Did I run a premortem? - [ ] Are my CIs wide enough (surprise test)? - [ ] Did I identify ways I could be wrong? **If NO to any → Overconfidence detected** ### 6. Fundamental Attribution Error - [ ] Did I attribute success to skill vs luck appropriately? - [ ] Did I consider situational factors, not just personal traits? - [ ] Did I avoid "great man" narratives? **If NO to any → Attribution error detected** ### Step 7: Prioritize and remediate **For each detected bias:** 1. **Severity:** High / Medium / Low 2. **Direction:** Pushing probability up or down? 3. **Magnitude:** Estimated percentage point impact **Remediation example:** | Bias | Severity | Direction | Adjustment | |------|----------|-----------|------------| | Confirmation | High | Up | -15% | | Availability | Medium | Up | -10% | | Affect heuristic | High | Up | -20% | **Net adjustment:** -45% → Move probability down by 45 points (e.g., 80% → 35%) **Next:** Return to [menu](#interactive-menu) --- ## 6. Learn the Framework **Deep dive into the methodology.** ### Resource Files 📄 **[Scout vs Soldier Mindset](resources/scout-vs-soldier.md)** - Julia Galef's framework - Motivated reasoning - Intellectual honesty - Identity and beliefs 📄 **[Cognitive Bias Catalog](resources/cognitive-bias-catalog.md)** - 20+ major biases - How they affect forecasting - Detection methods - Remediation strategies 📄 **[Debiasing Techniques](resources/debiasing-techniques.md)** - Systematic debiasing process - Pre-commitment strategies - External accountability - Algorithmic aids **Next:** Return to [menu](#interactive-menu) --- ## Quick Reference ### The Scout Commandments 1. **Truth over comfort** - Accuracy beats wishful thinking 2. **Seek disconfirmation** - Try to prove yourself wrong 3. **Hold beliefs lightly** - Probabilistic, not binary 4. **Update incrementally** - Change mind with evidence 5. **Separate wanting from expecting** - Desire ≠ Forecast 6. **Check your work** - Run bias audits routinely 7. **Stay calibrated** - Track accuracy over time > Scout mindset is the drive to see things as they are, not as you wish them to be. --- ## Resource Files 📁 **resources/** - [scout-vs-soldier.md](resources/scout-vs-soldier.md) - Mindset framework - [cognitive-bias-catalog.md](resources/cognitive-bias-catalog.md) - Comprehensive bias reference - [debiasing-techniques.md](resources/debiasing-techniques.md) - Remediation strategies --- **Ready to start? Choose a number from the [menu](#interactive-menu) above.**