9.2 KiB
Bayesian Reasoning Template
Workflow
Copy this checklist and track your progress:
Bayesian Update Progress:
- [ ] Step 1: State question and establish prior from base rates
- [ ] Step 2: Estimate likelihoods for evidence
- [ ] Step 3: Calculate posterior using Bayes' theorem
- [ ] Step 4: Perform sensitivity analysis
- [ ] Step 5: Calibrate and validate with quality checklist
Step 1: State question and establish prior from base rates
Define specific, testable hypothesis with timeframe and success criteria. Identify reference class and base rate, adjust for specific differences, and state prior explicitly with justification. See Step 1: State the Question and Step 2: Find Base Rates for guidance.
Step 2: Estimate likelihoods for evidence
Assess P(E|H) (probability of evidence if hypothesis TRUE) and P(E|¬H) (probability if FALSE), calculate likelihood ratio = P(E|H) / P(E|¬H), and interpret diagnostic strength. See Step 3: Estimate Likelihoods for examples and common mistakes.
Step 3: Calculate posterior using Bayes' theorem
Apply P(H|E) = [P(E|H) × P(H)] / P(E) or use simpler odds form: Posterior Odds = Prior Odds × LR. Interpret change in belief (prior → posterior) and strength of evidence. See Step 4: Calculate Posterior for calculation methods.
Step 4: Perform sensitivity analysis
Test how posterior changes with different prior values and likelihoods to assess robustness of conclusion. See Sensitivity Analysis section in template structure.
Step 5: Calibrate and validate with quality checklist
Check for overconfidence, base rate neglect, and extreme posteriors. Use Calibration Check and Quality Checklist to verify prior is justified, likelihoods have reasoning, evidence is diagnostic (LR ≠ 1), calculation correct, and assumptions stated.
Quick Template
# Bayesian Analysis: {Topic}
## Question
**Hypothesis**: {What are you testing?}
**Estimating**: P({specific outcome})
**Timeframe**: {When will outcome be known?}
**Matters because**: {What decision depends on this?}
---
## Prior Belief (Before Evidence)
### Base Rate
{What's the general frequency in similar cases?}
- Reference class: {Similar situations}
- Base rate: {X%}
### Adjustments
{How is this case different from base rate?}
- Factor 1: {Increases/decreases probability because...}
- Factor 2: {Increases/decreases probability because...}
### Prior Probability
**P(H) = {X%}**
**Justification**: {Why this prior?}
**Range if uncertain**: {min%} to {max%}
---
## Evidence
**What was observed**: {Specific evidence or data}
**How diagnostic**: {Does this distinguish hypothesis true vs false?}
### Likelihoods
**P(E|H) = {X%}** - Probability of seeing this evidence IF hypothesis is TRUE
- Reasoning: {Why this likelihood?}
**P(E|¬H) = {Y%}** - Probability of seeing this evidence IF hypothesis is FALSE
- Reasoning: {Why this likelihood?}
**Likelihood Ratio = {X/Y} = {ratio}**
- Interpretation: Evidence is {very strong / moderate / weak / not diagnostic}
---
## Bayesian Update
### Calculation
**Using probability form**:
P(H|E) = [P(E|H) × P(H)] / P(E)
where P(E) = [P(E|H) × P(H)] + [P(E|¬H) × P(¬H)]
P(E) = [{X%} × {Prior%}] + [{Y%} × {100-Prior%}] P(E) = {calculation}
P(H|E) = [{X%} × {Prior%}] / {P(E)} P(H|E) = {result%}
**Or using odds form** (often simpler):
Prior Odds = P(H) / P(¬H) = {Prior%} / {100-Prior%} = {odds} Likelihood Ratio = {LR} Posterior Odds = Prior Odds × LR = {odds} × {LR} = {result} Posterior Probability = Posterior Odds / (1 + Posterior Odds) = {result%}
### Posterior Probability
**P(H|E) = {result%}**
### Change in Belief
- Prior: {X%}
- Posterior: {Y%}
- Change: {+/- Z percentage points}
- Interpretation: Evidence {strongly supports / moderately supports / weakly supports / contradicts} hypothesis
---
## Sensitivity Analysis
**How sensitive is posterior to inputs?**
If Prior was {different value}:
- Posterior would be: {recalculated value}
If P(E|H) was {different value}:
- Posterior would be: {recalculated value}
**Robustness**: Conclusion is {robust / somewhat robust / sensitive} to assumptions
---
## Calibration Check
**Am I overconfident?**
- Did I anchor on initial belief? {yes/no - reasoning}
- Did I ignore base rates? {yes/no - reasoning}
- Is my posterior extreme (>90% or <10%)? {If yes, is evidence truly that strong?}
- Would an outside observer agree with my likelihoods? {check}
**Red flags**:
- ✗ Posterior is 100% or 0% (almost never justified)
- ✗ Large update from weak evidence (check LR)
- ✗ Prior ignores base rate entirely
- ✗ Likelihoods are guesses without reasoning
---
## Limitations & Assumptions
**Key assumptions**:
1. {Assumption 1}
2. {Assumption 2}
**What could invalidate this analysis**:
- {Condition that would change conclusion}
- {Different interpretation of evidence}
**Uncertainty**:
- Most uncertain about: {which input?}
- Could be wrong if: {what scenario?}
---
## Decision Implications
**Given posterior of {X%}**:
Recommended action: {what to do}
**If decision threshold is**:
- High confidence needed (>80%): {action}
- Medium confidence (>60%): {action}
- Low bar (>40%): {action}
**Next evidence to gather**: {What would further update belief?}
Step-by-Step Guide
Step 1: State the Question
Be specific and testable.
Good: "Will our product achieve >1000 DAU within 6 months?" Bad: "Will the product succeed?"
Define success criteria numerically when possible.
Step 2: Find Base Rates
Method:
- Identify reference class (similar situations)
- Look up historical frequency
- Adjust for known differences
Example:
- Question: Will our SaaS startup raise Series A?
- Reference class: B2B SaaS startups, seed stage, similar market
- Base rate: ~30% raise Series A within 2 years
- Adjustments: Strong traction (+), competitive market (-), experienced team (+)
- Adjusted prior: 45%
Common mistake: Ignoring base rates entirely ("inside view" bias)
Step 3: Estimate Likelihoods
Ask: "If hypothesis were true, how likely is this evidence?" Then: "If hypothesis were false, how likely is this evidence?"
Example - Medical test:
- Hypothesis: Patient has disease (prevalence 1%)
- Evidence: Positive test result
- P(positive test | has disease) = 90% (test sensitivity)
- P(positive test | no disease) = 5% (false positive rate)
- LR = 90% / 5% = 18 (strong evidence)
Common mistake: Confusing P(E|H) with P(H|E) - the "prosecutor's fallacy"
Step 4: Calculate Posterior
Odds form is often easier:
- Convert prior to odds: Odds = P / (1-P)
- Multiply by LR: Posterior Odds = Prior Odds × LR
- Convert back to probability: P = Odds / (1 + Odds)
Example:
- Prior: 30% → Odds = 0.3/0.7 = 0.43
- LR = 5
- Posterior Odds = 0.43 × 5 = 2.15
- Posterior Probability = 2.15 / 3.15 = 68%
Step 5: Calibrate
Calibration questions:
- If you made 100 predictions at X% confidence, would X actually occur?
- Are you systematically over/underconfident?
- Does your posterior pass the "outside view" test?
Calibration tips:
- Track your forecasts and outcomes
- Be especially skeptical of extreme probabilities (>95%, <5%)
- Consider opposite evidence (confirmation bias check)
Common Pitfalls
Ignoring base rates ("base rate neglect"):
- Bad: "Test is 90% accurate, so positive test means 90% chance of disease"
- Good: "Disease is rare (1%), so even with positive test, probability is only ~15%"
Confusing conditional probabilities:
- P(positive test | disease) ≠ P(disease | positive test)
- These can be very different!
Overconfident likelihoods:
- Claiming P(E|H) = 99% when evidence is ambiguous
- Not considering alternative explanations
Anchoring on prior:
- Weak evidence + starting at 50% = staying near 50%
- Solution: Use base rates, not 50% default
Treating all evidence as equally strong:
- Check likelihood ratio (LR)
- LR ≈ 1 means evidence is not diagnostic
Worked Example
Question: Will project finish on time?
Prior:
- Base rate: 60% of our projects finish on time
- This project: More complex than average (-), experienced team (+)
- Prior: 55%
Evidence: At 50% milestone, we're 1 week behind schedule
Likelihoods:
- P(behind at 50% | finish on time) = 30% (can recover)
- P(behind at 50% | miss deadline) = 80% (usually signals trouble)
- LR = 30% / 80% = 0.375 (evidence against on-time)
Calculation:
- Prior odds = 0.55 / 0.45 = 1.22
- Posterior odds = 1.22 × 0.375 = 0.46
- Posterior probability = 0.46 / 1.46 = 32%
Conclusion: Updated from 55% to 32% probability of on-time finish. Being behind at 50% is meaningful evidence of delay.
Decision: If deadline is flexible, continue. If hard deadline, consider descoping or adding resources.
Quality Checklist
- Prior is justified (base rate + adjustments)
- Likelihoods have reasoning (not just guesses)
- Evidence is diagnostic (LR significantly different from 1)
- Calculation is correct
- Posterior is in reasonable range (not 0% or 100%)
- Assumptions are stated
- Sensitivity analysis performed
- Decision implications clear