Files
2025-11-30 08:38:26 +08:00

364 lines
16 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Kill Criteria & Exit Ramps: Methodology
Advanced techniques for setting kill criteria, avoiding behavioral biases, managing portfolios, and building organizational culture around disciplined stopping.
---
## 1. Sunk Cost Psychology and Behavioral Economics
### Understanding Sunk Cost Fallacy
**Definition**: Tendency to continue project based on past investment rather than future value
**Cognitive mechanisms**:
- **Loss aversion**: Losses feel 2× more painful than equivalent gains (Kahneman & Tversky)
- **Escalation of commitment**: Justifying past decisions by doubling down
- **Status quo bias**: Preference for current state over change
- **Endowment effect**: Overvaluing what we already own/built
**Common rationalizations**:
- "We've invested $2M, can't quit now" (ignoring opportunity cost)
- "Just need a little more time" (moving goalposts)
- "Too close to finishing to stop" (completion bias)
- "Team morale will suffer if we quit" (social pressure)
### Behavioral Interventions
**Pre-commitment devices**:
1. **Signed kill criteria document** before launch (removes discretion)
2. **Third-party decision maker** without sunk cost attachment
3. **Time-locked gates** with automatic NO-GO if criteria not met
4. **Financial caps** ("Will not invest more than $X total")
**Framing techniques**:
1. **Pre-mortem inversion**: "If starting today with $0, would we do this?"
2. **Opportunity cost framing**: "What else could these resources achieve?"
3. **Reverse trial**: "To CONTINUE, we need to prove X" (default = kill)
4. **Outside view**: "What would we advise another company to do?"
**Example**: Company spent $5M on enterprise sales, 3 customers in 2 years (need 50 for viability)
- **Sunk cost framing**: "We've invested $5M, can't stop now"
- **Opportunity cost framing**: "Next $5M could yield 200 SMB customers based on pilot data"
- **Pre-mortem inversion**: "If starting today, would we choose enterprise over SMB?" → No
- **Decision**: Kill enterprise sales, reallocate to SMB
### Quantifying Opportunity Cost
**Formula**: Opportunity Cost = Value of Best Alternative Value of Current Project
**Process**:
1. Identify top 3 alternatives for resources (team, budget)
2. Estimate expected value for each: EV = Σ (Probability × Payoff)
3. Rank by EV / Resource Cost ratio
4. If current project ranks below alternatives → Kill signal
**Example**: SaaS with 3 options
- **Current project** (mobile app): EV = $2M, Cost = 5 engineers × 6 months = 30 eng-months, Ratio = $67k/eng-month
- **Alternative 1** (API platform): EV = $8M, Cost = 30 eng-months, Ratio = $267k/eng-month
- **Alternative 2** (integrations): EV = $5M, Cost = 20 eng-months, Ratio = $250k/eng-month
- **Decision**: Kill mobile app (lowest ratio), reallocate to API platform (highest ratio)
---
## 2. Portfolio Management Frameworks
### Portfolio Ranking Methodologies
**Method 1: Expected Value / Resource Cost**
Formula: `Rank Score = (Revenue × Probability of Success) / (Engineer-Months × Avg Cost)`
**Steps**:
1. Estimate revenue potential for each project (pessimistic, baseline, optimistic)
2. Assign probability of success (use reference class forecasting for calibration)
3. Calculate expected value: EV = Σ (Probability × Revenue)
4. Estimate resource cost (engineer-months, budget, opportunity cost)
5. Rank by EV/Cost ratio
6. Define kill threshold (e.g., bottom 30% of portfolio)
**Method 2: Weighted Scoring Model**
Formula: `Score = Σ (Weight_i × Rating_i)`
**Dimensions** (weights sum to 100%):
- Strategic fit (30%): Alignment with company vision
- Revenue potential (25%): Market size × conversion × pricing
- Probability of success (20%): Team capability, market readiness, technical risk
- Resource efficiency (15%): ROI, payback period, opportunity cost
- Competitive urgency (10%): Time-to-market importance
**Ratings**: 1-5 scale for each dimension
**Example**:
- **Project A**: Strategic fit (4), Revenue (5), Probability (3), Efficiency (4), Urgency (5) → Score = 0.30×4 + 0.25×5 + 0.20×3 + 0.15×4 + 0.10×5 = 4.05
- **Project B**: Strategic fit (3), Revenue (3), Probability (4), Efficiency (2), Urgency (2) → Score = 0.30×3 + 0.25×3 + 0.20×4 + 0.15×2 + 0.10×2 = 3.05
- **Ranking**: A > B
**Method 3: Real Options Analysis**
Treat projects as options (right but not obligation to continue)
**Value of option**:
- **Upside**: High if market/tech uncertainty resolves favorably
- **Downside**: Limited to incremental investment at each gate
- **Flexibility value**: Ability to pivot, expand, or abandon based on new info
**Decision rule**: Continue if `Option Value > Immediate Kill Value`
**Example**: R&D project with 3 gates
- Gate 1 ($50k): Learn if technology feasible (60% chance) → Continue if yes
- Gate 2 ($200k): Learn if market demand exists (40% chance) → Continue if yes
- Gate 3 ($1M): Full launch if both tech + market validated
- **Option value**: Flexibility to kill early if tech/market fails (limits downside)
- **Immediate kill value**: $0 but foregoes learning
**Decision**: Continue through gates (option value > kill value)
### Portfolio Rebalancing Cadence
**Quarterly portfolio review**:
1. Re-rank all projects using latest data
2. Identify projects below threshold (bottom 20-30%)
3. Evaluate kill vs. pivot for bottom performers
4. Reallocate resources from killed projects to top performers
5. Document decisions and communicate transparently
**Trigger-based review** (between quarterly reviews):
- Major market change (competitor launch, regulation, economic shift)
- Significant underperformance vs. projections (>30% variance)
- Resource constraints (hiring freeze, budget cut, key person departure)
- Strategic pivot (new company direction)
---
## 3. Decision Rights Frameworks
### RACI Matrix for Kill Decisions
**Roles**:
- **Responsible (R)**: Gathers data, monitors metrics, presents to decision-maker
- **Accountable (A)**: Makes final kill decision (single person, not committee)
- **Consulted (C)**: Provides input before decision (team, stakeholders)
- **Informed (I)**: Notified after decision (broader org, customers)
**Example**:
- **Kill Criteria Met → Kill Decision**:
- Responsible: Product Manager (gathers data)
- Accountable: Product VP (makes decision)
- Consulted: Engineering Lead, Finance, CEO (input)
- Informed: Team, customers, company (notification)
**Anti-pattern**: "Team decides" (everyone has sunk cost, leads to paralysis)
### Escalation Paths
**Standard path**:
1. **Metrics tracked** → Alert if approaching kill threshold
2. **Project owner evaluates** → Presents data + recommendation to decision authority
3. **Decision authority decides** → GO, NO-GO, or PIVOT
4. **Execute decision** → If kill, wind down within 1 month
**Override path** (use sparingly):
- **Override condition**: Decision authority wants to continue despite kill criteria met
- **Override process**: Written justification, senior approval (e.g., CEO), new kill criteria with shorter timeline
- **Override limit**: Max 1 override per project (prevents repeated goal-post moving)
**Example**: Product VP wants to override kill criteria for feature with 8% adoption (threshold: 10%)
- **Justification**: "Enterprise pilot starting next month, expect 15% adoption within 60 days"
- **New criteria**: "If <12% adoption after 60 days, kill immediately (no further overrides)"
- **CEO approval**: Required for override
### Governance Mechanisms
**Pre-launch approval**:
- Kill criteria document signed by project owner, decision authority, executive sponsor
- Changes to criteria require re-approval by all signatories
**Monitoring dashboard**:
- Real-time metrics vs. kill thresholds
- Traffic light system: Green (>20% above threshold), Yellow (within 20%), Red (below threshold)
- Automated alerts when entering Yellow or Red zones
**Postmortem requirement**:
- All killed projects require postmortem within 2 weeks
- Focus: learning, not blame
- Shared with broader org to normalize killing projects
---
## 4. Postmortem Processes and Learning Culture
### Postmortem Structure
**Timing**: Within 2 weeks of kill decision (while context fresh)
**Participants**: Project team + key stakeholders (5-10 people)
**Facilitator**: Neutral person (not project owner, avoids defensiveness)
**Duration**: 60-90 minutes
**Agenda**:
1. **Project recap** (5 min): Goals, kill criteria, timeline, outcome
2. **What went well?** (15 min): Successes, positive learnings
3. **What went wrong?** (20 min): Root causes of failure (not blame)
4. **What did we learn?** (20 min): Insights, surprises, assumptions invalidated
5. **What would we do differently?** (15 min): Specific changes for future projects
6. **Action items** (10 min): How to apply learnings (process changes, skill gaps, hiring needs)
**Output**: Written doc (2-3 pages) shared with broader team
### Blameless Postmortem Techniques
**Goal**: Learn from failure without blame culture
**Techniques**:
1. **Focus on systems, not individuals**: "Why did our process allow this?" not "Who screwed up?"
2. **Assume good intent**: Team made best decisions with info available at the time
3. **Prime Directive**: "Everyone did the best job they could, given what they knew, their skills, the resources available, and the situation at hand"
4. **"How might we" framing**: Forward-looking, solutions-focused
**Red flags** (blame culture):
- Finger-pointing ("PM should have known...")
- Defensiveness ("It wasn't my fault...")
- Punishment mindset ("Someone should be held accountable...")
- Learning avoidance ("Let's just move on...")
**Example**:
- **Blame framing**: "Why didn't PM validate market demand before building?"
- **Blameless framing**: "How might we improve our discovery process to validate demand earlier with less investment?"
### Normalizing Killing Projects
**Cultural shift**: Killing projects = good (disciplined capital allocation), not bad (failure)
**Messaging**:
- **Positive framing**: "We killed 3 projects this quarter, freeing resources for winners"
- **Celebrate discipline**: Acknowledge teams for recognizing kill criteria and executing quickly
- **Success metrics**: % of portfolio actively managed (killed or pivoted) each quarter (target: 20-30%)
**Leadership behaviors**:
- CEO/VPs publicly discuss projects they killed and why
- Reward PMs who kill projects early (before major resource drain)
- Promote "disciplined stopping" as core competency in performance reviews
**Anti-patterns**:
- Hiding killed projects (creates stigma)
- Only discussing successes (survivorship bias)
- Punishing teams for "failed" projects (discourages risk-taking)
---
## 5. Advanced Topics
### Real Options Theory
**Concept**: Treat uncertain projects as financial options
**Option types**:
1. **Option to defer**: Delay investment until uncertainty resolves
2. **Option to expand**: Scale up if initial results positive
3. **Option to contract**: Scale down if results negative
4. **Option to abandon**: Kill if results very negative
5. **Option to switch**: Pivot to alternative use
**Valuation**: Black-Scholes model adapted for projects
- **Underlying asset**: NPV of project
- **Strike price**: Investment required
- **Volatility**: Uncertainty in outcomes
- **Time to expiration**: Decision window
**Application**: Continue projects with high option value (high upside, limited downside, flexibility)
### Stage-Gate Process Design
**Optimal gate structure**:
- **3-5 gates** for major initiatives
- **Investment increases** by 3-5× at each gate (e.g., $10k → $50k → $200k → $1M)
- **Success criteria tighten** at each gate (higher bar as investment grows)
**Gate design**:
- **Gate 0 (Concept)**: $0-10k, 1-2 weeks, validate problem exists
- **Gate 1 (Discovery)**: $10-50k, 4-8 weeks, validate solution direction
- **Gate 2 (MVP)**: $50-200k, 2-3 months, validate product-market fit
- **Gate 3 (Scale)**: $200k-1M, 6-12 months, validate unit economics
- **Gate 4 (Growth)**: $1M+, ongoing, optimize and scale
**Kill rates by gate** (typical):
- Gate 0 → 1: Kill 50% (cheap to kill, many bad ideas)
- Gate 1 → 2: Kill 30% (learning reveals issues)
- Gate 2 → 3: Kill 20% (product-market fit hard)
- Gate 3 → 4: Kill 10% (unit economics don't work)
### Bayesian Updating for Kill Criteria
**Process**: Update kill probability as new data arrives
**Steps**:
1. **Prior probability** of kill: P(Kill) = initial estimate (e.g., 40% based on historical kill rate)
2. **Likelihood** of data given kill: P(Data | Kill) = how likely is this data if project should be killed?
3. **Likelihood** of data given success: P(Data | Success) = how likely is this data if project will succeed?
4. **Posterior probability** using Bayes theorem: P(Kill | Data) = P(Data | Kill) × P(Kill) / P(Data)
**Example**: SaaS feature with 10% adoption target (kill if <10% after 6 months)
- **Month 3 data**: 7% adoption
- **Prior**: P(Kill) = 40% (4 in 10 similar features killed historically)
- **Likelihood**: P(7% at month 3 | Kill) = 70% (projects that get killed typically have ~7% at halfway point)
- **Likelihood**: P(7% at month 3 | Success) = 30% (successful projects typically have ~12% at halfway point)
- **Posterior**: P(Kill | 7% adoption) = (0.70 × 0.40) / [(0.70 × 0.40) + (0.30 × 0.60)] = 0.28 / 0.46 = 61%
- **Interpretation**: 61% chance this project should be killed (up from 40% prior)
- **Action**: Evaluate closely, prepare pivot/kill plan
### Stopping Rules in Scientific Research
**Clinical trial stopping rules** (adapted for product development):
1. **Futility stopping**: Stop early if interim data shows unlikely to reach success criteria
- **Rule**: If <10% chance of reaching target at current trajectory → Stop
- **Application**: Monitor weekly, project trajectory, stop if trajectory misses by >30%
2. **Efficacy stopping**: Stop early if interim data shows clear success (reallocate resources)
- **Rule**: If >95% confident success criteria will be met → Graduate early
- **Application**: Feature with 25% adoption at month 3 (target: 15% at month 6) → Graduate to core product
3. **Safety stopping**: Stop if harmful unintended consequences detected
- **Rule**: If churn increases >20% or NPS drops >10 points → Stop immediately
- **Application**: New feature causing user confusion, support ticket spike → Kill
**Example**: Mobile app experiment
- **Target**: 20% weekly active users at month 6
- **Month 2 data**: 5% weekly active
- **Trajectory**: Projecting 10% at month 6 (50% below target)
- **Futility analysis**: 95% confidence interval for month 6: 8-12% (entirely below 20% target)
- **Decision**: Invoke futility stopping, kill experiment at month 2 (save 4 months)
---
## Key Principles Summary
1. **Set kill criteria before launch** (remove emotion, politics, sunk cost bias)
2. **Make criteria objective** (numbers, dates, not feelings)
3. **Assign clear decision rights** (single decision-maker, not committee)
4. **Don't move goalposts** (criteria are fixed; changes require formal process)
5. **Sunk costs are irrelevant** (only future value matters)
6. **Kill quickly** (wind down within 1 month, avoid zombie projects)
7. **Opportunity cost > sunk cost** (kill even if "almost done" if better options exist)
8. **Normalize killing** (celebrate discipline, share learnings, remove stigma)
9. **Portfolio thinking** (rank all projects, kill bottom 20-30% regularly)
10. **Learn from kills** (blameless postmortems, apply insights to future projects)
---
## Common Mistakes and Solutions
| Mistake | Symptom | Solution |
|---------|---------|----------|
| **Setting criteria after launch** | Goalposts move when results disappoint | Document criteria in PRD before launch, get sign-off |
| **Subjective criteria** | Debate over "low engagement" | Quantify: "<10% weekly active", not "poor adoption" |
| **Team consensus for kill** | Paralysis, no one wants to kill | Single decision-maker with clear authority |
| **Sunk cost justification** | "Invested $2M, can't quit" | Pre-mortem inversion: "Would we start this today?" |
| **Zombie projects** | Lingering for 6+ months | Wind down within 1 month of kill decision |
| **Stigma around killing** | Teams hide struggles, delay kill | Celebrate kills, share postmortems, normalize stopping |
| **Portfolio inertia** | Same projects year over year | Quarterly ranking + kill bottom 20-30% |
| **No postmortem** | Repeat same mistakes | Require postmortem within 2 weeks, share learnings |