16 KiB
Kill Criteria & Exit Ramps: Methodology
Advanced techniques for setting kill criteria, avoiding behavioral biases, managing portfolios, and building organizational culture around disciplined stopping.
1. Sunk Cost Psychology and Behavioral Economics
Understanding Sunk Cost Fallacy
Definition: Tendency to continue project based on past investment rather than future value
Cognitive mechanisms:
- Loss aversion: Losses feel 2× more painful than equivalent gains (Kahneman & Tversky)
- Escalation of commitment: Justifying past decisions by doubling down
- Status quo bias: Preference for current state over change
- Endowment effect: Overvaluing what we already own/built
Common rationalizations:
- "We've invested $2M, can't quit now" (ignoring opportunity cost)
- "Just need a little more time" (moving goalposts)
- "Too close to finishing to stop" (completion bias)
- "Team morale will suffer if we quit" (social pressure)
Behavioral Interventions
Pre-commitment devices:
- Signed kill criteria document before launch (removes discretion)
- Third-party decision maker without sunk cost attachment
- Time-locked gates with automatic NO-GO if criteria not met
- Financial caps ("Will not invest more than $X total")
Framing techniques:
- Pre-mortem inversion: "If starting today with $0, would we do this?"
- Opportunity cost framing: "What else could these resources achieve?"
- Reverse trial: "To CONTINUE, we need to prove X" (default = kill)
- Outside view: "What would we advise another company to do?"
Example: Company spent $5M on enterprise sales, 3 customers in 2 years (need 50 for viability)
- Sunk cost framing: "We've invested $5M, can't stop now"
- Opportunity cost framing: "Next $5M could yield 200 SMB customers based on pilot data"
- Pre-mortem inversion: "If starting today, would we choose enterprise over SMB?" → No
- Decision: Kill enterprise sales, reallocate to SMB
Quantifying Opportunity Cost
Formula: Opportunity Cost = Value of Best Alternative − Value of Current Project
Process:
- Identify top 3 alternatives for resources (team, budget)
- Estimate expected value for each: EV = Σ (Probability × Payoff)
- Rank by EV / Resource Cost ratio
- If current project ranks below alternatives → Kill signal
Example: SaaS with 3 options
- Current project (mobile app): EV = $2M, Cost = 5 engineers × 6 months = 30 eng-months, Ratio = $67k/eng-month
- Alternative 1 (API platform): EV = $8M, Cost = 30 eng-months, Ratio = $267k/eng-month
- Alternative 2 (integrations): EV = $5M, Cost = 20 eng-months, Ratio = $250k/eng-month
- Decision: Kill mobile app (lowest ratio), reallocate to API platform (highest ratio)
2. Portfolio Management Frameworks
Portfolio Ranking Methodologies
Method 1: Expected Value / Resource Cost
Formula: Rank Score = (Revenue × Probability of Success) / (Engineer-Months × Avg Cost)
Steps:
- Estimate revenue potential for each project (pessimistic, baseline, optimistic)
- Assign probability of success (use reference class forecasting for calibration)
- Calculate expected value: EV = Σ (Probability × Revenue)
- Estimate resource cost (engineer-months, budget, opportunity cost)
- Rank by EV/Cost ratio
- Define kill threshold (e.g., bottom 30% of portfolio)
Method 2: Weighted Scoring Model
Formula: Score = Σ (Weight_i × Rating_i)
Dimensions (weights sum to 100%):
- Strategic fit (30%): Alignment with company vision
- Revenue potential (25%): Market size × conversion × pricing
- Probability of success (20%): Team capability, market readiness, technical risk
- Resource efficiency (15%): ROI, payback period, opportunity cost
- Competitive urgency (10%): Time-to-market importance
Ratings: 1-5 scale for each dimension
Example:
- Project A: Strategic fit (4), Revenue (5), Probability (3), Efficiency (4), Urgency (5) → Score = 0.30×4 + 0.25×5 + 0.20×3 + 0.15×4 + 0.10×5 = 4.05
- Project B: Strategic fit (3), Revenue (3), Probability (4), Efficiency (2), Urgency (2) → Score = 0.30×3 + 0.25×3 + 0.20×4 + 0.15×2 + 0.10×2 = 3.05
- Ranking: A > B
Method 3: Real Options Analysis
Treat projects as options (right but not obligation to continue)
Value of option:
- Upside: High if market/tech uncertainty resolves favorably
- Downside: Limited to incremental investment at each gate
- Flexibility value: Ability to pivot, expand, or abandon based on new info
Decision rule: Continue if Option Value > Immediate Kill Value
Example: R&D project with 3 gates
- Gate 1 ($50k): Learn if technology feasible (60% chance) → Continue if yes
- Gate 2 ($200k): Learn if market demand exists (40% chance) → Continue if yes
- Gate 3 ($1M): Full launch if both tech + market validated
- Option value: Flexibility to kill early if tech/market fails (limits downside)
- Immediate kill value: $0 but foregoes learning
Decision: Continue through gates (option value > kill value)
Portfolio Rebalancing Cadence
Quarterly portfolio review:
- Re-rank all projects using latest data
- Identify projects below threshold (bottom 20-30%)
- Evaluate kill vs. pivot for bottom performers
- Reallocate resources from killed projects to top performers
- Document decisions and communicate transparently
Trigger-based review (between quarterly reviews):
- Major market change (competitor launch, regulation, economic shift)
- Significant underperformance vs. projections (>30% variance)
- Resource constraints (hiring freeze, budget cut, key person departure)
- Strategic pivot (new company direction)
3. Decision Rights Frameworks
RACI Matrix for Kill Decisions
Roles:
- Responsible (R): Gathers data, monitors metrics, presents to decision-maker
- Accountable (A): Makes final kill decision (single person, not committee)
- Consulted (C): Provides input before decision (team, stakeholders)
- Informed (I): Notified after decision (broader org, customers)
Example:
- Kill Criteria Met → Kill Decision:
- Responsible: Product Manager (gathers data)
- Accountable: Product VP (makes decision)
- Consulted: Engineering Lead, Finance, CEO (input)
- Informed: Team, customers, company (notification)
Anti-pattern: "Team decides" (everyone has sunk cost, leads to paralysis)
Escalation Paths
Standard path:
- Metrics tracked → Alert if approaching kill threshold
- Project owner evaluates → Presents data + recommendation to decision authority
- Decision authority decides → GO, NO-GO, or PIVOT
- Execute decision → If kill, wind down within 1 month
Override path (use sparingly):
- Override condition: Decision authority wants to continue despite kill criteria met
- Override process: Written justification, senior approval (e.g., CEO), new kill criteria with shorter timeline
- Override limit: Max 1 override per project (prevents repeated goal-post moving)
Example: Product VP wants to override kill criteria for feature with 8% adoption (threshold: 10%)
- Justification: "Enterprise pilot starting next month, expect 15% adoption within 60 days"
- New criteria: "If <12% adoption after 60 days, kill immediately (no further overrides)"
- CEO approval: Required for override
Governance Mechanisms
Pre-launch approval:
- Kill criteria document signed by project owner, decision authority, executive sponsor
- Changes to criteria require re-approval by all signatories
Monitoring dashboard:
- Real-time metrics vs. kill thresholds
- Traffic light system: Green (>20% above threshold), Yellow (within 20%), Red (below threshold)
- Automated alerts when entering Yellow or Red zones
Postmortem requirement:
- All killed projects require postmortem within 2 weeks
- Focus: learning, not blame
- Shared with broader org to normalize killing projects
4. Postmortem Processes and Learning Culture
Postmortem Structure
Timing: Within 2 weeks of kill decision (while context fresh)
Participants: Project team + key stakeholders (5-10 people)
Facilitator: Neutral person (not project owner, avoids defensiveness)
Duration: 60-90 minutes
Agenda:
- Project recap (5 min): Goals, kill criteria, timeline, outcome
- What went well? (15 min): Successes, positive learnings
- What went wrong? (20 min): Root causes of failure (not blame)
- What did we learn? (20 min): Insights, surprises, assumptions invalidated
- What would we do differently? (15 min): Specific changes for future projects
- Action items (10 min): How to apply learnings (process changes, skill gaps, hiring needs)
Output: Written doc (2-3 pages) shared with broader team
Blameless Postmortem Techniques
Goal: Learn from failure without blame culture
Techniques:
- Focus on systems, not individuals: "Why did our process allow this?" not "Who screwed up?"
- Assume good intent: Team made best decisions with info available at the time
- Prime Directive: "Everyone did the best job they could, given what they knew, their skills, the resources available, and the situation at hand"
- "How might we" framing: Forward-looking, solutions-focused
Red flags (blame culture):
- Finger-pointing ("PM should have known...")
- Defensiveness ("It wasn't my fault...")
- Punishment mindset ("Someone should be held accountable...")
- Learning avoidance ("Let's just move on...")
Example:
- Blame framing: "Why didn't PM validate market demand before building?"
- Blameless framing: "How might we improve our discovery process to validate demand earlier with less investment?"
Normalizing Killing Projects
Cultural shift: Killing projects = good (disciplined capital allocation), not bad (failure)
Messaging:
- Positive framing: "We killed 3 projects this quarter, freeing resources for winners"
- Celebrate discipline: Acknowledge teams for recognizing kill criteria and executing quickly
- Success metrics: % of portfolio actively managed (killed or pivoted) each quarter (target: 20-30%)
Leadership behaviors:
- CEO/VPs publicly discuss projects they killed and why
- Reward PMs who kill projects early (before major resource drain)
- Promote "disciplined stopping" as core competency in performance reviews
Anti-patterns:
- Hiding killed projects (creates stigma)
- Only discussing successes (survivorship bias)
- Punishing teams for "failed" projects (discourages risk-taking)
5. Advanced Topics
Real Options Theory
Concept: Treat uncertain projects as financial options
Option types:
- Option to defer: Delay investment until uncertainty resolves
- Option to expand: Scale up if initial results positive
- Option to contract: Scale down if results negative
- Option to abandon: Kill if results very negative
- Option to switch: Pivot to alternative use
Valuation: Black-Scholes model adapted for projects
- Underlying asset: NPV of project
- Strike price: Investment required
- Volatility: Uncertainty in outcomes
- Time to expiration: Decision window
Application: Continue projects with high option value (high upside, limited downside, flexibility)
Stage-Gate Process Design
Optimal gate structure:
- 3-5 gates for major initiatives
- Investment increases by 3-5× at each gate (e.g., $10k → $50k → $200k → $1M)
- Success criteria tighten at each gate (higher bar as investment grows)
Gate design:
- Gate 0 (Concept): $0-10k, 1-2 weeks, validate problem exists
- Gate 1 (Discovery): $10-50k, 4-8 weeks, validate solution direction
- Gate 2 (MVP): $50-200k, 2-3 months, validate product-market fit
- Gate 3 (Scale): $200k-1M, 6-12 months, validate unit economics
- Gate 4 (Growth): $1M+, ongoing, optimize and scale
Kill rates by gate (typical):
- Gate 0 → 1: Kill 50% (cheap to kill, many bad ideas)
- Gate 1 → 2: Kill 30% (learning reveals issues)
- Gate 2 → 3: Kill 20% (product-market fit hard)
- Gate 3 → 4: Kill 10% (unit economics don't work)
Bayesian Updating for Kill Criteria
Process: Update kill probability as new data arrives
Steps:
- Prior probability of kill: P(Kill) = initial estimate (e.g., 40% based on historical kill rate)
- Likelihood of data given kill: P(Data | Kill) = how likely is this data if project should be killed?
- Likelihood of data given success: P(Data | Success) = how likely is this data if project will succeed?
- Posterior probability using Bayes theorem: P(Kill | Data) = P(Data | Kill) × P(Kill) / P(Data)
Example: SaaS feature with 10% adoption target (kill if <10% after 6 months)
- Month 3 data: 7% adoption
- Prior: P(Kill) = 40% (4 in 10 similar features killed historically)
- Likelihood: P(7% at month 3 | Kill) = 70% (projects that get killed typically have ~7% at halfway point)
- Likelihood: P(7% at month 3 | Success) = 30% (successful projects typically have ~12% at halfway point)
- Posterior: P(Kill | 7% adoption) = (0.70 × 0.40) / [(0.70 × 0.40) + (0.30 × 0.60)] = 0.28 / 0.46 = 61%
- Interpretation: 61% chance this project should be killed (up from 40% prior)
- Action: Evaluate closely, prepare pivot/kill plan
Stopping Rules in Scientific Research
Clinical trial stopping rules (adapted for product development):
-
Futility stopping: Stop early if interim data shows unlikely to reach success criteria
- Rule: If <10% chance of reaching target at current trajectory → Stop
- Application: Monitor weekly, project trajectory, stop if trajectory misses by >30%
-
Efficacy stopping: Stop early if interim data shows clear success (reallocate resources)
- Rule: If >95% confident success criteria will be met → Graduate early
- Application: Feature with 25% adoption at month 3 (target: 15% at month 6) → Graduate to core product
-
Safety stopping: Stop if harmful unintended consequences detected
- Rule: If churn increases >20% or NPS drops >10 points → Stop immediately
- Application: New feature causing user confusion, support ticket spike → Kill
Example: Mobile app experiment
- Target: 20% weekly active users at month 6
- Month 2 data: 5% weekly active
- Trajectory: Projecting 10% at month 6 (50% below target)
- Futility analysis: 95% confidence interval for month 6: 8-12% (entirely below 20% target)
- Decision: Invoke futility stopping, kill experiment at month 2 (save 4 months)
Key Principles Summary
- Set kill criteria before launch (remove emotion, politics, sunk cost bias)
- Make criteria objective (numbers, dates, not feelings)
- Assign clear decision rights (single decision-maker, not committee)
- Don't move goalposts (criteria are fixed; changes require formal process)
- Sunk costs are irrelevant (only future value matters)
- Kill quickly (wind down within 1 month, avoid zombie projects)
- Opportunity cost > sunk cost (kill even if "almost done" if better options exist)
- Normalize killing (celebrate discipline, share learnings, remove stigma)
- Portfolio thinking (rank all projects, kill bottom 20-30% regularly)
- Learn from kills (blameless postmortems, apply insights to future projects)
Common Mistakes and Solutions
| Mistake | Symptom | Solution |
|---|---|---|
| Setting criteria after launch | Goalposts move when results disappoint | Document criteria in PRD before launch, get sign-off |
| Subjective criteria | Debate over "low engagement" | Quantify: "<10% weekly active", not "poor adoption" |
| Team consensus for kill | Paralysis, no one wants to kill | Single decision-maker with clear authority |
| Sunk cost justification | "Invested $2M, can't quit" | Pre-mortem inversion: "Would we start this today?" |
| Zombie projects | Lingering for 6+ months | Wind down within 1 month of kill decision |
| Stigma around killing | Teams hide struggles, delay kill | Celebrate kills, share postmortems, normalize stopping |
| Portfolio inertia | Same projects year over year | Quarterly ranking + kill bottom 20-30% |
| No postmortem | Repeat same mistakes | Require postmortem within 2 weeks, share learnings |