Files
2025-11-30 08:38:26 +08:00

298 lines
9.4 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Portfolio Roadmapping Bets Methodology
## Table of Contents
1. [Horizon Planning Frameworks](#1-horizon-planning-frameworks)
2. [Bet Sizing Methodologies](#2-bet-sizing-methodologies)
3. [Portfolio Balancing Techniques](#3-portfolio-balancing-techniques)
4. [Dependency Mapping](#4-dependency-mapping)
5. [Exit & Scale Criteria](#5-exit--scale-criteria)
6. [Portfolio Review](#6-portfolio-review)
7. [Anti-Patterns & Fixes](#7-anti-patterns--fixes)
---
## 1. Horizon Planning Frameworks
### McKinsey Three Horizons
**H1: Extend & Defend Core** (70%)
- Timeline: 0-12mo | Risk: Low | Return: 10-30% | Examples: Feature improvements, optimizations
**H2: Build Emerging Businesses** (20%)
- Timeline: 6-24mo | Risk: Medium | Return: 2-5x | Examples: New product lines, geographies
**H3: Create Transformational Options** (10%)
- Timeline: 12-36+mo | Risk: High | Return: 10x+ | Examples: Moonshots, new business models
**Adjustments**: Startup (50/30/20), Enterprise (80/15/5), Scale-up (70/20/10)
### Now-Next-Later
**Now** (Shipping this quarter): >80% confidence, clear reqs, in development
**Next** (Starting 1-2 quarters): ~60% confidence, mostly clear, in planning
**Later** (Future quarters): ~40% confidence, unclear, in research
Use when: Teams uncomfortable with 6-12-24mo planning
### Dual-Track Agile
**Discovery** (Learn): User research, prototypes, experiments → Decide what to build
**Delivery** (Ship): Build, ship, monitor, iterate
**Application**: H1 = delivery, H2 = mix, H3 = discovery. Discovery runs 1-2 sprints ahead.
---
## 2. Bet Sizing Methodologies
### RICE Scoring
**Formula**: (Reach × Impact × Confidence) / Effort
- **Reach**: Users affected per quarter
- **Impact**: 0.25 (minimal) to 3 (massive)
- **Confidence**: 50% (low) to 100% (high)
- **Effort**: Person-months
**Example**: (5000 × 2 × 80%) / 4 = 2000 score
### ICE Scoring
**Formula**: (Impact + Confidence + Ease) / 3 or Impact × Confidence × Ease
- **Impact**: 1-10 scale
- **Confidence**: 1-10 scale
- **Ease**: 1-10 scale (inverse of effort)
Use when: Quick prioritization without reach data
### Effort/Impact Matrix
**Quadrants**:
- High Impact, Low Effort → Quick wins (do first)
- High Impact, High Effort → Strategic (plan carefully, H2/H3)
- Low Impact, Low Effort → Fill-ins (if spare capacity)
- Low Impact, High Effort → Avoid
### Kano Model
**Basic Needs**: Must-have (if missing, dissatisfied) → H1
**Performance Needs**: Linear satisfaction (more is better) → H1/H2
**Delight Needs**: Unexpected wow factors → H2/H3
---
## 3. Portfolio Balancing Techniques
### 70-20-10 Rule
- **70% Core**: Optimize existing (low risk, predictable return)
- **20% Adjacent**: Extend to new (medium risk, substantial return)
- **10% Transformational**: Create new (high risk, breakthrough potential)
**Measure by**: Bet count or effort. **Red flags**: >80% core (too safe), >30% transformational (too risky)
### Risk-Return Diversification
**Low Risk, Low Return** (Core): 80-90% win rate, 1.2-1.5x return
**Medium Risk, Medium Return** (Adjacent): 50-60% win rate, 2-3x return
**High Risk, High Return** (Transformational): 10-30% win rate, 10x+ return
**Portfolio construction**: Combine to achieve desired risk/return profile
### Barbell Strategy
**Structure**: 80-90% very safe + 10-20% very risky, 0% medium
**Rationale**: Safe bets sustain, risky bets create upside, avoid "meh" middle
### Pacing by Cycle Time
**Fast** (days-weeks): A/B tests, experiments → 50%
**Medium** (months): Features, initiatives → 30%
**Slow** (quarters-years): Platform, R&D → 20%
---
## 4. Dependency Mapping
### Critical Path Method
1. List all bets and dependencies
2. Map dependencies (A → B → C)
3. Calculate duration for each path
4. Identify critical path (longest)
5. Accelerate critical path
**Example**: Path A (3mo) → C (4mo) = 7mo ← Critical. Path B (2mo) → C (4mo) = 6mo (1mo slack)
### Dependency Types
- **Technical**: Infrastructure, APIs, data pipelines
- **Learning**: Insights from experiments
- **Strategic**: Prior bet must validate market
- **Resource**: Team availability
### Learning-Based Sequencing
**Pattern**: Small experiment (H1) → Validate → Large bet (H2) → Scale
**Example**: H1: 2-week prototype ($5K) | Exit if CTR <5% | Scale: H2: Full build ($500K) if CTR >10%
---
## 5. Exit & Scale Criteria
### North Star Metric Thresholds
**Example**: North Star = WAU
- Exit: If WAU lift <5% after 60 days, kill
- Scale: If WAU lift >15%, expand to all users
### Staged Funding
**Stage 1** (Seed): $50K → Prototype, 100 users, 20% engagement
- Exit if <20%, fund $200K for alpha if ≥20%
**Stage 2** (Series A): $200K → Alpha, 1000 users, 10% conversion
- Exit if <10%, fund $1M for full build if ≥10%
### Kill Criteria Examples
- **Time**: "If not validated in 90 days, kill"
- **Metric**: "If adoption <5%, kill"
- **Cost**: "If CAC >$100, kill"
- **Strategic**: "If competitor launches first, reassess"
### Scale Criteria Examples
- **Adoption**: "If >20% adopt in 30 days, expand"
- **Engagement**: "If usage >3x baseline, add features"
- **Revenue**: "If ARR >$100K, hire team"
- **Efficiency**: "If LTV/CAC >5, increase budget 3x"
---
## 6. Portfolio Review
### Review Cadence
- **H1**: Monthly (check progress, blockers, kill/pivot/scale)
- **H2**: Quarterly (ready to start? dependencies? promote to H1 or push to H3?)
- **H3**: Semi-annually (still strategic? market shifts? add/kill)
### Kill / Pivot / Persevere / Scale
**For each bet**:
- **Kill**: Criteria not met, no path to success
- **Pivot**: Partially working, adjust approach
- **Persevere**: On track, continue
- **Scale**: Exceeding expectations, double-down
### Portfolio Health Metrics
**Velocity**: Bets shipped/quarter (target: 5-10)
**Win Rate**: % meeting scale criteria (target: 20-40%), % exited (target: 10-30%)
**Impact**: Portfolio contribution to North Star, ROI (target: >3x)
**Balance**: Risk 70/20/10, Horizon 50/30/20
**Red flags**: Win <10% (too risky), Win >80% (too conservative), Exit <5% (not killing), Exit >50% (too risky)
---
## 7. Anti-Patterns & Fixes
### #1: Everything High Priority
**Symptom**: All must-have, no trade-offs
**Fix**: Force-rank (only top 3 high), MoSCoW (20% must, 30% should, 30% could, 20% won't), capacity-constrain
### #2: No Exit Criteria
**Symptom**: Bets continue indefinitely, zombie projects
**Fix**: Set criteria upfront, review monthly, celebrate killing
### #3: All Bets in H1
**Symptom**: Wish list, unrealistic
**Fix**: Capacity-constrain H1, move excess to H2/H3, set expectations
### #4: No H3 Pipeline
**Symptom**: Only H1/H2, no future exploration
**Fix**: Reserve 10-20% for H3, run experiments, refresh quarterly
### #5: All Core, No Transformational
**Symptom**: 100% incremental
**Fix**: Mandate 10% transformational, innovation sprints, measure % revenue from <3yr products (target 20%+)
### #6: Dependencies Ignored
**Symptom**: H2 depends on H1 infrastructure not prioritized
**Fix**: Map dependencies, prioritize blockers, review critical path
### #7: No Review Discipline
**Symptom**: Roadmap created once, never updated
**Fix**: Monthly H1, quarterly portfolio review, version control
### #8: Metrics-Free Bets
**Symptom**: No success metrics, unclear if worked
**Fix**: Require metrics per bet, instrument before ship, review post-launch
### #9: Over-Optimistic Impact
**Symptom**: Every bet "10x potential"
**Fix**: Use baselines, benchmark, risk-adjust (assume 50% success)
### #10: No Portfolio Balance
**Symptom**: All small (busy work) or all large (nothing ships)
**Fix**: Mix sizes (50% S, 30% M, 15% L, 5% XL), cycles (fast/medium/slow), risk (70/20/10)
---
## Quick Reference
### When to Use Each Framework
**Horizon Planning**: McKinsey (classic), Now-Next-Later (adaptive), Dual-Track (continuous)
**Bet Sizing**: RICE (quantitative), ICE (quick), Effort/Impact (visual), Kano (user satisfaction)
**Balancing**: 70-20-10 (risk), Risk-Return (diversification), Barbell (extremes), Pacing (cycles)
**Sequencing**: CPM (critical path), Dependency Matrix (complex), Learning-Based (de-risk)
**Criteria**: North Star (aligned), Staged Funding (VC model), Time/Metric/Cost/Strategic (varied)
**Review**: Monthly/Quarterly/Semi-annual (by horizon), Kill/Pivot/Persevere/Scale (framework)
### Common Patterns
**Product**: H1: Quick wins + strategic features | H2: Major features + platform | H3: Exploratory | 60% incremental, 30% substantial, 10% breakthrough
**Tech**: H1: Stability + migration start | H2: Complete migration + improvements | H3: Next-gen research | 50% maintain, 30% improve, 20% transform
**Innovation**: H1: Scale validated + new tests | H2: Strategic bets + experiments | H3: Moonshots | 70% core, 20% adjacent, 10% transformational
**Marketing**: H1: Optimize proven + test new | H2: Scale winners + brand | H3: Positioning + market entry | 70% performance, 20% growth, 10% brand
### Success Criteria
✓ Strategic theme clear & measurable
✓ Bets sized (S/M/L/XL) & impact quantified (1x/3x/10x)
✓ Sequenced across H1/H2/H3 with dependencies mapped
✓ Exit & scale criteria defined per bet
✓ Portfolio balanced (risk, horizon, size)
✓ Capacity feasible (effort ≤ capacity × 0.8)
✓ Impact ladders to theme (risk-adjusted)
✓ Review cadence established
### Red Flags
❌ No theme → wish list
❌ All "Large" → no prioritization
❌ No exit criteria → zombies
❌ Imbalanced (all core or all moonshots)
❌ Dependencies ignored → blocking
❌ Overcommitted (>80% capacity)
❌ Impact below goal
❌ No review → stale roadmap