# Prioritization: Advanced Methodologies ## Table of Contents 1. [Advanced Scoring Frameworks](#1-advanced-scoring-frameworks) 2. [Alternative Prioritization Models](#2-alternative-prioritization-models) 3. [Stakeholder Alignment Techniques](#3-stakeholder-alignment-techniques) 4. [Data-Driven Prioritization](#4-data-driven-prioritization) 5. [Roadmap Optimization](#5-roadmap-optimization) 6. [Common Pitfalls and Solutions](#6-common-pitfalls-and-solutions) --- ## 1. Advanced Scoring Frameworks ### RICE Score (Reach × Impact × Confidence / Effort) **Formula**: `Score = (Reach × Impact × Confidence) / Effort` **When to use**: Product backlogs where user reach and confidence matter more than simple impact **Components**: - **Reach**: Users/customers affected per time period (e.g., 1000 users/quarter) - **Impact**: Value per user (1=minimal, 2=low, 3=medium, 4=high, 5=massive) - **Confidence**: How certain are we? (100%=high, 80%=medium, 50%=low) - **Effort**: Person-months (e.g., 2 = 2 engineers for 1 month, or 1 engineer for 2 months) **Example**: - Feature A: (5000 users/qtr × 3 impact × 100% confidence) / 2 effort = **7500 score** - Feature B: (500 users/qtr × 5 impact × 50% confidence) / 1 effort = **1250 score** - **Decision**: Feature A scores 6× higher despite lower per-user impact **Advantages**: More nuanced than 2x2 matrix, accounts for uncertainty **Disadvantages**: Requires estimation of reach (hard for new features) ### ICE Score (Impact × Confidence × Ease) **Formula**: `Score = Impact × Confidence × Ease` **When to use**: Growth experiments, marketing campaigns, quick prioritization **Components**: - **Impact**: Potential value (1-10 scale) - **Confidence**: How sure are we this will work? (1-10 scale) - **Ease**: How easy to implement? (1-10 scale, inverse of effort) **Example**: - Experiment A: 8 impact × 9 confidence × 7 ease = **504 score** - Experiment B: 10 impact × 3 confidence × 5 ease = **150 score** - **Decision**: A scores higher due to confidence, even with lower max impact **Advantages**: Simpler than RICE, no time period needed **Disadvantages**: Multiplicative scoring can exaggerate differences ### Weighted Scoring (Custom Criteria) **When to use**: Complex decisions with multiple evaluation dimensions beyond effort/impact **Process**: 1. Define criteria (e.g., Revenue, Strategic Fit, User Pain, Complexity, Risk) 2. Weight each criterion (sum to 100%) 3. Score each item on each criterion (1-5) 4. Calculate weighted score: `Score = Σ(criterion_score × criterion_weight)` **Example**: | Criteria | Weight | Feature A Score | Feature A Weighted | Feature B Score | Feature B Weighted | |----------|--------|----------------|-------------------|----------------|-------------------| | Revenue | 40% | 4 | 1.6 | 2 | 0.8 | | Strategic | 30% | 5 | 1.5 | 4 | 1.2 | | User Pain | 20% | 3 | 0.6 | 5 | 1.0 | | Complexity | 10% | 2 (low) | 0.2 (penalty) | 4 (high) | 0.4 (penalty) | | **Total** | **100%** | - | **3.9** | - | **3.0** | **Decision**: Feature A scores higher (3.9 vs 3.0) due to revenue and strategic fit **Advantages**: Transparent, customizable to organization's values **Disadvantages**: Can become analysis paralysis with too many criteria ### Kano Model (Customer Satisfaction) **When to use**: Understanding which features delight vs must-have vs don't-matter **Categories**: - **Must-Be (Basic)**: Absence causes dissatisfaction, presence is expected (e.g., product works, no security bugs) - **Performance (Linear)**: More is better, satisfaction increases linearly (e.g., speed, reliability) - **Delighters (Excitement)**: Unexpected features that wow users (e.g., dark mode before it was common) - **Indifferent**: Users don't care either way (e.g., internal metrics dashboard for end users) - **Reverse**: Some users actively dislike (e.g., forced tutorials, animations) **Survey technique**: - Ask functional question: "How would you feel if we had feature X?" (Satisfied / Expected / Neutral / Tolerate / Dissatisfied) - Ask dysfunctional question: "How would you feel if we did NOT have feature X?" - Map responses to category **Prioritization strategy**: 1. **Must-Be first**: Fix broken basics (dissatisfiers) 2. **Delighters for differentiation**: Quick wins that wow 3. **Performance for competitiveness**: Match or exceed competitors 4. **Avoid indifferent/reverse**: Don't waste time **Example**: - Must-Be: Fix crash on login (everyone expects app to work) - Performance: Improve search speed from 2s to 0.5s - Delighter: Add "undo send" for emails (unexpected delight) - Indifferent: Add 50 new color themes (most users don't care) ### Value vs Complexity (Alternative Labels) **When to use**: Same as effort-impact, but emphasizes business value and technical complexity **Axes**: - **X-axis: Complexity** (technical difficulty, dependencies, unknowns) - **Y-axis: Value** (business value, strategic value, user value) **Quadrants** (same concept, different framing): - **High Value, Low Complexity** = Quick Wins (same) - **High Value, High Complexity** = Strategic Investments (same as Big Bets) - **Low Value, Low Complexity** = Nice-to-Haves (same as Fill-Ins) - **Low Value, High Complexity** = Money Pits (same as Time Sinks) **When to use this framing**: Technical audiences respond to "complexity", business audiences to "value" --- ## 2. Alternative Prioritization Models ### MoSCoW (Must/Should/Could/Won't) **When to use**: Fixed-scope projects (e.g., product launch, migration) with deadline constraints **Categories**: - **Must Have**: Non-negotiable, project fails without (e.g., core functionality, security, legal requirements) - **Should Have**: Important but not critical, defer if needed (e.g., nice UX, analytics) - **Could Have**: Desirable if time/budget allows (e.g., polish, edge cases) - **Won't Have (this time)**: Out of scope, revisit later (e.g., advanced features, integrations) **Process**: 1. List all requirements 2. Stakeholders categorize each (force hard choices: only 60% can be Must/Should) 3. Build Must-Haves first, then Should-Haves, then Could-Haves if time remains 4. Communicate Won't-Haves to set expectations **Example (product launch)**: - **Must**: User registration, core workflow, payment processing, security - **Should**: Email notifications, basic analytics, help docs - **Could**: Dark mode, advanced filters, mobile app - **Won't**: Integrations, API, white-labeling (v2 scope) **Advantages**: Forces scope discipline, clear for deadline-driven work **Disadvantages**: Doesn't account for effort (may put high-effort items in Must) ### Cost of Delay (CD3: Cost, Duration, Delay) **When to use**: Time-sensitive decisions where delaying has quantifiable revenue/strategic cost **Formula**: `CD3 Score = Cost of Delay / Duration` - **Cost of Delay**: $/month of delaying this (revenue loss, market share, customer churn) - **Duration**: Months to complete **Example**: - Feature A: $100K/mo delay cost, 2 months duration → **$50K/mo score** - Feature B: $200K/mo delay cost, 5 months duration → **$40K/mo score** - **Decision**: Feature A higher score (faster time-to-value despite lower total CoD) **When to use**: Competitive markets, revenue-critical features, time-limited opportunities (e.g., seasonal) **Advantages**: Explicitly values speed to market **Disadvantages**: Requires estimating revenue impact (often speculative) ### Opportunity Scoring (Jobs-to-be-Done) **When to use**: Understanding which user jobs are underserved (high importance, low satisfaction) **Survey**: - Ask users to rate each "job" on: - **Importance**: How important is this outcome? (1-5) - **Satisfaction**: How well does current solution satisfy? (1-5) **Opportunity Score**: `Importance + max(Importance - Satisfaction, 0)` - Score ranges 0-10 - **>8 = High opportunity** (important but poorly served) - **<5 = Low opportunity** (either unimportant or well-served) **Example**: - Job: "Quickly find relevant past conversations" - Importance: 5 (very important) - Satisfaction: 2 (very dissatisfied) - Opportunity: 5 + (5-2) = **8 (high opportunity)** → prioritize search improvements - Job: "Customize notification sounds" - Importance: 2 (not important) - Satisfaction: 3 (neutral) - Opportunity: 2 + 0 = **2 (low opportunity)** → deprioritize **Advantages**: User-centric, identifies gaps between need and solution **Disadvantages**: Requires user research, doesn't account for effort --- ## 3. Stakeholder Alignment Techniques ### Silent Voting (Avoid Anchoring Bias) **Problem**: First person to score influences others (anchoring bias), HIPPO dominates **Solution**: Everyone scores independently, then discuss **Process**: 1. Each stakeholder writes scores on sticky notes (don't share yet) 2. Reveal all scores simultaneously 3. Discuss discrepancies (why did eng score Effort=5 but product scored Effort=2?) 4. Converge on consensus score **Tools**: Planning poker (common in agile), online voting (Miro, Mural) ### Forced Ranking (Avoid "Everything is High Priority") **Problem**: Stakeholders rate everything 4-5, no differentiation **Solution**: Force stack ranking (only one item can be #1) **Process**: 1. List all items 2. Stakeholders must rank 1, 2, 3, ..., N (no ties allowed) 3. Convert ranks to scores (e.g., top 20% = 5, next 20% = 4, middle 20% = 3, etc.) **Variant: $100 Budget**: - Give stakeholders $100 to "invest" across all items - They allocate dollars based on priority ($30 to A, $25 to B, $20 to C, ...) - Items with most investment are highest priority ### Weighted Stakeholder Input (Account for Expertise) **Problem**: Not all opinions are equal (eng knows effort, sales knows customer pain) **Solution**: Weight scores by expertise domain **Example**: - Effort score = 100% from Engineering (they know effort best) - Impact score = 40% Product + 30% Sales + 30% Customer Success (all know value) **Process**: 1. Define who estimates what (effort, user impact, revenue, etc.) 2. Assign weights (e.g., 60% engineering + 40% product for effort) 3. Calculate weighted average: `Score = Σ(stakeholder_score × weight)` ### Pre-Mortem for Controversial Items **Problem**: Stakeholders disagree on whether item is Quick Win or Time Sink **Solution**: Run pre-mortem to surface hidden risks/assumptions **Process**: 1. Assume item failed spectacularly ("We spent 6 months and it failed") 2. Each stakeholder writes down "what went wrong" (effort blowups, impact didn't materialize) 3. Discuss: Are these risks real? Should we adjust scores or descope? **Example**: - Item: "Build mobile app" (scored Impact=5, Effort=3) - Pre-mortem reveals: "App store approval took 3 months", "iOS/Android doubled effort", "Users didn't download" - **Revised score**: Impact=3 (uncertain), Effort=5 (doubled for two platforms) → Time Sink, defer --- ## 4. Data-Driven Prioritization ### Usage Analytics (Prioritize High-Traffic Areas) **When to use**: Product improvements where usage data is available **Metrics**: - **Page views / Feature usage**: Improve high-traffic areas first (more users benefit) - **Conversion funnel**: Fix drop-off points with biggest impact (e.g., 50% drop at checkout) - **Support tickets**: High ticket volume = high user pain **Example**: - Dashboard page: 1M views/mo, 10% bounce rate → **100K frustrated users** → high impact - Settings page: 10K views/mo, 50% bounce rate → **5K frustrated users** → lower impact - **Decision**: Fix dashboard first (20× more impact) **Advantages**: Objective, quantifies impact **Disadvantages**: Ignores new features (no usage data yet), low-traffic areas may be high-value for specific users ### A/B Test Results (Validate Impact Assumptions) **When to use**: Uncertain impact; run experiment to measure before committing **Process**: 1. Build minimal version of feature (1-2 weeks) 2. A/B test with 10% of users 3. Measure impact (conversion, retention, revenue, NPS) 4. If impact validated, commit to full version; if not, deprioritize **Example**: - Hypothesis: "Adding social login will increase signups 20%" - A/B test: 5% increase (not 20%) - **Decision**: Impact=3 (not 5 as assumed), deprioritize vs other items **Advantages**: Reduces uncertainty, validates assumptions before big bets **Disadvantages**: Requires experimentation culture, time to run tests ### Customer Request Frequency (Vote Counting) **When to use**: Feature requests from sales, support, customers **Metrics**: - **Request count**: Number of unique customers asking - **Revenue at risk**: Total ARR of customers requesting (enterprise vs SMB) - **Churn risk**: Customers threatening to leave without feature **Example**: - Feature A: 50 SMB customers requesting ($5K ARR each) = **$250K ARR** - Feature B: 2 Enterprise customers requesting ($200K ARR each) = **$400K ARR** - **Decision**: Feature B higher impact (more revenue at risk) despite fewer requests **Guardrail**: Don't just count votes (10 vocal users ≠ real demand), weight by revenue/strategic value ### NPS/CSAT Drivers Analysis **When to use**: Understanding which improvements drive customer satisfaction most **Process**: 1. Collect NPS/CSAT scores 2. Ask open-ended: "What's the one thing we could improve?" 3. Categorize feedback (performance, features, support, etc.) 4. Correlate categories with NPS (which issues are mentioned by detractors most?) **Example**: - Detractors (NPS 0-6) mention "slow performance" 80% of time - Passives (NPS 7-8) mention "missing integrations" 60% - **Decision**: Fix performance first (bigger impact on promoter score) --- ## 5. Roadmap Optimization ### Dependency Mapping (Critical Path) **Problem**: Items with dependencies can't start until prerequisites complete **Solution**: Map dependency graph, identify critical path **Process**: 1. List all items with dependencies (A depends on B, C depends on A and D) 2. Draw dependency graph (use tools: Miro, Mural, project management software) 3. Identify critical path (longest chain of dependencies) 4. Parallelize non-dependent work **Example**: - Quick Win A (2 weeks) → Big Bet B (8 weeks) → Quick Win C (1 week) = **11 week critical path** - Quick Win D (2 weeks, no dependencies) can run in parallel with A - **Optimized timeline**: 11 weeks instead of 13 weeks (if sequential) ### Team Velocity and Capacity Planning **Problem**: Overcommitting to more work than team can deliver **Solution**: Use historical velocity to forecast capacity **Process**: 1. Measure past velocity (effort points completed per sprint/quarter) 2. Estimate total capacity (team size × utilization × time period) 3. Don't plan >70-80% of capacity (buffer for unknowns, support, bugs) **Example**: - Team completes 20 effort points/quarter historically - Next quarter roadmap: 30 effort points planned - **Problem**: 150% overcommitted - **Fix**: Cut lowest-priority items (time sinks, fill-ins) to fit 16 effort points (80% of 20) **Guardrail**: If you consistently complete <50% of roadmap, estimation is broken (not just overcommitted) ### Incremental Delivery (Break Big Bets into Phases) **Problem**: Big Bet takes 6 months; no value delivery until end **Solution**: Break into phases that deliver incremental value **Example**: - **Original**: "Rebuild reporting system" (6 months, Effort=5) - **Phased**: - Phase 1: Migrate 3 most-used reports (1 month, Effort=2, Impact=3) - Phase 2: Add drill-down capability (1 month, Effort=2, Impact=4) - Phase 3: Real-time data (2 months, Effort=3, Impact=5) - Phase 4: Custom dashboards (2 months, Effort=3, Impact=3) - **Benefit**: Ship value after 1 month (not 6), can adjust based on feedback **Advantages**: Faster time-to-value, reduces risk, allows pivoting **Disadvantages**: Requires thoughtful phasing (some work can't be incrementalized) ### Portfolio Balancing (Mix of Quick Wins and Big Bets) **Problem**: Roadmap is all quick wins (no strategic depth) or all big bets (no momentum) **Solution**: Balance portfolio across quadrants and time horizons **Target distribution**: - **70% Quick Wins + Fill-Ins** (short-term value, momentum) - **30% Big Bets** (long-term strategic positioning) - OR by time: **Now (0-3 months)**: 60%, **Next (3-6 months)**: 30%, **Later (6-12 months)**: 10% **Example**: - Q1: 5 quick wins + 1 big bet Phase 1 - Q2: 3 quick wins + 1 big bet Phase 2 - Q3: 4 quick wins + 1 new big bet start - **Result**: Consistent value delivery (quick wins) + strategic progress (big bets) --- ## 6. Common Pitfalls and Solutions ### Pitfall 1: Solo Scoring (No Stakeholder Input) **Problem**: PM scores everything alone, misses engineering effort or sales context **Solution**: Multi-stakeholder scoring session (2-hour workshop) **Workshop agenda**: - 0-15min: Align on scoring scales (calibrate with examples) - 15-60min: Silent voting on effort/impact for all items - 60-90min: Discuss discrepancies, converge on consensus - 90-120min: Plot matrix, identify quadrants, sequence roadmap ### Pitfall 2: Analysis Paralysis (Perfect Scores) **Problem**: Spending days debating if item is Effort=3.2 or Effort=3.4 **Solution**: Good-enough > perfect; prioritization is iterative **Guardrail**: Limit scoring session to 2 hours; if still uncertain, default to conservative (higher effort, lower impact) ### Pitfall 3: Ignoring Dependencies **Problem**: Quick Win scored Effort=2, but depends on Effort=5 migration completing first **Solution**: Score standalone effort AND prerequisite effort separately **Example**: - Item: "Add SSO login" (Effort=2 standalone) - Depends on: "Migrate to new auth system" (Effort=5) - **True effort**: 5 (for new roadmaps) or 2 (if migration already planned) ### Pitfall 4: Strategic Override ("CEO Wants It") **Problem**: Exec declares item "high priority" without scoring, bypasses process **Solution**: Make execs participate in scoring, apply same framework **Response**: "Let's score this using our framework so we can compare to other priorities. If it's truly high impact and low effort, it'll naturally rise to the top." ### Pitfall 5: Sunk Cost Fallacy (Continuing Time Sinks) **Problem**: "We already spent 2 months on X, we can't stop now" (even if impact is low) **Solution**: Sunk costs are sunk; evaluate based on future effort/impact only **Decision rule**: If you wouldn't start this project today knowing what you know, stop it now ### Pitfall 6: Neglecting Maintenance (Assuming 100% Project Capacity) **Problem**: Roadmap plans 100% of team time, ignoring support/bugs/tech debt/meetings **Solution**: Reduce capacity by 20-50% for non-project work **Realistic capacity**: - 100% time - 20% support/bugs - 10% meetings - 10% code review/pairing = **60% project capacity** - If team is 5 people × 40 hrs/wk × 12 weeks = 2400 hrs, only 1440 hrs available for roadmap ### Pitfall 7: Ignoring User Research (Opinion-Based Scoring) **Problem**: Impact scores based on team intuition, not user data **Solution**: Validate impact with user research (surveys, interviews, usage data) **Quick validation**: - Survey 50 users: "How important is feature X?" (1-5) - If <50% say 4-5, impact is not as high as assumed - Adjust scores based on data ### Pitfall 8: Scope Creep During Execution **Problem**: Quick Win (Effort=2) grows to Big Bet (Effort=5) during implementation **Solution**: Timebox quick wins; if effort exceeds estimate, cut scope or defer **Guardrail**: "If this takes >1 week, we stop and re-evaluate whether it's still worth it" ### Pitfall 9: Forgetting to Say "No" **Problem**: Roadmap keeps growing (never remove items), becomes unexecutable **Solution**: Explicitly cut time sinks, communicate what you're NOT doing **Communication template**: "We prioritized X, Y, Z based on impact/effort. This means we're NOT doing A, B, C this quarter because [reason]. We'll revisit in [timeframe]." ### Pitfall 10: One-Time Prioritization (Never Re-Score) **Problem**: Scores from 6 months ago are stale (context changed, new data available) **Solution**: Re-score quarterly, adjust roadmap based on new information **Triggers for re-scoring**: - Quarterly planning cycles - Major context changes (new competitor, customer churn, team size change) - Big bets complete (update dependent items' scores) - User research reveals new insights