286 lines
7.6 KiB
Markdown
286 lines
7.6 KiB
Markdown
---
|
||
name: prioritization-methods
|
||
description: Apply RICE, ICE, MoSCoW, Kano, and Value vs Effort frameworks. Use when prioritizing features, roadmap planning, or making trade-off decisions.
|
||
---
|
||
|
||
# Prioritization Methods & Frameworks
|
||
|
||
## Overview
|
||
|
||
Data-driven frameworks for feature prioritization, backlog ranking, and MVP scoping. Choose the right framework based on your context: data availability, team size, and decision type.
|
||
|
||
## When to Use This Skill
|
||
|
||
**Auto-loaded by agents**:
|
||
- `feature-prioritizer` - For RICE/ICE scoring, MVP scoping, and backlog ranking
|
||
|
||
**Use when you need**:
|
||
- Choosing between competing features
|
||
- Building quarterly roadmaps
|
||
- Backlog prioritization
|
||
- Saying "no" with evidence
|
||
- Clear prioritization decisions
|
||
- Resource allocation decisions
|
||
- MVP scoping decisions
|
||
|
||
---
|
||
|
||
## Seven Core Frameworks
|
||
|
||
### 1. RICE Scoring (Intercom)
|
||
|
||
**Formula**: (Reach × Impact × Confidence) / Effort
|
||
|
||
**Best for**: Large backlogs (20+ items) with quantitative data
|
||
|
||
**Components**:
|
||
- **Reach**: Users impacted per quarter
|
||
- **Impact**: 0.25 (minimal) to 3 (massive)
|
||
- **Confidence**: 50% (low data) to 100% (high data)
|
||
- **Effort**: Person-months to ship
|
||
|
||
**Example**:
|
||
```
|
||
Dark Mode: (10,000 × 2.0 × 0.80) / 1.5 = 10,667
|
||
```
|
||
|
||
**When to use**: Post-PMF with metrics, need defendable priorities, data-driven culture
|
||
|
||
**Template**: `assets/rice-scoring-template.md`
|
||
|
||
---
|
||
|
||
### 2. ICE Scoring (Sean Ellis)
|
||
|
||
**Formula**: (Impact + Confidence + Ease) / 3
|
||
|
||
**Best for**: Quick experiments, early-stage products, limited data
|
||
|
||
**Components** (each 1-10):
|
||
- **Impact**: How much will this move the needle?
|
||
- **Confidence**: How sure are we?
|
||
- **Ease**: How simple to implement?
|
||
|
||
**Example**:
|
||
```
|
||
Email Notifications: (8 + 9 + 7) / 3 = 8.0
|
||
```
|
||
|
||
**When to use**: Growth experiments, startups, need speed over rigor
|
||
|
||
**Template**: `assets/ice-scoring-template.md`
|
||
|
||
---
|
||
|
||
### 3. Value vs Effort Matrix (2×2)
|
||
|
||
**Quadrants**:
|
||
- **Quick Wins** (high value, low effort) - Do first
|
||
- **Big Bets** (high value, high effort) - Strategic
|
||
- **Fill-Ins** (low value, low effort) - If capacity
|
||
- **Time Sinks** (low value, high effort) - Avoid
|
||
|
||
**Best for**: Visual presentations, portfolio planning, quick assessments
|
||
|
||
**When to use**: Clear communication, strategic planning, need visualization
|
||
|
||
**Template**: `assets/value-effort-matrix-template.md`
|
||
|
||
---
|
||
|
||
### 4. MoSCoW Method
|
||
|
||
**Categories**:
|
||
- **Must Have** (60%) - Critical for launch
|
||
- **Should Have** (20%) - Important but not critical
|
||
- **Could Have** (20%) - Nice-to-have
|
||
- **Won't Have** - Explicitly out of scope
|
||
|
||
**Best for**: MVP scoping, release planning, clear scope decisions
|
||
|
||
**When to use**: Fixed timeline, need to cut scope, binary go/no-go decisions
|
||
|
||
**Template**: `assets/moscow-prioritization-template.md`
|
||
|
||
---
|
||
|
||
### 5. Kano Model
|
||
|
||
**Categories**:
|
||
- **Basic Needs (Must-Be)**: Expected, dissatisfiers if absent
|
||
- **Performance Needs**: More is better, linear satisfaction
|
||
- **Excitement Needs (Delighters)**: Unexpected joy
|
||
- **Indifferent**: Users don't care
|
||
- **Reverse**: Users prefer without it
|
||
|
||
**Best for**: Understanding user expectations, competitive positioning, roadmap sequencing
|
||
|
||
**When to use**: Strategic planning, differentiation strategy, multi-release planning
|
||
|
||
**Template**: `assets/kano-model-template.md`
|
||
|
||
---
|
||
|
||
### 6. Weighted Scoring
|
||
|
||
**Process**:
|
||
1. Define criteria (User Value, Revenue, Strategic Fit, Effort)
|
||
2. Assign weights (must sum to 100%)
|
||
3. Score features (1-10) on each criterion
|
||
4. Calculate weighted score
|
||
|
||
**Example**:
|
||
```
|
||
Criteria: User Value 40%, Revenue 30%, Strategic 20%, Ease 10%
|
||
Feature: (8 × 0.40) + (6 × 0.30) + (9 × 0.20) + (5 × 0.10) = 7.3
|
||
```
|
||
|
||
**Best for**: Multiple criteria, complex trade-offs, custom needs
|
||
|
||
**When to use**: Balancing priorities, transparent decisions
|
||
|
||
**Template**: `assets/weighted-scoring-template.md`
|
||
|
||
---
|
||
|
||
### 7. Opportunity Scoring (Jobs-to-be-Done)
|
||
|
||
**Formula**: Importance + Max(Importance - Satisfaction, 0)
|
||
|
||
**Process**:
|
||
1. Identify customer jobs (outcomes, not features)
|
||
2. Survey: Rate importance (1-5) and satisfaction (1-5)
|
||
3. Calculate opportunity = importance + gap
|
||
4. Prioritize high-opportunity jobs (>7.0)
|
||
|
||
**Best for**: Outcome-driven innovation, understanding underserved needs, feature gap analysis
|
||
|
||
**When to use**: JTBD methodology, finding innovation opportunities, validation
|
||
|
||
**Template**: `assets/opportunity-scoring-template.md`
|
||
|
||
---
|
||
|
||
## Choosing the Right Framework
|
||
|
||
**Need speed?** → ICE (fastest)
|
||
|
||
**Have user data?** → RICE (most rigorous)
|
||
|
||
**Visual presentation?** → Value/Effort (clear visualization)
|
||
|
||
**MVP scoping?** → MoSCoW (forces cuts)
|
||
|
||
**User expectations?** → Kano (strategic insights)
|
||
|
||
**Complex criteria?** → Weighted Scoring (custom)
|
||
|
||
**Outcome-focused?** → Opportunity Scoring (JTBD)
|
||
|
||
**Detailed comparison**: `references/framework-selection-guide.md`
|
||
|
||
Complete decision tree, framework comparison table, combining strategies
|
||
|
||
---
|
||
|
||
## Best Practices
|
||
|
||
**1. Be Consistent**
|
||
- Use same framework across team
|
||
- Document assumptions explicitly
|
||
- Update scores as you learn
|
||
|
||
**2. Combine Frameworks**
|
||
- RICE for ranking + Value/Effort for visualization
|
||
- MoSCoW for release + RICE for roadmap
|
||
- Kano for strategy + ICE for tactics
|
||
|
||
**3. Avoid Common Pitfalls**
|
||
- Don't prioritize by HiPPO (Highest Paid Person's Opinion)
|
||
- Don't ignore effort (value alone insufficient)
|
||
- Don't set-and-forget (re-prioritize regularly)
|
||
- Don't game the system (honest scoring)
|
||
|
||
**4. Clear Communication**
|
||
- Show your work (transparent criteria)
|
||
- Visualize priorities clearly
|
||
- Explain trade-offs explicitly
|
||
- Document "why not" for rejected items
|
||
|
||
**5. Iterate and Learn**
|
||
- Track actual vs estimated impact
|
||
- Refine scoring over time
|
||
- Calibrate team estimates
|
||
- Learn from misses
|
||
|
||
---
|
||
|
||
## Templates and References
|
||
|
||
### Assets (Ready-to-Use Templates)
|
||
|
||
Copy-paste these for immediate use:
|
||
- `assets/rice-scoring-template.md` - Reach × Impact × Confidence / Effort
|
||
- `assets/ice-scoring-template.md` - Impact + Confidence + Ease / 3
|
||
- `assets/value-effort-matrix-template.md` - 2×2 visualization
|
||
- `assets/moscow-prioritization-template.md` - Must/Should/Could/Won't
|
||
- `assets/kano-model-template.md` - Expectation analysis
|
||
- `assets/weighted-scoring-template.md` - Custom criteria scoring
|
||
- `assets/opportunity-scoring-template.md` - Jobs-to-be-done prioritization
|
||
|
||
### References (Deep Dives)
|
||
|
||
When you need comprehensive guidance:
|
||
- `references/framework-selection-guide.md` - Choose the right framework, comparison table, combining strategies, decision tree
|
||
|
||
---
|
||
|
||
## Quick Reference
|
||
|
||
```
|
||
Problem: Too many features, limited resources
|
||
Solution: Use prioritization framework
|
||
|
||
Context-Based Selection:
|
||
├─ Lots of data? → RICE
|
||
├─ Need speed? → ICE
|
||
├─ Visual presentation? → Value/Effort
|
||
├─ MVP scoping? → MoSCoW
|
||
├─ User expectations? → Kano
|
||
├─ Complex criteria? → Weighted Scoring
|
||
└─ Outcome-focused? → Opportunity Scoring
|
||
|
||
Always: Document, communicate, iterate
|
||
```
|
||
|
||
---
|
||
|
||
## Resources
|
||
|
||
**Books**:
|
||
- "Intercom on Product Management" (RICE framework)
|
||
- "Hacking Growth" by Sean Ellis (ICE scoring)
|
||
- "Jobs to be Done" by Anthony Ulwick (Opportunity scoring)
|
||
|
||
**Tools**:
|
||
- Airtable/Notion for scoring
|
||
- ProductPlan for roadmaps
|
||
- Aha!, ProductBoard for frameworks
|
||
|
||
**Articles**:
|
||
- "RICE: Simple prioritization for product managers" - Intercom
|
||
- "How to use ICE Scoring" - Sean Ellis
|
||
- "The Kano Model" - UX Magazine
|
||
|
||
---
|
||
|
||
## Related Skills
|
||
|
||
- `roadmap-frameworks` - Turn priorities into roadmaps
|
||
- `specification-techniques` - Spec prioritized features
|
||
- `product-positioning` - Strategic positioning and differentiation
|
||
|
||
---
|
||
|
||
**Key Principle**: Choose one framework, use it consistently, iterate. Don't over-analyze - prioritization should enable decisions, not paralyze them.
|