371 lines
13 KiB
Markdown
371 lines
13 KiB
Markdown
# Decision Matrix Template
|
||
|
||
## Workflow
|
||
|
||
Copy this checklist and track your progress:
|
||
|
||
```
|
||
Decision Matrix Progress:
|
||
- [ ] Step 1: Frame the decision
|
||
- [ ] Step 2: Identify criteria and assign weights
|
||
- [ ] Step 3: Score alternatives
|
||
- [ ] Step 4: Calculate and analyze results
|
||
- [ ] Step 5: Validate and deliver
|
||
```
|
||
|
||
**Step 1: Frame the decision** - Clarify decision context, list alternatives, identify must-haves. See [Decision Framing](#decision-framing).
|
||
|
||
**Step 2: Identify criteria and assign weights** - Determine what factors matter, assign percentage weights. See [Criteria Identification](#criteria-identification) and [Weighting Techniques](#weighting-techniques).
|
||
|
||
**Step 3: Score alternatives** - Rate each option on each criterion (1-10 scale). See [Scoring Guidance](#scoring-guidance).
|
||
|
||
**Step 4: Calculate and analyze results** - Compute weighted scores, rank options, check sensitivity. See [Matrix Calculation](#matrix-calculation) and [Interpretation](#interpretation).
|
||
|
||
**Step 5: Validate and deliver** - Quality check against [Quality Checklist](#quality-checklist), deliver with recommendation.
|
||
|
||
---
|
||
|
||
## Decision Framing
|
||
|
||
### Input Questions
|
||
|
||
Ask user to clarify:
|
||
|
||
**1. Decision context:**
|
||
- What are we deciding? (Be specific: "Choose CRM platform" not "improve sales")
|
||
- Why now? (Triggering event, deadline, opportunity)
|
||
- What happens if we don't decide or choose wrong?
|
||
|
||
**2. Alternatives:**
|
||
- What are ALL the options we're considering? (Get exhaustive list)
|
||
- Include "do nothing" or status quo as an option if relevant
|
||
- Are these mutually exclusive or can we combine them?
|
||
|
||
**3. Must-have requirements (filters):**
|
||
- Are there absolute dealbreakers? (Budget cap, compliance requirement, technical constraint)
|
||
- Which options fail must-haves and can be eliminated immediately?
|
||
- Distinguish between "must have" (filter) and "nice to have" (criterion)
|
||
|
||
**4. Stakeholders:**
|
||
- Who needs to agree with this decision?
|
||
- Who will be affected by it?
|
||
- Do different stakeholders have different priorities?
|
||
|
||
### Framing Template
|
||
|
||
```markdown
|
||
## Decision Context
|
||
- **Decision:** [Specific choice to be made]
|
||
- **Timeline:** [When decision needed by]
|
||
- **Stakeholders:** [Who needs to agree]
|
||
- **Consequences of wrong choice:** [What we risk]
|
||
|
||
## Alternatives
|
||
1. [Option A name]
|
||
2. [Option B name]
|
||
3. [Option C name]
|
||
4. [Option D name - if applicable]
|
||
5. [Do nothing / Status quo - if applicable]
|
||
|
||
## Must-Have Requirements (Pass/Fail)
|
||
- [ ] [Requirement 1] - All options must meet this
|
||
- [ ] [Requirement 2] - Eliminates options that don't pass
|
||
- [ ] [Requirement 3] - Non-negotiable constraint
|
||
|
||
**Options eliminated:** [List any that fail must-haves]
|
||
**Remaining options:** [List that pass filters]
|
||
```
|
||
|
||
---
|
||
|
||
## Criteria Identification
|
||
|
||
### Process
|
||
|
||
**Step 1: Brainstorm factors**
|
||
|
||
Ask: "What makes one option better than another?"
|
||
|
||
Common categories:
|
||
- **Cost:** Upfront, ongoing, total cost of ownership
|
||
- **Performance:** Speed, quality, reliability, scalability
|
||
- **Risk:** Implementation risk, reversibility, vendor lock-in
|
||
- **Strategic:** Alignment with goals, competitive advantage, future flexibility
|
||
- **Operational:** Ease of use, maintenance, training, support
|
||
- **Stakeholder:** Team preference, customer impact, executive buy-in
|
||
|
||
**Step 2: Validate criteria**
|
||
|
||
Each criterion should be:
|
||
- [ ] **Measurable or scorable** (can assign 1-10 rating)
|
||
- [ ] **Differentiating** (options vary on this dimension)
|
||
- [ ] **Relevant** (actually matters for this decision)
|
||
- [ ] **Independent** (not redundant with other criteria)
|
||
|
||
**Remove:**
|
||
- Criteria where all options score the same (no differentiation)
|
||
- Duplicate criteria that measure same thing
|
||
- Criteria that should be must-haves (pass/fail, not scored)
|
||
|
||
**Step 3: Keep list manageable**
|
||
|
||
- **Ideal:** 4-7 criteria (enough to capture trade-offs, not overwhelming)
|
||
- **Minimum:** 3 criteria (otherwise too simplistic)
|
||
- **Maximum:** 10 criteria (beyond this, hard to weight meaningfully)
|
||
|
||
If you have >10 criteria, group related ones into categories with sub-criteria.
|
||
|
||
### Criteria Template
|
||
|
||
```markdown
|
||
## Evaluation Criteria
|
||
|
||
| # | Criterion | Definition | How We'll Measure |
|
||
|---|-----------|------------|-------------------|
|
||
| 1 | [Name] | [What this measures] | [Data source or scoring approach] |
|
||
| 2 | [Name] | [What this measures] | [Data source or scoring approach] |
|
||
| 3 | [Name] | [What this measures] | [Data source or scoring approach] |
|
||
| 4 | [Name] | [What this measures] | [Data source or scoring approach] |
|
||
| 5 | [Name] | [What this measures] | [Data source or scoring approach] |
|
||
```
|
||
|
||
---
|
||
|
||
## Weighting Techniques
|
||
|
||
### Technique 1: Direct Allocation (Fastest)
|
||
Solo decision or aligned stakeholders. Assign percentages summing to 100%. Start with most important (30-50%), avoid weights <5%, round to 5% increments.
|
||
|
||
**Example:** Cost 30%, Performance 25%, Ease of use 20%, Risk 15%, Team preference 10% = 100%
|
||
|
||
### Technique 2: Pairwise Comparison (Most Rigorous)
|
||
Difficult to weight directly or need justification. Compare each pair ("Is A more important than B?"), tally wins, convert to percentages.
|
||
|
||
**Example:** Cost vs Performance → Performance wins. After all pairs, Performance has 4 wins (40%), Cost has 2 wins (20%), etc.
|
||
|
||
### Technique 3: Stakeholder Averaging (Group Decisions)
|
||
Multiple stakeholders with different priorities. Each assigns weights independently, then average. Large variance reveals disagreement → discuss before proceeding.
|
||
|
||
**Example:** If stakeholders assign Cost weights of 40%, 20%, 30% → Average is 30%, but variance suggests need for alignment discussion.
|
||
|
||
---
|
||
|
||
## Scoring Guidance
|
||
|
||
### Scoring Scale
|
||
|
||
**Use 1-10 scale** (better granularity than 1-5):
|
||
|
||
- **10:** Exceptional, best-in-class
|
||
- **8-9:** Very good, exceeds requirements
|
||
- **6-7:** Good, meets requirements
|
||
- **4-5:** Acceptable, meets minimum
|
||
- **2-3:** Poor, below requirements
|
||
- **1:** Fails, unacceptable
|
||
|
||
**Consistency tips:**
|
||
- Define what 10 means for each criterion before scoring
|
||
- Score all options on one criterion at a time (easier to compare)
|
||
- Use half-points (7.5) if needed for precision
|
||
|
||
### Scoring Process
|
||
|
||
**For objective criteria (cost, speed, measurable metrics):**
|
||
|
||
1. Get actual data (quotes, benchmarks, measurements)
|
||
2. Convert to 1-10 scale using formula:
|
||
- **Lower is better** (cost, time): Score = 10 × (Best value / This value)
|
||
- **Higher is better** (performance, capacity): Score = 10 × (This value / Best value)
|
||
|
||
**Example (Cost - lower is better):**
|
||
- Option A: $50K → Score = 10 × ($30K / $50K) = 6.0
|
||
- Option B: $30K → Score = 10 × ($30K / $30K) = 10.0
|
||
- Option C: $40K → Score = 10 × ($30K / $40K) = 7.5
|
||
|
||
**For subjective criteria (ease of use, team preference):**
|
||
|
||
1. Define what 10, 7, and 4 look like for this criterion
|
||
2. Score relative to those anchors
|
||
3. Document reasoning/assumptions
|
||
|
||
**Example (Ease of Use):**
|
||
- 10 = No training needed, intuitive UI, users productive day 1
|
||
- 7 = 1-week training, moderate learning curve
|
||
- 4 = Significant training (1 month), complex UI
|
||
|
||
**Calibration questions:**
|
||
- Would I bet money on this score being accurate?
|
||
- Is this score relative to alternatives or absolute?
|
||
- What would change this score by ±2 points?
|
||
|
||
### Scoring Template
|
||
|
||
```markdown
|
||
## Scoring Matrix
|
||
|
||
| Option | Criterion 1 (Weight%) | Criterion 2 (Weight%) | Criterion 3 (Weight%) | Criterion 4 (Weight%) |
|
||
|--------|-----------------------|-----------------------|-----------------------|-----------------------|
|
||
| Option A | [Score] | [Score] | [Score] | [Score] |
|
||
| Option B | [Score] | [Score] | [Score] | [Score] |
|
||
| Option C | [Score] | [Score] | [Score] | [Score] |
|
||
|
||
**Data sources and assumptions:**
|
||
- Criterion 1: [Where scores came from, what assumptions]
|
||
- Criterion 2: [Where scores came from, what assumptions]
|
||
- Criterion 3: [Where scores came from, what assumptions]
|
||
- Criterion 4: [Where scores came from, what assumptions]
|
||
```
|
||
|
||
---
|
||
|
||
## Matrix Calculation
|
||
|
||
### Calculation Process
|
||
|
||
**For each option:**
|
||
1. Multiply criterion score by criterion weight
|
||
2. Sum all weighted scores
|
||
3. This is the option's total score
|
||
|
||
**Formula:** Total Score = Σ (Criterion Score × Criterion Weight)
|
||
|
||
**Example:**
|
||
|
||
| Option | Cost (30%) | Performance (40%) | Risk (20%) | Ease (10%) | **Total** |
|
||
|--------|-----------|------------------|-----------|-----------|---------|
|
||
| Option A | 7 × 0.30 = 2.1 | 9 × 0.40 = 3.6 | 6 × 0.20 = 1.2 | 8 × 0.10 = 0.8 | **7.7** |
|
||
| Option B | 9 × 0.30 = 2.7 | 6 × 0.40 = 2.4 | 8 × 0.20 = 1.6 | 6 × 0.10 = 0.6 | **7.3** |
|
||
| Option C | 5 × 0.30 = 1.5 | 8 × 0.40 = 3.2 | 7 × 0.20 = 1.4 | 9 × 0.10 = 0.9 | **7.0** |
|
||
|
||
**Winner: Option A (7.7)**
|
||
|
||
### Final Matrix Template
|
||
|
||
```markdown
|
||
## Decision Matrix Results
|
||
|
||
| Option | [Criterion 1] ([W1]%) | [Criterion 2] ([W2]%) | [Criterion 3] ([W3]%) | [Criterion 4] ([W4]%) | **Weighted Total** | **Rank** |
|
||
|--------|----------------------|----------------------|----------------------|----------------------|-------------------|----------|
|
||
| [Option A] | [S] ([S×W1]) | [S] ([S×W2]) | [S] ([S×W3]) | [S] ([S×W4]) | **[Total]** | [Rank] |
|
||
| [Option B] | [S] ([S×W1]) | [S] ([S×W2]) | [S] ([S×W3]) | [S] ([S×W4]) | **[Total]** | [Rank] |
|
||
| [Option C] | [S] ([S×W1]) | [S] ([S×W2]) | [S] ([S×W3]) | [S] ([S×W4]) | **[Total]** | [Rank] |
|
||
|
||
**Weights:** [Criterion 1] ([W1]%), [Criterion 2] ([W2]%), [Criterion 3] ([W3]%), [Criterion 4] ([W4]%)
|
||
|
||
**Scoring scale:** 1-10 (10 = best)
|
||
```
|
||
|
||
---
|
||
|
||
## Interpretation
|
||
|
||
### Analysis Checklist
|
||
|
||
After calculating scores, analyze:
|
||
|
||
**1. Clear winner vs close call**
|
||
- [ ] **Margin >10%:** Clear winner, decision is robust
|
||
- [ ] **Margin 5-10%:** Moderate confidence, validate assumptions
|
||
- [ ] **Margin <5%:** Toss-up, need more data or stakeholder discussion
|
||
|
||
**2. Dominant criterion check**
|
||
- [ ] Does one criterion drive entire decision? (accounts for >50% of score difference)
|
||
- [ ] Is that appropriate or is weight too high?
|
||
|
||
**3. Surprising results**
|
||
- [ ] Does the winner match gut instinct?
|
||
- [ ] If not, what does the matrix reveal? (Trade-off you hadn't considered)
|
||
- [ ] Or are weights/scores wrong?
|
||
|
||
**4. Sensitivity questions**
|
||
- [ ] If we swapped top two criterion weights, would winner change?
|
||
- [ ] If we adjusted one score by ±1 point, would winner change?
|
||
- [ ] Which scores are most uncertain? (Could they change with more data)
|
||
|
||
### Recommendation Template
|
||
|
||
```markdown
|
||
## Recommendation
|
||
|
||
**Recommended Option:** [Option name] (Score: [X.X])
|
||
|
||
**Rationale:**
|
||
- [Option] scores highest overall ([X.X] vs [Y.Y] for runner-up)
|
||
- Key strengths: [What it excels at based on criterion scores]
|
||
- Acceptable trade-offs: [Where it scores lower but weight is low enough]
|
||
|
||
**Key Trade-offs:**
|
||
- **Winner:** Strong on [Criterion A, B] ([X]% of total weight)
|
||
- **Runner-up:** Strong on [Criterion C] but weaker on [Criterion A]
|
||
- **Decision driver:** [Criterion A] matters most ([X]%), where [Winner] excels
|
||
|
||
**Confidence Level:**
|
||
- [ ] **High (>10% margin):** Decision is robust to reasonable assumption changes
|
||
- [ ] **Moderate (5-10% margin):** Sensitive to [specific assumption], recommend validating
|
||
- [ ] **Low (<5% margin):** Effectively a tie, consider [additional data needed] or [stakeholder input]
|
||
|
||
**Sensitivity:**
|
||
- [Describe any sensitivity - e.g., "If Risk weight increased from 20% to 35%, Option B would win"]
|
||
|
||
**Next Steps:**
|
||
1. [Immediate action - e.g., "Get final pricing from vendor"]
|
||
2. [Validation - e.g., "Confirm technical feasibility with engineering"]
|
||
3. [Communication - e.g., "Present to steering committee by [date]"]
|
||
```
|
||
|
||
---
|
||
|
||
## Quality Checklist
|
||
|
||
Before delivering, verify:
|
||
|
||
**Decision framing:**
|
||
- [ ] Decision is specific and well-defined
|
||
- [ ] All viable alternatives included
|
||
- [ ] Must-haves clearly separated from nice-to-haves
|
||
- [ ] Stakeholders identified
|
||
|
||
**Criteria:**
|
||
- [ ] 3-10 criteria (enough to capture trade-offs, not overwhelming)
|
||
- [ ] Each criterion is measurable/scorable
|
||
- [ ] Criteria differentiate between options (not all scored the same)
|
||
- [ ] No redundancy between criteria
|
||
- [ ] Weights sum to 100%
|
||
- [ ] Weight distribution reflects true priorities
|
||
|
||
**Scoring:**
|
||
- [ ] Scores use consistent 1-10 scale
|
||
- [ ] Objective criteria based on data (not guesses)
|
||
- [ ] Subjective criteria have clear definitions/anchors
|
||
- [ ] Assumptions and data sources documented
|
||
- [ ] Scores are defensible (could explain to stakeholder)
|
||
|
||
**Analysis:**
|
||
- [ ] Weighted scores calculated correctly
|
||
- [ ] Options ranked by total score
|
||
- [ ] Sensitivity analyzed (close calls identified)
|
||
- [ ] Recommendation includes rationale and trade-offs
|
||
- [ ] Next steps identified
|
||
|
||
**Communication:**
|
||
- [ ] Matrix table is clear and readable
|
||
- [ ] Weights shown in column headers
|
||
- [ ] Weighted scores shown (not just raw scores)
|
||
- [ ] Recommendation stands out visually
|
||
- [ ] Assumptions and limitations noted
|
||
|
||
---
|
||
|
||
## Common Pitfalls
|
||
|
||
| Pitfall | Fix |
|
||
|---------|-----|
|
||
| **Too many criteria (>10)** | Consolidate related criteria into categories |
|
||
| **Redundant criteria** | Combine criteria that always score the same |
|
||
| **Arbitrary weights** | Use pairwise comparison or stakeholder discussion |
|
||
| **Scores are guesses** | Gather data for objective criteria, define anchors for subjective |
|
||
| **Confirmation bias** | Weight criteria BEFORE scoring options |
|
||
| **Ignoring sensitivity** | Always check if small changes flip the result |
|
||
| **False precision** | Match precision to confidence level |
|
||
| **Missing "do nothing"** | Include status quo as an option to evaluate |
|