13 KiB
Metrics Tree Template
How to Use This Template
Follow this structure to create a metrics tree for your product or business:
- Start with North Star metric definition
- Apply appropriate decomposition method
- Map action metrics for each input
- Identify leading indicators
- Prioritize experiments using ICE framework
- Output to
metrics-tree.md
Part 1: North Star Metric
Define Your North Star
North Star Metric: [Name of metric]
Definition: [Precise definition including time window] Example: "Number of unique users who complete at least one transaction per week"
Rationale: [Why this metric?]
- ✓ Captures value delivered to customers: [how]
- ✓ Reflects business model: [revenue connection]
- ✓ Measurable and trackable: [data source]
- ✓ Actionable by teams: [who can influence]
Current Value: [Number] as of [Date]
Target: [Goal] by [Date]
North Star Selection Checklist
- Customer value: Does it measure value delivered to customers?
- Business correlation: Does it predict revenue/business success?
- Actionable: Can teams influence it through their work?
- Measurable: Do we have reliable data?
- Not vanity: Does it reflect actual usage/value, not just interest?
- Time-bounded: Does it have a clear time window (daily/weekly/monthly)?
Part 2: Metric Decomposition
Choose the decomposition method that best fits your North Star:
Method 1: Additive Decomposition
Use when: North Star is sum of independent segments
Formula:
North Star = Component A + Component B + Component C + ...
Template:
[North Star] =
+ [New users/customers]
+ [Retained users/customers]
+ [Resurrected users/customers]
+ [Other segment]
Example (SaaS WAU):
Weekly Active Users =
+ New activated users this week (30%)
+ Retained from previous week (60%)
+ Resurrected (inactive→active) (10%)
Method 2: Multiplicative Decomposition
Use when: North Star is product of rates/factors
Formula:
North Star = Factor A × Factor B × Factor C × ...
Template:
[North Star] =
[Total addressable users/visits]
× [Conversion rate at step 1]
× [Conversion rate at step 2]
× [Value per conversion]
Example (E-commerce Revenue):
Monthly Revenue =
Monthly site visitors
× Purchase conversion rate (3%)
× Average order value ($75)
Method 3: Funnel Decomposition
Use when: North Star is end of sequential conversion process
Formula:
North Star = Top of funnel → Step 1 → Step 2 → ... → Final conversion
Template:
[North Star] =
[Total entries]
× [Step 1 conversion %]
× [Step 2 conversion %]
× [Final conversion %]
Example (Paid SaaS Customers):
New paid customers/month =
Free signups
× Activation rate (complete onboarding) (40%)
× Trial start rate (25%)
× Trial→Paid conversion rate (20%)
Math: 1000 signups × 0.4 × 0.25 × 0.2 = 20 paid customers
Method 4: Cohort Decomposition
Use when: Retention is key driver, need to separate acquisition from retention
Formula:
North Star = Σ (Cohort Size_t × Retention Rate_t,n) for all cohorts
Template:
[North Star today] =
[Users from Month 0] × [Month 0 retention rate]
+ [Users from Month 1] × [Month 1 retention rate]
+ ...
+ [Users from Month N] × [Month N retention rate]
Example (Subscription Service MAU):
March Active Users =
Jan signups (500) × Month 2 retention (50%) = 250
+ Feb signups (600) × Month 1 retention (70%) = 420
+ Mar signups (700) × Month 0 retention (100%) = 700
= 1,370 MAU
Part 3: Input Metrics (L2)
For each component in your decomposition, define as input metric:
Input Metric Template
Input Metric 1: [Name]
- Definition: [Precise definition]
- Current value: [Number]
- Target: [Goal]
- Owner: [Team/person]
- Relationship to North Star: [How it affects NS, with estimated coefficient] Example: "Increasing activation rate by 10% → 5% increase in WAU"
Input Metric 2: [Name] [Repeat for 3-5 input metrics]
Validation Questions
- Are all input metrics mutually exclusive? (No double-counting)
- Do they collectively exhaust the North Star? (Nothing missing)
- Can each be owned by a single team?
- Is each measurable with existing/planned instrumentation?
- Are they all at same level of abstraction?
Part 4: Action Metrics (L3)
For each input metric, identify specific user behaviors that drive it:
Action Metrics Template
For Input Metric: [Name of L2 metric]
Action 1: [Specific user behavior]
- Measurement: [How to track it]
- Frequency: [How often it happens]
- Impact: [Estimated effect on input metric]
- Current rate: [% of users doing this]
Action 2: [Another behavior] [Repeat for 3-5 actions per input]
Example (For input metric "Retained Users"):
Action 1: User completes core workflow
- Measurement: Track "workflow_completed" event
- Frequency: 5x per week average for active users
- Impact: Users with 3+ completions have 80% retention vs 20% baseline
- Current rate: 45% of users complete workflow at least once
Action 2: User invites teammate
- Measurement: "invite_sent" event with "invite_accepted" event
- Frequency: 1.2 invites per user on average
- Impact: Users who invite have 90% retention vs 40% baseline
- Current rate: 20% of users send at least one invite
Part 5: Leading Indicators
Identify early signals that predict North Star movement:
Leading Indicator Template
Leading Indicator 1: [Metric name]
- Definition: [What it measures]
- Timing: [How far in advance it predicts] Example: "Predicts week 4 retention"
- Correlation: [Strength of relationship] Example: "r=0.75 with 30-day retention"
- Actionability: [How teams can move it]
- Current value: [Number]
Example:
Leading Indicator: Day 1 Activation Rate
- Definition: % of new users who complete 3 key actions on first day
- Timing: Predicts 7-day and 30-day retention (measured day 1, predicts weeks ahead)
- Correlation: r=0.82 with 30-day retention. Users with Day 1 activation have 70% retention vs 15% without
- Actionability: Improve onboarding flow, reduce time-to-value, send activation nudges
- Current value: 35%
How to Find Leading Indicators
Method 1: Cohort analysis
- Segment users by early behavior (first day, first week)
- Measure long-term outcomes (retention, LTV)
- Find behaviors that predict positive outcomes
Method 2: Correlation analysis
- List all early-funnel metrics
- Calculate correlation with North Star or key inputs
- Select metrics with r > 0.6 and actionable
Method 3: High-performer analysis
- Identify users in top 20% for North Star metric
- Look at their first week/month behavior
- Find patterns that distinguish them from average users
Part 6: Experiment Prioritization
Use ICE framework to prioritize which metrics to improve:
ICE Scoring Template
Impact (1-10): How much will improving this metric affect North Star?
- 10 = Direct, large effect (e.g., 10% improvement → 8% NS increase)
- 5 = Moderate effect (e.g., 10% improvement → 3% NS increase)
- 1 = Small effect (e.g., 10% improvement → 0.5% NS increase)
Confidence (1-10): How certain are we about the relationship?
- 10 = Proven causal relationship with data
- 5 = Correlated, plausible causation
- 1 = Hypothesis, no data yet
Ease (1-10): How easy is it to move this metric?
- 10 = Simple change, 1-2 weeks
- 5 = Moderate effort, 1-2 months
- 1 = Major project, 6+ months
ICE Score = (Impact + Confidence + Ease) / 3
Prioritization Table
| Metric/Experiment | Impact | Confidence | Ease | ICE Score | Rank |
|---|---|---|---|---|---|
| [Experiment 1] | [1-10] | [1-10] | [1-10] | [Avg] | [#] |
| [Experiment 2] | [1-10] | [1-10] | [1-10] | [Avg] | [#] |
| [Experiment 3] | [1-10] | [1-10] | [1-10] | [Avg] | [#] |
Top 3 Experiments
Experiment 1: [Name - highest ICE score]
- Hypothesis: [What we believe will happen]
- Metric to move: [Target metric]
- Expected impact: [Quantified prediction]
- Timeline: [Duration]
- Success criteria: [How we'll know it worked]
Experiment 2: [Second highest] [Repeat structure]
Experiment 3: [Third highest] [Repeat structure]
Part 7: Metric Relationships Diagram
Create visual representation of your metrics tree:
ASCII Tree Format
North Star: [Metric Name] = [Current Value]
│
├─ Input Metric 1: [Name] = [Value]
│ ├─ Action 1.1: [Behavior] = [Rate]
│ ├─ Action 1.2: [Behavior] = [Rate]
│ └─ Action 1.3: [Behavior] = [Rate]
│
├─ Input Metric 2: [Name] = [Value]
│ ├─ Action 2.1: [Behavior] = [Rate]
│ ├─ Action 2.2: [Behavior] = [Rate]
│ └─ Action 2.3: [Behavior] = [Rate]
│
└─ Input Metric 3: [Name] = [Value]
├─ Action 3.1: [Behavior] = [Rate]
├─ Action 3.2: [Behavior] = [Rate]
└─ Action 3.3: [Behavior] = [Rate]
Leading Indicators:
→ [Indicator 1]: Predicts [what] by [timing]
→ [Indicator 2]: Predicts [what] by [timing]
Example (Complete Tree)
North Star: Weekly Active Users = 10,000
│
├─ New Activated Users = 3,000/week (30%)
│ ├─ Complete onboarding: 40% of signups
│ ├─ Connect data source: 25% of signups
│ └─ Invite teammate: 20% of signups
│
├─ Retained Users = 6,000/week (60%)
│ ├─ Use core feature 3+ times: 45% of users
│ ├─ Create content: 30% of users
│ └─ Engage with team: 25% of users
│
└─ Resurrected Users = 1,000/week (10%)
├─ Receive reactivation email: 50% open rate
├─ See new feature announcement: 30% click rate
└─ Get @mentioned by teammate: 40% return rate
Leading Indicators:
→ Day 1 activation rate (35%): Predicts 30-day retention
→ 3 key actions in first session (22%): Predicts weekly usage
Output Format
Create metrics-tree.md with this structure:
# Metrics Tree: [Product/Business Name]
**Date:** [YYYY-MM-DD]
**Owner:** [Team/Person]
**Review Frequency:** [Weekly/Monthly]
## North Star Metric
**Metric:** [Name]
**Current:** [Value] as of [Date]
**Target:** [Goal] by [Date]
**Rationale:** [Why this metric]
## Decomposition Method
[Additive/Multiplicative/Funnel/Cohort]
**Formula:**
[Mathematical relationship]
## Input Metrics (L2)
### 1. [Input Metric Name]
- **Current:** [Value]
- **Target:** [Goal]
- **Owner:** [Team]
- **Impact:** [Effect on NS]
#### Actions (L3):
1. [Action 1]: [Current rate]
2. [Action 2]: [Current rate]
3. [Action 3]: [Current rate]
[Repeat for all input metrics]
## Leading Indicators
1. **[Indicator 1]:** [Definition]
- Timing: [When it predicts]
- Correlation: [Strength]
- Current: [Value]
2. **[Indicator 2]:** [Definition]
[Repeat structure]
## Prioritized Experiments
### Experiment 1: [Name] (ICE: [Score])
- **Hypothesis:** [What we believe]
- **Metric:** [Target]
- **Expected Impact:** [Quantified]
- **Timeline:** [Duration]
- **Success Criteria:** [Threshold]
[Repeat for top 3 experiments]
## Metrics Tree Diagram
[Include ASCII or visual diagram]
## Notes
- [Assumptions made]
- [Data gaps or limitations]
- [Next review date]
Quick Examples by Business Model
SaaS Example (Slack-style)
North Star: Teams sending 100+ messages per week
Decomposition (Additive):
Active Teams = New Active Teams + Retained Active Teams + Resurrected Teams
Input Metrics:
- New active teams: Complete onboarding + hit 100 messages in week 1
- Retained active teams: Hit 100 messages this week and last week
- Resurrected teams: Hit 100 messages this week but not last 4 weeks
Leading Indicators:
- 10 members invited in first day (predicts team activation)
- 50 messages sent in first week (predicts long-term retention)
E-commerce Example
North Star: Monthly Revenue
Decomposition (Multiplicative):
Revenue = Visitors × Purchase Rate × Average Order Value
Input Metrics:
- Monthly unique visitors (owned by Marketing)
- Purchase conversion rate (owned by Product)
- Average order value (owned by Merchandising)
Leading Indicators:
- Add-to-cart rate (predicts purchase)
- Product page views per session (predicts purchase intent)
Marketplace Example (Airbnb-style)
North Star: Nights Booked
Decomposition (Multi-sided):
Nights Booked = (Active Listings × Availability Rate) × (Searches × Booking Rate)
Input Metrics:
- Active host supply: Listings with ≥1 available night
- Guest demand: Unique searches
- Match rate: Searches resulting in booking
Leading Indicators:
- Host completes first listing (predicts long-term hosting)
- Guest saves listings (predicts future booking)