Files
2025-11-30 08:38:26 +08:00

176 lines
6.4 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Prototyping & Pretotyping Experiment Template
## Workflow
```
Prototyping Progress:
- [ ] Step 1: Identify riskiest assumption to test
- [ ] Step 2: Choose pretotype/prototype approach
- [ ] Step 3: Design and build minimum test
- [ ] Step 4: Run experiment and collect data
- [ ] Step 5: Analyze results and decide
```
## Experiment Design Template
### 1. Assumption to Test
**Assumption**: [What are we assuming? E.g., "Users will pay $49/mo for AI-powered analytics"]
**Why risky**: [Why might this be wrong? Impact if wrong?]
**Risk score**: [Probability wrong (1-5) × Impact if wrong (1-5) = Risk (1-25)]
### 2. Test Method
**Approach**: [Pretotype / Paper / Clickable / Coded / MVP]
**Fidelity choice rationale**: [Why this fidelity level? What question does it answer?]
**Estimated cost**: [$X or X hours]
**Timeline**: [X days to build, Y days to test]
### 3. Success Criteria
**Primary metric**: [E.g., "10% landing page → sign-up conversion"]
**Secondary metrics**: [E.g., "50% complete onboarding, 5 min avg session"]
**Minimum sample**: [n=X users/observations]
**Decision rule**:
- **Persevere** (build it): [Metric ≥ X means validated]
- **Pivot** (change direction): [Metric < Y means assumption wrong]
- **Iterate** (refine and re-test): [X > Metric ≥ Y means unclear, need more data]
### 4. Experiment Build
**What we're building**: [Landing page, paper prototype, working feature, etc.]
**Components needed**:
- [ ] [Component 1, e.g., Landing page copy/design]
- [ ] [Component 2, e.g., Sign-up form]
- [ ] [Component 3, e.g., Analytics tracking]
**Fake vs Real**:
- **Faking**: [What appears real but isn't? E.g., "Buy Now button shows 'Coming Soon'"]
- **Real**: [What must actually work? E.g., "Email capture must work"]
### 5. Participant Recruitment
**Target users**: [Who are we testing with? Demographics, behaviors, context]
**Sample size**: [n=X, reasoning: qualitative vs quantitative]
**Recruitment method**: [Ads, existing users, outreach, intercepts]
**Screening**: [How do we ensure target users? Screener questions]
### 6. Data Collection Plan
**Quantitative data**:
| Metric | How measured | Tool | Target |
|--------|--------------|------|--------|
| [Sign-ups] | [Form submissions] | [Google Analytics] | [≥100] |
| [Conversion] | [Sign-ups / Visitors] | [GA] | [≥10%] |
**Qualitative data**:
| Method | N | Questions/Tasks |
|--------|---|-----------------|
| [User interview] | [5-10] | [What problem were you trying to solve? Did prototype help?] |
| [Task observation] | [10] | [Complete checkout, note errors/confusion] |
### 7. Results
**Quantitative**:
| Metric | Target | Actual | Status |
|--------|--------|--------|--------|
| [Sign-ups] | [≥100] | [X] | [✓ / ✗] |
| [Conversion] | [≥10%] | [Y%] | [✓ / ✗] |
**Qualitative**:
- **Observation 1**: [E.g., "7/10 users confused by pricing page"]
- **Observation 2**: [E.g., "All users expected 'Export' feature"]
- **Quote 1**: [User said...]
- **Quote 2**: [User said...]
### 8. Decision
**Decision**: [Persevere / Pivot / Iterate]
**Rationale**: [Why? Which criteria met/not met?]
**Next steps**:
- [ ] [If Persevere: Build MVP with features X, Y, Z]
- [ ] [If Pivot: Test alternative approach A]
- [ ] [If Iterate: Refine prototype addressing issues 1, 2, 3, re-test in 2 weeks]
**Learnings**:
1. [What we learned about assumption]
2. [What surprised us]
3. [What to test next]
---
## Quick Patterns
### Pretotype Methods
**Fake Door Test** (Test demand):
- Build: Landing page "New Feature X - Coming Soon" with "Notify Me" button
- Measure: Click rate, email sign-ups
- Example: "500 visitors, 50 sign-ups (10%) → validates demand"
**Concierge MVP** (Test workflow manually before automating):
- Build: Manual service delivery (no automation)
- Measure: Customer satisfaction, willingness to pay, time spent
- Example: "Manually curate recommendations for 10 users → learn what good looks like before building algorithm"
**Wizard of Oz** (Appear automated, human-powered):
- Build: UI looks automated, humans behind scenes
- Measure: User perception, task success, performance expectations
- Example: "Chatbot UI, humans answering questions → test if users accept chatbot interaction before building NLP"
**Single-Feature MVP** (Test one feature well):
- Build: One core feature, ignore rest
- Measure: Usage, retention, WTP
- Example: "Instagram v1: photo filters only → test if core value enough before building stories/reels"
### Prototype Methods
**Paper Prototype** (Test workflow):
- Build: Hand-drawn screens on paper/cards
- Test: Users "click" on paper, swap screens, observe
- Measure: Task completion, errors, confusion points
- Example: "10 users complete checkout, 3 confused by shipping step → redesign before coding"
**Clickable Prototype** (Test UI/UX):
- Build: Interactive mockup in Figma/InVision (no real code)
- Test: Users complete tasks, measure success/time
- Measure: Completion rate, time, errors, satisfaction
- Example: "20 users, 85% complete task <3 min → validates flow"
**Coded Prototype** (Test feasibility):
- Build: Working code, limited features/data
- Test: Real users, real tasks, measure performance
- Measure: Latency, error rate, scalability, cost
- Example: "Search 10K docs <500ms → validates approach, ready to scale to 10M docs"
### Measurement Approaches
**Quantitative (n=100+)**:
- Conversion rates (landing page → sign-up, sign-up → payment)
- Task completion rates (% who finish checkout)
- Time on task (how long to complete)
- Error rates (clicks on wrong element, form errors)
**Qualitative (n=5-10)**:
- Think-aloud protocol (users narrate thought process)
- Retrospective interview (after task, ask about confusion/delight)
- Observation notes (where they pause, retry, look confused)
- Open-ended feedback (what worked, what didn't)
**Behavioral > Opinions**:
- ✓ "50 clicked 'Buy', 5 completed payment" (behavior)
- ❌ "Users said they'd pay $99" (opinion, unreliable)
---
## Quality Checklist
- [ ] Assumption is risky (high probability wrong × high impact if wrong)
- [ ] Fidelity matches question (not overbuilt)
- [ ] Success criteria set before testing (no moving goalposts)
- [ ] Recruited real target users (not friends/family)
- [ ] Sample size appropriate (n=5-10 qualitative, n=100+ quantitative)
- [ ] Measuring behavior (clicks, conversions), not just opinions
- [ ] Clear decision rule (persevere/pivot/iterate thresholds)
- [ ] Results documented and shared with team