Files
gh-jamesrochabrun-skills-al…/skills/design-brief-generator/references/user_research_methods.md
2025-11-29 18:48:58 +08:00

920 lines
17 KiB
Markdown

# User Research Methods
Comprehensive guide to user research methods for design projects.
---
## Why User Research Matters
**Reduces risk:** Validate assumptions before building
**Saves time:** Fix issues before development
**Builds empathy:** Understand users deeply
**Drives decisions:** Data-informed design choices
**Increases success:** Solutions that actually work
**Rule:** The cost of fixing a problem increases 10x at each stage (design → development → post-launch)
---
## When to Use Which Method
### Discovery Phase (Understanding)
**Goal:** Learn about users, their context, and needs
**Methods:**
- User interviews
- Contextual inquiry
- Diary studies
- Surveys
- Analytics review
**Questions answered:**
- Who are our users?
- What are their goals?
- What problems do they face?
- How do they currently solve this?
---
### Exploration Phase (Ideation)
**Goal:** Generate and evaluate design directions
**Methods:**
- Co-design workshops
- Card sorting
- Concept testing
- Competitive analysis
- Design critiques
**Questions answered:**
- What solutions might work?
- How should we organize information?
- What do users expect?
- What works elsewhere?
---
### Validation Phase (Testing)
**Goal:** Evaluate specific designs before development
**Methods:**
- Usability testing
- Prototype testing
- A/B testing
- First-click tests
- Tree testing
**Questions answered:**
- Can users complete tasks?
- Where do they struggle?
- What works well?
- Which version performs better?
---
### Measurement Phase (Optimization)
**Goal:** Understand performance and iterate
**Methods:**
- Analytics analysis
- Heatmaps
- Session recordings
- Surveys (post-use)
- A/B testing
**Questions answered:**
- How are users actually using it?
- Where do they drop off?
- What's working vs. not?
- How can we improve?
---
## Research Methods Guide
### 1. User Interviews
**What:** One-on-one conversations with users
**When to use:**
- Discovery phase
- Understanding motivations
- Complex topics needing depth
- Building empathy
**Duration:** 30-60 minutes per interview
**Participants:** 5-8 users per user segment
**Format:**
- Semi-structured (planned questions with flexibility)
- Open-ended questions
- Follow-up probes
**Sample questions:**
- "Tell me about a time when..."
- "What were you trying to accomplish?"
- "How did that make you feel?"
- "What would make this better?"
**Best practices:**
- Start broad, get specific
- Ask "why" 5 times
- Listen more than talk
- Avoid leading questions
- Record (with permission)
**Deliverables:**
- Interview transcripts
- Key findings summary
- Quotes and themes
- User journey maps
---
### 2. Contextual Inquiry
**What:** Observing users in their environment
**When to use:**
- Understanding workflows
- Physical context matters
- Complex processes
- B2B/enterprise products
**Duration:** 1-3 hours per session
**Participants:** 3-6 users
**Format:**
- Observe users in natural environment
- Ask questions as they work
- Minimal interruption
- Take photos/notes
**Four principles:**
1. **Context:** Go to the user's environment
2. **Partnership:** User and researcher collaborate
3. **Interpretation:** Develop shared understanding
4. **Focus:** Guide based on research questions
**Best practices:**
- Master-apprentice model
- Think-aloud protocol
- Photograph artifacts
- Note workarounds
- Ask about exceptions
**Deliverables:**
- Observation notes
- Photos of workspace/tools
- Workflow diagrams
- Pain points and opportunities
---
### 3. Surveys
**What:** Quantitative data from many users
**When to use:**
- Large sample sizes needed
- Measuring satisfaction
- Prioritizing features
- Demographic data
**Duration:** 5-10 minutes (completion time)
**Participants:** 100+ for statistical significance
**Types of questions:**
- Multiple choice
- Likert scales (1-5 rating)
- Ranking
- Open-ended (limit to 1-2)
**Best practices:**
- Keep it short (< 10 questions)
- One question per page
- Avoid bias in wording
- Test before sending
- Incentivize completion
**Common formats:**
- NPS (Net Promoter Score)
- CSAT (Customer Satisfaction)
- SUS (System Usability Scale)
- Custom satisfaction surveys
**Deliverables:**
- Response data
- Statistical analysis
- Charts and visualizations
- Key findings report
---
### 4. Usability Testing
**What:** Users attempt tasks with your design
**When to use:**
- Validating designs
- Finding usability issues
- Comparing design options
- Pre-launch validation
**Duration:** 45-60 minutes per session
**Participants:** 5 users per user type
- Nielsen Norman: 5 users find 85% of issues
**Format:**
1. Introduction
2. Background questions
3. Task scenarios (3-5 tasks)
4. Post-task questions
5. Debrief
**Task example:**
"You need to update your credit card information. Show me how you would do that."
**Best practices:**
- Test with prototype or real product
- Think-aloud protocol
- Don't help or lead
- Note time to completion
- Record screen + audio
- Use realistic scenarios
**Metrics to track:**
- Task success rate
- Time on task
- Errors made
- Satisfaction rating
**Deliverables:**
- Usability issues (with severity)
- Task success rates
- Video highlights
- Recommendations
---
### 5. A/B Testing
**What:** Compare two versions with real users
**When to use:**
- Optimizing designs
- Data-driven decisions
- High-traffic pages
- Incremental improvements
**Duration:** 1-4 weeks (until statistical significance)
**Participants:** Varies (need large enough sample)
**Format:**
- Version A (control)
- Version B (variant)
- Randomly assign users
- Measure conversion
**Statistical requirements:**
- Sample size calculator
- 95% confidence level
- 80% statistical power
- Account for baseline conversion
**Best practices:**
- Test one variable at a time
- Run until statistically significant
- Consider external factors (holidays, etc.)
- Don't stop test early
- Account for novelty effect
**Tools:**
- Optimizely
- Google Optimize
- VWO
- Custom implementation
**Deliverables:**
- Winning variant
- Conversion lift
- Statistical confidence
- Implementation recommendation
---
### 6. Card Sorting
**What:** Users organize content into categories
**When to use:**
- Information architecture
- Navigation design
- Taxonomy creation
- Menu structure
**Duration:** 20-30 minutes per session
**Participants:** 15-30 for reliable patterns
**Types:**
**Open card sort:**
- Users create their own categories
- Discovers mental models
- Use early in design
**Closed card sort:**
- Users sort into predefined categories
- Validates existing structure
- Use later in design
**Format:**
1. Give users cards with content items
2. Ask them to group related items
3. Ask them to name groups
4. Analyze patterns
**Tools:**
- Optimal Workshop
- UserZoom
- Miro (remote)
- Physical cards (in-person)
**Deliverables:**
- Dendrogram (similarity matrix)
- Common groupings
- Category names
- IA recommendations
---
### 7. Tree Testing
**What:** Users find content in text-only hierarchy
**When to use:**
- Testing information architecture
- Before visual design
- Validating navigation
- After card sorting
**Duration:** 10-15 minutes
**Participants:** 50+ for statistical significance
**Format:**
1. Present text-only site structure
2. Give user a task
3. User clicks through structure
4. Measure success and path
**Metrics:**
- Success rate
- Directness (optimal path?)
- Time taken
- First click
**Best practices:**
- Test before visual design
- 5-10 tasks
- Realistic task scenarios
- Don't test too deep (3-4 levels)
**Tools:**
- Optimal Workshop
- Treejack
- UserZoom
**Deliverables:**
- Success rates per task
- Problem areas
- Recommended structure changes
- Piecharts showing paths taken
---
### 8. Diary Studies
**What:** Users log experiences over time
**When to use:**
- Long-term behavior
- Infrequent events
- Context switching
- Habitual use
**Duration:** 3-14 days
**Participants:** 10-20 users
**Format:**
- Users log entries daily
- Photo + text entries
- Structured prompts
- Follow-up interview
**Prompts example:**
- "When did you use [product] today?"
- "What were you trying to do?"
- "Photo of where you were"
- "What worked/didn't work?"
**Best practices:**
- Keep it lightweight
- Daily reminders
- Incentivize completion
- Follow up with interview
**Tools:**
- dscout
- Indeemo
- Custom forms
- Mobile apps
**Deliverables:**
- Usage patterns over time
- Context of use
- Pain points
- Opportunity areas
---
### 9. Competitive Analysis
**What:** Evaluate competitor products
**When to use:**
- Understanding landscape
- Identifying opportunities
- Benchmarking
- Inspiration
**Duration:** 1-2 weeks
**Competitors to analyze:** 3-5 direct + 2-3 indirect
**Framework:**
1. Identify competitors
2. Define criteria
3. Evaluate each
4. Synthesize findings
**Criteria examples:**
- Features offered
- User experience quality
- Pricing/model
- Visual design
- Performance
- User reviews
**Best practices:**
- Sign up and use products
- Complete key tasks
- Read user reviews
- Screenshot key flows
- Note strengths/weaknesses
**Deliverables:**
- Competitive matrix
- Feature comparison
- UX evaluation
- Screenshots
- Opportunities
---
### 10. Analytics Review
**What:** Analyze quantitative usage data
**When to use:**
- Understanding current behavior
- Identifying issues
- Measuring impact
- Continuous improvement
**Data sources:**
- Google Analytics
- Product analytics (Mixpanel, Amplitude)
- Heatmaps (Hotjar, Crazy Egg)
- Session recordings
- Error logs
**Key metrics:**
- Page views
- Bounce rate
- Time on page
- Conversion rate
- Drop-off points
- Feature adoption
**Best practices:**
- Define questions first
- Look for anomalies
- Segment users
- Combine with qualitative
- Track over time
**Deliverables:**
- Usage dashboard
- Key metrics report
- Problem areas identified
- Hypotheses for testing
---
## Research Planning
### Research Questions
**Bad questions:**
- "Do users like this design?"
- "Is this easy to use?"
**Good questions:**
- "Can users find the checkout button within 10 seconds?"
- "What prevents users from completing signup?"
- "How do users currently manage their projects?"
### Research Plan Template
```markdown
## Research Plan: [Project Name]
### Background
[Why we're doing this research]
### Goals
- [Goal 1]
- [Goal 2]
### Research Questions
1. [Question 1]
2. [Question 2]
3. [Question 3]
### Method
[Which research method(s)]
### Participants
- **Who:** [User segment]
- **Number:** [How many]
- **Recruiting:** [Where/how]
- **Compensation:** [Incentive]
### Timeline
- Planning: [Dates]
- Recruiting: [Dates]
- Sessions: [Dates]
- Analysis: [Dates]
- Reporting: [Dates]
### Deliverables
- [Deliverable 1]
- [Deliverable 2]
### Team
- Researcher: [Name]
- Designer: [Name]
- Note-taker: [Name]
```
---
## Recruiting Participants
### Screening Criteria
**Example screener questions:**
```markdown
1. How often do you [relevant behavior]?
- Daily (CONTINUE)
- Weekly (CONTINUE)
- Monthly (TERMINATE)
- Never (TERMINATE)
2. What tools do you currently use for [task]?
- [Competitor A] (CONTINUE)
- [Competitor B] (CONTINUE)
- Other: _____ (CONTINUE)
- None (TERMINATE)
3. What is your role?
- [Target role] (CONTINUE)
- Other (TERMINATE)
```
### Recruiting Sources
**Internal:**
- Customer database
- Email campaigns
- In-app recruitment
- Support ticket users
**External:**
- User Interviews (platform)
- Respondent.io
- Ethnio
- Craigslist (with screening)
- Social media
### Incentives
**Typical rates:**
- 30 min interview: $50-75
- 60 min interview: $75-125
- 90 min session: $100-200
- B2B participants: $150-300
- Executives: $300-500
---
## Conducting Research
### Interview Guide Template
```markdown
## [Project] Interview Guide
**Duration:** 60 minutes
### Introduction (5 min)
- Thank you for joining
- Recording with permission
- No right/wrong answers
- Can stop anytime
- Questions before we start?
### Warm-up (5 min)
- Tell me about your role
- Typical day at work
- [Relevant context questions]
### Main Questions (40 min)
**Section 1: Current Behavior**
- Walk me through how you [task]
- What tools do you use?
- What works well?
- What's frustrating?
**Section 2: [Topic]**
- [Questions]
**Section 3: [Topic]**
- [Questions]
### Concept Test (if applicable) (10 min)
- Show prototype/mockup
- What do you see here?
- What would you do first?
- What questions do you have?
### Wrap-up (5 min)
- Anything we didn't cover?
- Final thoughts?
- Thank you + compensation
```
### Facilitation Best Practices
**DO:**
- ✅ Build rapport first
- ✅ Ask open-ended questions
- ✅ Probe deeper ("Tell me more...")
- ✅ Embrace silence
- ✅ Stay neutral
- ✅ Take notes (or have note-taker)
**DON'T:**
- ❌ Lead the witness
- ❌ Explain your design
- ❌ Defend your choices
- ❌ Ask yes/no questions
- ❌ Put words in their mouth
- ❌ Multi-part questions
---
## Analyzing Research
### Affinity Mapping
**Process:**
1. Write observations on sticky notes
2. Group related notes
3. Name themes
4. Identify patterns
5. Prioritize insights
**Tools:**
- Miro
- Mural
- FigJam
- Physical wall + stickies
### Finding Themes
**Look for:**
- Repeated phrases
- Common pain points
- Similar behaviors
- Workarounds
- Strong emotions
**Quotes to capture:**
- Memorable phrases
- Representative of theme
- Emotionally resonant
- Actionable insights
### Prioritizing Insights
**Framework:**
- **Frequency:** How many users mentioned it?
- **Severity:** How big a problem is it?
- **Impact:** How much would solving it help?
**Priority Matrix:**
```
High Impact, High Frequency → P0 (Must fix)
High Impact, Low Frequency → P1 (Should fix)
Low Impact, High Frequency → P2 (Nice to fix)
Low Impact, Low Frequency → P3 (Maybe later)
```
---
## Research Deliverables
### 1. Research Report
**Structure:**
```markdown
# [Project] Research Report
## Executive Summary
- Key findings (3-5 bullets)
- Recommendations (3-5 bullets)
## Background
- Research questions
- Methods
- Participants
## Findings
### Theme 1: [Name]
- Finding
- Evidence (quotes, data)
- Implications
### Theme 2: [Name]
[Repeat]
## Recommendations
1. [Action item with priority]
2. [Action item with priority]
## Next Steps
- [What comes next]
## Appendix
- Full transcripts
- Detailed data
```
### 2. User Personas
**Components:**
- Name and photo
- Demographics
- Goals
- Frustrations
- Behaviors
- Quote
- Scenario
**Keep it:**
- Based on research
- Focused on behaviors
- Referenced in decisions
- Updated regularly
### 3. Journey Maps
**Components:**
- Stages of experience
- User actions
- Thoughts and feelings
- Pain points
- Opportunities
**Format:**
- Timeline across top
- Emotional curve
- Touchpoints
- Behind-the-scenes actions
---
## Research Ethics
### Informed Consent
**Must include:**
- Purpose of research
- What you'll do with data
- Recording disclosure
- Right to stop anytime
- How data will be stored
- Contact information
### Privacy & Data
**Best practices:**
- De-identify data
- Secure storage
- Limited access
- Retention policy
- GDPR/privacy compliance
- Don't share recordings externally
### Participant Wellbeing
**Considerations:**
- Don't cause distress
- Allow breaks
- Respect "I don't know"
- Compensate fairly
- Follow through on promises
---
## Research Checklist for Design Briefs
- [ ] **Research questions** defined
- [ ] **Methods** selected and justified
- [ ] **Participants** criteria specified
- [ ] **Sample size** determined
- [ ] **Timeline** planned
- [ ] **Recruiting** strategy defined
- [ ] **Incentives** budgeted
- [ ] **Interview guide** created
- [ ] **Consent form** prepared
- [ ] **Analysis plan** defined
- [ ] **Deliverables** specified
- [ ] **Stakeholder** review scheduled
---
## Resources
### Books
- "Just Enough Research" - Erika Hall
- "The User Experience Team of One" - Leah Buley
- "Interviewing Users" - Steve Portigal
- "Observing the User Experience" - Goodman, Kuniavsky, Moed
### Tools
- **Recruiting:** User Interviews, Respondent
- **Remote testing:** UserTesting, Lookback, Maze
- **Surveys:** Typeform, Google Forms, Qualtrics
- **Analysis:** Dovetail, Airtable, Notion
- **Card sorting:** Optimal Workshop, UserZoom
### Templates
- [Research Plan Template](https://www.nngroup.com/articles/ux-research-cheat-sheet/)
- [Interview Guide Template](https://userinterviews.com/blog/interview-guide-template)
- [Usability Test Plan](https://www.usability.gov/how-to-and-tools/resources/templates.html)
---
## Quick Reference
### Research Method Selection
| Need | Method | Participants | Duration |
|------|--------|--------------|----------|
| Understand users | Interviews | 5-8 | 30-60 min |
| Observe context | Contextual inquiry | 3-6 | 1-3 hours |
| Large sample | Survey | 100+ | 5-10 min |
| Test usability | Usability test | 5 | 45-60 min |
| Compare options | A/B test | 1000+ | 1-4 weeks |
| Organize content | Card sorting | 15-30 | 20-30 min |
| Test IA | Tree testing | 50+ | 10-15 min |
| Long-term use | Diary study | 10-20 | 3-14 days |
| Competitor insight | Competitive analysis | N/A | 1-2 weeks |
| Usage patterns | Analytics | N/A | Ongoing |
---
**Remember:** Research is not a phase, it's a practice. Build it into every project.