Files
gh-lyndonkl-claude/skills/design-of-experiments/resources/template.md
2025-11-30 08:38:26 +08:00

14 KiB
Raw Blame History

Design of Experiments - Template

Workflow

Copy this checklist and track your progress:

DOE Template Progress:
- [ ] Step 1: Define experiment objective
- [ ] Step 2: List factors and levels
- [ ] Step 3: Select design type
- [ ] Step 4: Generate design matrix
- [ ] Step 5: Randomize and document protocol
- [ ] Step 6: Finalize experiment plan

Step 1: Define experiment objective

Specify what you're trying to learn (screening, optimization, response surface, robust design), primary response metric(s), and success criteria. See Objective Definition for examples.

Step 2: List factors and levels

Identify all factors (controllable inputs), specify levels for each (2-3 initially), distinguish control vs noise factors, and define measurable responses. See Factor Table Template for structure.

Step 3: Select design type

Based on objective:

Step 4: Generate design matrix

Create run-by-run table with factor settings for each experimental run. See Design Matrix Examples for format.

Step 5: Randomize and document protocol

Randomize run order, specify blocking if needed, detail measurement procedures, and plan replication strategy. See Execution Details for guidance.

Step 6: Finalize experiment plan

Create complete design-of-experiments.md document using Document Structure Template. Self-check with quality criteria in Quality Checklist.


Document Structure Template

Use this structure for the final design-of-experiments.md file:

# Design of Experiments: [Experiment Name]

## 1. Objective

**Goal**: [Screening | Optimization | Response Surface | Robust Design]

**Context**: [1-2 sentences describing the system/process being studied]

**Success Criteria**: [What constitutes a successful experiment? Measurable outcomes.]

**Constraints**:
- Budget: [Maximum number of runs allowed]
- Time: [Deadline or duration per run]
- Resources: [Equipment, personnel, materials]

## 2. Factors and Levels

| Factor | Type | Low Level (-1) | High Level (+1) | Center (0) | Units | Rationale |
|--------|------|----------------|-----------------|------------|-------|-----------|
| A: [Name] | Control | [value] | [value] | [value] | [units] | [Why this factor?] |
| B: [Name] | Control | [value] | [value] | [value] | [units] | [Why this factor?] |
| C: [Name] | Noise | [value] | [value] | - | [units] | [Uncontrollable variation] |

**Factor Selection Rationale**: [Why these factors? Any excluded? Assumptions?]

## 3. Response Variables

| Response | Description | Measurement Method | Target | Units |
|----------|-------------|-------------------|---------|-------|
| Y1: [Name] | [What it measures] | [How measured] | [Maximize/Minimize/Target value] | [units] |
| Y2: [Name] | [What it measures] | [How measured] | [Maximize/Minimize/Target value] | [units] |

**Response Selection Rationale**: [Why these responses? Any tradeoffs?]

## 4. Experimental Design

**Design Type**: [Full Factorial 2^k | Fractional Factorial 2^(k-p) | Plackett-Burman | Central Composite | Box-Behnken]

**Resolution**: [For fractional factorials: III, IV, or V]

**Runs**:
- Design points: [number]
- Center points: [number of replicates at center]
- Total runs: [design + center]

**Design Rationale**: [Why this design? What can/can't it detect?]

## 5. Design Matrix

| Run | Order | Block | A | B | C | Y1 | Y2 | Notes |
|-----|-------|-------|---|---|---|----|----|-------|
| 1 | 5 | 1 | -1 | -1 | -1 | | | |
| 2 | 12 | 1 | +1 | -1 | -1 | | | |
| 3 | 3 | 1 | -1 | +1 | -1 | | | |
| 4 | 8 | 1 | +1 | +1 | -1 | | | |
| 5 | 1 | 2 | -1 | -1 | +1 | | | |
| ... | ... | ... | ... | ... | ... | | | |

**Randomization**: Run order randomized using [method]. Original design point order preserved in "Run" column.

**Blocking**: [If used] Runs blocked by [day/batch/operator/etc.] to control for [nuisance variable].

## 6. Execution Protocol

**Preparation**:
- [ ] [Equipment setup/calibration steps]
- [ ] [Material preparation]
- [ ] [Personnel training]

**Run Procedure**:
1. [Step-by-step protocol for each run]
2. [Factor settings to apply]
3. [Wait/equilibration time]
4. [Response measurement procedure]
5. [Recording method]

**Quality Controls**:
- [Measurement calibration checks]
- [Process stability verification]
- [Outlier detection procedure]

**Timeline**: [Start date, duration per run, expected completion]

## 7. Analysis Plan

**Primary Analysis**:
- Calculate main effects for factors A, B, C
- Calculate 2-way interaction effects (AB, AC, BC)
- Fit linear model: Y = β0 + β1·A + β2·B + β3·C + β12·AB + ...
- ANOVA to test significance (α = 0.05)
- Residual diagnostics (normality, constant variance, independence)

**Graphical Analysis**:
- Main effects plot
- Interaction plot
- Pareto chart of standardized effects
- Residual plots (normal probability, vs fitted, vs order)

**Decision Criteria**:
- Effects significant at p < 0.05 are considered important
- Interaction present if p(interaction) < 0.05
- Optimal settings chosen to [maximize/minimize] Y1 while [constraint on Y2]

**Follow-up**:
- If curvature detected → Run [response surface design]
- If additional factors identified → Run [screening design]
- Confirmation runs: [Number] at predicted optimum settings

## 8. Assumptions and Limitations

**Assumptions**:
- [Linear relationship between factors and response]
- [No strong higher-order interactions]
- [Homogeneous variance across factor space]
- [Errors are independent and normally distributed]
- [Process is stable during experiment]

**Limitations**:
- [Design resolution limits  e.g., 2-way interactions confounded]
- [Factor range restrictions]
- [Measurement precision limits]
- [External validity  generalization beyond tested region]

**Risks**:
- [What could invalidate results?]
- [Mitigation strategies]

## 9. Expected Outcomes

**If screening design**:
- Pareto chart identifying 3-5 critical factors from [N] candidates
- Effect size estimates with confidence intervals
- Shortlist for follow-up optimization experiment

**If optimization design**:
- Optimal factor settings: A = [value], B = [value], C = [value]
- Predicted response at optimum: Y1 = [value] ± [CI]
- Interaction insights: [Which factors interact? How?]

**If response surface**:
- Response surface equation: Y = [polynomial model]
- Contour/surface plots showing optimal region
- Sensitivity analysis showing robustness

**Deliverables**:
- This experiment plan document
- Completed design matrix with results (after execution)
- Analysis report with plots and recommendations

Objective Definition

Screening: Screen 12 software config parameters to identify 3-5 affecting API response time. Success: Reduce candidates 60%+. Constraint: Max 16 runs.

Optimization: Optimize injection molding (temp, pressure, time) to minimize defect rate while cycle time < 45s. Success: < 2% defects (currently 8%). Constraint: Max 20 runs, 2 days.

Response Surface: Map yield vs temperature/pH, find maximum, model curvature. Success: R² > 0.90, optimal region. Constraint: Max 15 runs.


Factor Table Template

Factor Type Low (-1) High (+1) Center (0) Units Rationale
A: Temperature Control 150°C 200°C 175°C °C Literature suggests 150-200 range optimal
B: Pressure Control 50 psi 100 psi 75 psi psi Equipment operates 50-100, nonlinear expected
C: Time Control 10 min 30 min 20 min min Longer times may improve but cost increases
D: Humidity Noise 30% 70% - %RH Uncontrollable environmental variation

Type definitions:

  • Control: Factors you can set deliberately in the experiment
  • Noise: Factors that vary but can't be controlled (for robust design)
  • Held constant: Factors fixed at one level (not in design)

Level selection guidance:

  • 2 levels: Start here for screening/optimization. Detects linear effects and interactions.
  • 3 levels: Add center point to detect curvature. Required for response surface designs.
  • Categorical: Use coded values (-1, +1) for categories (e.g., Supplier A = -1, Supplier B = +1)

Full Factorial Designs

When to use: 2-5 factors, want to estimate all main effects and interactions, budget allows 2^k runs.

Design structure: Test all combinations of factor levels.

Example: 2³ factorial (3 factors, 2 levels each = 8 runs)

Run A B C
1 - - -
2 + - -
3 - + -
4 + + -
5 - - +
6 + - +
7 - + +
8 + + +

Advantages:

  • Estimates all main effects and 2-way/3-way interactions
  • No confounding
  • Maximum precision for given number of factors

Limitations:

  • Runs grow exponentially: 2³ = 8, 2⁴ = 16, 2⁵ = 32, 2⁶ = 64
  • Inefficient for screening (wastes runs on unimportant factors)

Add center points: Replicate 3-5 runs at center (0, 0, 0) to detect curvature and estimate pure error.


Fractional Factorial Designs

When to use: 5+ factors, limited budget, willing to sacrifice some interaction information.

Design structure: Test a fraction (1/2, 1/4, 1/8) of full factorial, deliberately confounding higher-order interactions.

Example: 2⁵⁻¹ design (5 factors, 16 runs instead of 32)

Resolution IV: Main effects clear, 2-way interactions confounded with each other.

Run A B C D E
1 - - - - +
2 + - - - -
3 - + - - -
4 + + - - +
5 - - + - -
... ... ... ... ... ...

Generator: E = ABCD (defining relation: I = ABCDE)

Confounding structure:

  • A confounded with BCDE
  • AB confounded with CDE
  • ABC confounded with DE

Resolution levels:

  • Resolution III: Main effects confounded with 2-way interactions. Use for screening only.
  • Resolution IV: Main effects clear, 2-way confounded with 2-way. Good for screening + some optimization.
  • Resolution V: Main effects and 2-way clear, 2-way confounded with 3-way. Preferred for optimization.

Choosing fraction: Use standard designs (tables available) or design software to ensure desired resolution.


Plackett-Burman Screening

When to use: Screen 8-15 factors with minimal runs, only care about main effects.

Design structure: Orthogonal design with runs = next multiple of 4 above number of factors.

Example: 12-run Plackett-Burman for up to 11 factors

Run A B C D E F G H J K L
1 + + - + + + - - - + -
2 + - + + + - - - + - +
3 - + + + - - - + - + +
... ... ... ... ... ... ... ... ... ... ... ...

Advantages:

  • Very efficient: Screen 11 factors in 12 runs (vs 2048 for full factorial)
  • Main effects estimated independently

Limitations:

  • 2-way interactions completely confounded with main effects
  • Only use when interactions unlikely or unimportant
  • Cannot estimate interactions

Use case: Early-stage screening to reduce 15 candidates to 4-5 for follow-up factorial design.


Design Matrix Examples

Format: Each row = one run. Columns: Run (design point #), Order (randomized sequence), Block (if used), Factors (coded -1/0/+1 or actual values), Responses (blank until execution), Notes (observations).

Run Order A: Temp (°C) B: Press (psi) C: Time (min) Y1: Yield (%) Y2: Cost ($)
1 3 150 50 10
2 7 200 50 10
3 1 150 100 10
4 5 200 100 10

Execution Details

Randomization: Eliminates bias from time trends/drift. Method: (1) List runs, (2) Assign random numbers, (3) Sort by random number = execution order, (4) Document both orders. Exception: Don't randomize hard-to-change factors (use split-plot design, see methodology.md).

Blocking: Use when runs span days/batches/operators. Method: Divide into 2-4 balanced blocks, randomize within each, analyze with block as factor. Example: 16 runs over 2 days → 2 blocks of 8.

Replication: True replication (repeat entire run), repeated measures (multiple measurements per run), or center points (3-5 replicates at center for pure error). Guidance: Always include 3-5 center points for continuous factors.


Quality Checklist

Before finalizing the experiment plan, verify:

Objective & Scope:

  • Goal clearly stated (screening | optimization | response surface | robust)
  • Success criteria are measurable and realistic
  • Constraints documented (runs, time, cost)

Factors:

  • All important factors included
  • Levels span meaningful range (not too narrow, not outside feasible region)
  • Factor types identified (control vs noise)
  • Rationale for each factor documented

Responses:

  • Responses are objective and quantitative
  • Measurement method specified and validated
  • Target direction clear (maximize | minimize | hit target)

Design:

  • Design type appropriate for objective and budget
  • Design resolution adequate (e.g., Resolution IV+ if interactions matter)
  • Run count justified (power analysis or practical limit)
  • Design matrix correct (orthogonal, balanced)

Execution:

  • Randomization method specified
  • Blocking used if runs span nuisance variable levels
  • Replication plan documented (center points, full replicates)
  • Protocol detailed enough for independent execution
  • Timeline realistic

Analysis:

  • Analysis plan specified before data collection
  • Significance level (α) stated
  • Decision criteria clear
  • Residual diagnostics planned
  • Follow-up strategy identified

Assumptions & Risks:

  • Key assumptions stated explicitly
  • Limitations acknowledged (resolution, range, measurement)
  • Risks identified with mitigation plans