16 KiB
Design of Experiments - Advanced Methodology
Workflow
Copy this checklist for advanced DOE cases:
Advanced DOE Progress:
- [ ] Step 1: Assess complexity and choose advanced technique
- [ ] Step 2: Design experiment using specialized method
- [ ] Step 3: Plan execution with advanced considerations
- [ ] Step 4: Analyze with appropriate statistical methods
- [ ] Step 5: Iterate or confirm findings
Step 1: Assess complexity
Identify which advanced technique applies: screening 8+ factors, response surface with curvature, robust design, mixture constraints, hard-to-change factors, or irregular factor space. See technique selection criteria below.
Step 2: Design experiment
Apply specialized design method from sections: Screening Designs, Response Surface Methodology, Taguchi Methods, Optimal Designs, Mixture Experiments, or Split-Plot Designs.
Step 3: Plan execution
Address advanced considerations: blocking for nuisance variables, replication for variance estimation, center points for curvature detection, and sequential strategies. See Sequential Experimentation for iterative approaches.
Step 4: Analyze
Use appropriate analysis for design type: effect estimation, ANOVA, regression modeling, response surface equations, contour plots, and residual diagnostics. See Analysis Techniques.
Step 5: Iterate or confirm
Based on findings, run confirmation experiments, refine factor ranges, add center/axial points for RSM, or screen additional factors. See Sequential Experimentation.
1. Screening Designs
When to use: 8-30 candidate factors, limited experimental budget, goal is to identify 3-5 vital factors for follow-up optimization.
Plackett-Burman Designs
Structure: Orthogonal designs with runs = multiple of 4. Screens k factors in k+1 runs.
Standard designs:
- 12 runs → screen up to 11 factors
- 16 runs → screen up to 15 factors
- 20 runs → screen up to 19 factors
- 24 runs → screen up to 23 factors
Example: 12-run Plackett-Burman generator matrix:
Run 1: + + - + + + - - - + -
Subsequent runs: Cycle previous row left, last run is all minus
Analysis: Fit linear model Y = β₀ + β₁X₁ + β₂X₂ + ... + βₖXₖ. Rank factors by |βᵢ|. Select top 3-5 for optimization. Pareto chart: cumulative % variance explained.
Limitation: Main effects confounded with 2-way interactions. Only valid if interactions negligible (sparsity-of-effects principle).
Fractional Factorial Screening
When to use: 5-8 factors, need to estimate some 2-way interactions, Resolution IV or V required.
Common designs:
- 2⁵⁻¹ (Resolution V): 16 runs, 5 factors. Main effects and 2-way interactions clear. Generator: I = ABCDE.
- 2⁶⁻² (Resolution IV): 16 runs, 6 factors. Main effects clear, 2-way confounded with 2-way. Generators: I = ABCE, I = BCDF.
- 2⁷⁻³ (Resolution IV): 16 runs, 7 factors. Generators: I = ABCD, I = ABEF, I = ACEG.
Confounding analysis: Use defining relation to determine alias structure. Example for 2⁵⁻¹ with I = ABCDE:
- A aliased with BCDE
- AB aliased with CDE
- ABC aliased with DE
Fold-over technique: If screening reveals ambiguous confounding, run fold-over design (flip all signs) to de-alias. 16 runs + 16 fold-over = 32 runs = full 2⁵ factorial.
2. Response Surface Methodology
When to use: 2-5 factors already identified as important, need to find optimum, expect curvature (quadratic relationship).
Central Composite Design (CCD)
Structure: Factorial points + axial points + center points
Components:
- Factorial points: 2^k corner points (±1 for all factors)
- Axial points: 2k points on axes (±α, 0, 0, ...) where α determines rotatability
- Center points: 3-5 replicates at origin (0, 0, ..., 0)
Total runs: 2^k + 2k + nc (nc = number of center points)
Example: CCD for 3 factors (8 + 6 + 5 = 19 runs):
| Type | X₁ | X₂ | X₃ |
|---|---|---|---|
| Factorial | -1 | -1 | -1 |
| ... | ... | ... | ... |
| Axial | -α | 0 | 0 |
| Axial | +α | 0 | 0 |
| Axial | 0 | -α | 0 |
| Axial | 0 | +α | 0 |
| Axial | 0 | 0 | -α |
| Axial | 0 | 0 | +α |
| Center | 0 | 0 | 0 |
Rotatability: Choose α = (2^k)^(1/4) for equal prediction variance at equal distance from center. For 3 factors: α = 1.682.
Model: Fit quadratic: Y = β₀ + Σβᵢxᵢ + Σβᵢᵢxᵢ² + Σβᵢⱼxᵢxⱼ
Analysis: Canonical analysis to find stationary point (maximum, minimum, or saddle). Ridge analysis if optimum outside design region.
Box-Behnken Design
Structure: 3-level design that avoids extreme corners (all factors at ±1 simultaneously).
Advantages: Fewer runs than CCD, safer when extreme combinations may damage equipment or produce out-of-spec product.
Example: Box-Behnken for 3 factors (12 + 3 = 15 runs):
| X₁ | X₂ | X₃ |
|---|---|---|
| -1 | -1 | 0 |
| +1 | -1 | 0 |
| -1 | +1 | 0 |
| +1 | +1 | 0 |
| -1 | 0 | -1 |
| +1 | 0 | -1 |
| -1 | 0 | +1 |
| +1 | 0 | +1 |
| 0 | -1 | -1 |
| 0 | +1 | -1 |
| 0 | -1 | +1 |
| 0 | +1 | +1 |
| 0 | 0 | 0 |
Model: Same quadratic as CCD.
Trade-off: Slightly less efficient than CCD for prediction, but avoids extreme points.
3. Taguchi Methods (Robust Parameter Design)
When to use: Product/process must perform well despite uncontrollable variation (noise factors). Goal: Find control factor settings that minimize sensitivity to noise.
Inner-Outer Array Structure
Inner array: Control factors (factors you can set in production) Outer array: Noise factors (environmental conditions, material variation, user variation)
Approach: Cross inner array with outer array. Each inner array run is repeated at all outer array conditions.
Example: L₈ inner × L₄ outer (8 control combinations × 4 noise conditions = 32 runs):
Inner array (control factors A, B, C):
| Run | A | B | C |
|---|---|---|---|
| 1 | -1 | -1 | -1 |
| 2 | -1 | -1 | +1 |
| ... | ... | ... | ... |
| 8 | +1 | +1 | +1 |
Outer array (noise factors N₁, N₂):
| Noise | N₁ | N₂ |
|---|---|---|
| 1 | -1 | -1 |
| 2 | -1 | +1 |
| 3 | +1 | -1 |
| 4 | +1 | +1 |
Data collection: For each inner run, measure response Y at all 4 noise conditions. Calculate mean (Ȳ) and variance (s²) or signal-to-noise ratio (SNR).
Signal-to-Noise Ratios:
- Larger-is-better: SNR = -10 log₁₀(Σ(1/Y²)/n)
- Smaller-is-better: SNR = -10 log₁₀(ΣY²/n)
- Target-is-best: SNR = 10 log₁₀(Ȳ²/s²)
Analysis: Choose control factor settings that maximize SNR (robust to noise) while achieving target mean.
Two-step optimization:
- Maximize SNR to reduce variability
- Adjust mean to target using control factors that don't affect SNR
4. Optimal Designs
When to use: Irregular factor space (constraints, categorical factors, unequal ranges), custom run budget, or standard designs don't fit.
D-Optimal Designs
Criterion: Minimize determinant of (X'X)⁻¹ (maximize information, minimize variance of coefficient estimates).
Algorithm: Computer-generated. Start with candidate set of all feasible runs, select subset that maximizes |X'X|.
Use cases:
- Mixture experiments with additional process variables
- Constrained factor spaces (e.g., temperature + pressure can't both be high)
- Irregular grids (e.g., existing data points + new runs)
- Unequal factor ranges
Software: Use R (AlgDesign package), JMP, Design-Expert, or Python (pyDOE).
A-Optimal and G-Optimal
A-optimal: Minimize average variance of predictions (trace of (X'X)⁻¹). G-optimal: Minimize maximum variance across design space (minimax criterion).
Choice: D-optimal for parameter estimation, G-optimal for prediction across entire space, A-optimal for average prediction quality.
5. Mixture Experiments
When to use: Factors are proportions that must sum to 100% (e.g., chemical formulations, blend compositions, budget allocations).
Simplex-Lattice Designs
Constraints: x₁ + x₂ + ... + xₖ = 1, all xᵢ ≥ 0
{q,k} designs: q = number of levels for each component, k = number of components
Example: {2,3} simplex-lattice (3 components at 0%, 50%, 100%):
| Run | x₁ | x₂ | x₃ |
|---|---|---|---|
| 1 | 1.0 | 0.0 | 0.0 |
| 2 | 0.0 | 1.0 | 0.0 |
| 3 | 0.0 | 0.0 | 1.0 |
| 4 | 0.5 | 0.5 | 0.0 |
| 5 | 0.5 | 0.0 | 0.5 |
| 6 | 0.0 | 0.5 | 0.5 |
Model: Scheffé canonical polynomials. Linear: Y = β₁x₁ + β₂x₂ + β₃x₃. Quadratic: Y = Σβᵢxᵢ + Σβᵢⱼxᵢxⱼ.
Simplex-Centroid Designs
Structure: Pure components + binary blends + ternary blends + overall centroid.
Example: 3-component simplex-centroid (7 runs):
| Run | x₁ | x₂ | x₃ |
|---|---|---|---|
| 1 | 1.0 | 0 | 0 |
| 2 | 0 | 1.0 | 0 |
| 3 | 0 | 0 | 1.0 |
| 4 | 0.5 | 0.5 | 0 |
| 5 | 0.5 | 0 | 0.5 |
| 6 | 0 | 0.5 | 0.5 |
| 7 | 0.33 | 0.33 | 0.33 |
Constraints: Add lower/upper bounds if components have minimum/maximum limits. Use D-optimal design for constrained mixture space.
6. Split-Plot Designs
When to use: Some factors are hard to change (e.g., temperature requires hours to stabilize), others are easy to change. Randomizing all factors fully is impractical or expensive.
Structure
Whole-plot factors: Hard to change (temperature, batch, supplier) Subplot factors: Easy to change (concentration, time, operator)
Design: Randomize whole-plot factors at top level, randomize subplot factors within each whole-plot level.
Example: 2² split-plot (Temperature = whole-plot, Time = subplot, 2 replicates):
| Whole-plot | Temp | Subplot | Time | Run order |
|---|---|---|---|---|
| 1 | Low | 1 | Short | 1 |
| 1 | Low | 2 | Long | 2 |
| 2 | High | 3 | Short | 4 |
| 2 | High | 4 | Long | 3 |
| (Replicate block 2) |
Analysis: Mixed model with whole-plot error and subplot error terms. Whole-plot factors tested with lower precision (fewer degrees of freedom).
Trade-off: Allows practical execution when full randomization impossible, but reduces statistical power for hard-to-change factors.
7. Sequential Experimentation
Philosophy: Learn iteratively, adapt design based on results. Minimize total runs while maximizing information.
Stage 1: Screening
Objective: Reduce 10-20 candidates to 3-5 critical factors. Design: Plackett-Burman or 2^(k-p) fractional factorial (Resolution III-IV). Runs: 12-20. Output: Ranked factor list, effect sizes with uncertainty.
Stage 2: Steepest Ascent/Descent
Objective: Move quickly toward optimal region using main effects from screening. Method: Calculate path of steepest ascent (gradient = effect estimates). Run experiments along this path until response stops improving. Example: If screening finds temp effect = +10, pressure effect = +5, move in direction (temp: +2, pressure: +1).
Stage 3: Factorial Optimization
Objective: Explore region around best settings from steepest ascent, estimate interactions. Design: 2^k full factorial or Resolution V fractional factorial with 3-5 factors. Runs: 16-32. Output: Optimal settings, interaction effects, linear model.
Stage 4: Response Surface Refinement
Objective: Fit curvature, find true optimum. Design: CCD or Box-Behnken centered at best settings from Stage 3. Runs: 15-20. Output: Quadratic model, stationary point (optimum), contour plots.
Stage 5: Confirmation
Objective: Validate predicted optimum. Design: 3-5 replication runs at predicted optimal settings. Output: Confidence interval for response at optimum. If prediction interval contains observed mean, model validated.
Total runs example: Screening (16) + Steepest ascent (4) + Factorial (16) + RSM (15) + Confirmation (3) = 54 runs. Compare to one-shot full factorial for 10 factors = 1024 runs.
8. Analysis Techniques
Effect Estimation
Factorial designs: Estimate main effect of factor A as: Effect(A) = (Ȳ₊ - Ȳ₋) where Ȳ₊ = mean response when A is high, Ȳ₋ = mean when A is low.
Interaction effect: Effect(AB) = [(Ȳ₊₊ + Ȳ₋₋) - (Ȳ₊₋ + Ȳ₋₊)] / 2
Standard error: SE(effect) = 2σ/√n, where σ estimated from replicates or center points.
ANOVA
Purpose: Test statistical significance of effects. Null hypothesis: Effect = 0. Test statistic: F = MS(effect) / MS(error), compare to F-distribution. Significance: p < 0.05 (or chosen α level) → reject H₀, effect is significant.
Regression Modeling
Linear model: Y = β₀ + β₁x₁ + β₂x₂ + β₁₂x₁x₂ + ε Quadratic model (RSM): Y = β₀ + Σβᵢxᵢ + Σβᵢᵢxᵢ² + Σβᵢⱼxᵢxⱼ + ε
Fit: Least squares (minimize Σ(Yᵢ - Ŷᵢ)²). Assessment: R², adjusted R², RMSE, residual plots.
Residual Diagnostics
Check assumptions:
- Normal probability plot: Residuals should fall on straight line. Non-normality indicates transformation needed.
- Residuals vs fitted: Random scatter around zero. Funnel shape indicates non-constant variance (transform Y).
- Residuals vs run order: Random. Trend indicates time drift, lack of randomization.
- Residuals vs factors: Random. Pattern indicates missing interaction or curvature.
Transformations: Log(Y) for multiplicative effects, √Y for count data, 1/Y for rate data, Box-Cox for data-driven choice.
Optimization
Contour plots: Visualize response surface, identify optimal region, assess tradeoffs. Desirability functions: Multi-response optimization. Convert each response to 0-1 scale (0 = unacceptable, 1 = ideal). Maximize geometric mean of desirabilities. Canonical analysis: Find stationary point (∂Y/∂xᵢ = 0), classify as maximum, minimum, or saddle point based on eigenvalues of Hessian matrix.
9. Sample Size and Power Analysis
Before designing experiment, determine required runs:
Power: Probability of detecting true effect if it exists (1 - β). Standard: power ≥ 0.80.
Effect size (δ): Minimum meaningful difference. Example: "Must detect 10% yield improvement."
Noise (σ): Process variability. Estimate from historical data, pilot runs, or engineering judgment.
Formula for factorial designs: n ≥ 2(Zα/2 + Zβ)²σ² / δ² per cell.
Example: Detect δ = 5 units, σ = 3 units, α = 0.05, power = 0.80.
- n ≥ 2(1.96 + 0.84)²(3²) / 5² = 2(7.84)(9) / 25 ≈ 6 replicates per factor level combination.
For screening: Use effect sparsity assumption. If testing 10 factors, expect 2-3 active. Size design to detect large effects (1-2σ).
Software: Use G*Power, R (pwr package), JMP, or online calculators.
10. Common Pitfalls and Solutions
Pitfall 1: Ignoring confounding in screening designs
- Problem: Plackett-Burman confounds main effects with 2-way interactions. If interactions exist, main effect estimates are biased.
- Solution: Use only when sparsity-of-effects applies (most interactions negligible). Follow up ambiguous results with Resolution IV/V design or fold-over.
Pitfall 2: Extrapolating beyond design region
- Problem: Response surface models are local approximations. Predicting outside tested factor ranges is unreliable.
- Solution: Expand design if optimum appears outside current region. Run steepest ascent, then new RSM centered on improved region.
Pitfall 3: Inadequate replication
- Problem: Without replicates, cannot estimate pure error or test lack-of-fit.
- Solution: Always include 3-5 center point replicates. For critical experiments, replicate entire design (2-3 times).
Pitfall 4: Changing protocols mid-experiment
- Problem: Breaks orthogonality, confounds design structure with time.
- Solution: Complete design as planned. If protocol change necessary, analyze before/after separately or treat as blocking factor.
Pitfall 5: Treating categorical factors as continuous
- Problem: Assigning arbitrary numeric codes (-1, 0, +1) to unordered categories (e.g., Supplier A/B/C) implies ordering that doesn't exist.
- Solution: Use indicator variables (dummy coding) or separate experiments for each category level.