How to Design a Media Optimization Study Using DOE

May 2026 16 min read Bioprocess Engineering

Key Takeaways

Contents

  1. Why DOE for Media Optimization?
  2. The Sequential DOE Workflow
  3. Step 1: Screening — Identifying Critical Components
  4. Step 2: Optimization — Mapping the Response Surface
  5. Definitive Screening Designs: Screen and Optimize in One Step
  6. Worked Example: CHO mAb Media Optimization
  7. Common Pitfalls and How to Avoid Them
  8. Frequently Asked Questions

Why DOE for Media Optimization?

Design of experiments (DOE) is the most efficient statistical method for optimizing cell culture media because it reveals both individual component effects and interactions between components in far fewer runs than one-factor-at-a-time (OFAT) approaches. A chemically defined CHO medium typically contains 50–80 components, and OFAT testing even 10 of these at 3 levels would require 30 independent experiments that miss every interaction effect. DOE detects interactions by design.

Media composition directly controls cell growth, product titer, and critical quality attributes like glycosylation. Amino acids alone can account for 25–40% of total media cost in large-scale mammalian cell culture, making optimization both a productivity and an economics problem. DOE-guided media optimization routinely improves titers by 30–70%, with published studies reporting mAb titers increasing from 3–4 g/L to 7–10 g/L after systematic amino acid rebalancing.

The core advantage of DOE over OFAT is the ability to detect interaction effects. If glutamine and asparagine both individually increase growth by 10%, their combination might yield a 30% increase (synergy) or only 5% (antagonism). OFAT experiments cannot detect this—only DOE with appropriately designed factor combinations can.

OFAT vs DOE: Design Space Coverage Left panel shows OFAT exploring one factor at a time along axes only. Right panel shows DOE exploring combinations across the entire design space, enabling detection of interactions and optimal regions missed by OFAT. OFAT Approach Glutamine (mM) Asparagine (mM) Interactions missed DOE Approach Glutamine (mM) Asparagine (mM) Optimum Factorial Center Axial
Figure 1 — OFAT explores one variable at a time along the axes, missing the interior of the design space where interactions and the true optimum reside. DOE places runs at strategic combinations (factorial corners, center points, axial points) to map the full response surface.
Diagram comparing OFAT and DOE approaches. OFAT tests points along axes only, while DOE distributes points across the entire design space using factorial, center, and axial points to detect interactions.

The Sequential DOE Workflow

Media optimization follows a staged approach that progressively narrows the design space from dozens of candidate components down to the optimal concentration of 3–5 critical factors. Each stage uses a different DOE design type matched to its objective.

Media Optimization DOE Workflow Four-stage workflow: Component Screening (Plackett-Burman, 12-20 runs) narrows to Augmentation/Path of Steepest Ascent (6-10 runs), then Response Surface Optimization (CCD/BBD, 15-30 runs), and finally Verification (3-5 runs at predicted optimum). 1. Screening Plackett–Burman or Frac. Factorial 12–20 runs 2. Augmentation Path of Steepest Ascent 6–10 runs 3. Optimization CCD or Box– Behnken 15–30 runs 4. Verification Confirm at Predicted Optimum 3–5 runs Weeks 1–3 Weeks 3–5 Weeks 5–9 Weeks 9–11
Figure 2 — The sequential media optimization DOE workflow. Each stage narrows the design space and changes the design type to match the objective. Total timeline: 8–12 weeks, 40–60 experiments.
Four-stage workflow diagram showing screening (12-20 runs), augmentation (6-10 runs), optimization (15-30 runs), and verification (3-5 runs) with approximate timelines.
Table 1 — DOE Design Types for Each Stage of Media Optimization
Stage Design Type Factors Levels Runs (typical) What It Reveals
Screening Plackett–Burman 7–15 2 12–20 Main effects only
Screening Fractional factorial (Res III–IV) 5–11 2 8–32 Main effects + some 2FI
Screening + curvature Definitive screening (DSD) 5–16 3 13–33 Main effects + quadratics + some 2FI
Optimization Central composite (CCD) 2–5 5 9–32 Full quadratic model
Optimization Box–Behnken (BBD) 3–5 3 13–46 Full quadratic (avoids extremes)
Runs include center-point replicates. 2FI = two-factor interaction. Resolution III designs alias main effects with 2FI; Resolution IV designs alias 2FI with each other but not with main effects.

Step 1: Screening — Identifying Critical Components

The screening stage identifies which media components significantly affect your responses from a large initial list. For a CHO fed-batch process, you might start with 10–15 candidate components drawn from spent media analysis, literature, and supplier recommendations. The goal is to reduce this to 3–5 critical factors for detailed optimization.

Choosing Candidate Components

Start by analyzing spent media from your baseline process. Components that are depleted below 20% of their initial concentration by harvest are candidates for supplementation. Amino acids are almost always the highest-impact group: glutamine, asparagine, serine, and cysteine are commonly limiting in CHO cultures. Trace metals (iron, zinc, copper, manganese) and vitamins (B12, biotin, folic acid) are also frequent screening candidates.

Group related components when the total factor count exceeds 15. For example, test "branched-chain amino acids" (leucine, isoleucine, valine) as a single group factor at low and high levels rather than as three separate factors. This reduces the design size while preserving the ability to detect group-level effects.

Setting Factor Levels

For screening, use two levels per factor: a "low" level (typically 0.5–0.75× the current concentration) and a "high" level (1.5–2.0× the current concentration). The range must be wide enough to produce a measurable response difference but not so extreme that it causes toxicity or precipitation.

The Plackett–Burman Design

A Plackett–Burman (PB) design is the most run-efficient screening design, requiring only N+1 experiments (rounded to the nearest multiple of 4) for N factors. It estimates main effects independently but aliases all two-factor interactions with main effects, so any significant "main effect" might actually be a disguised interaction. This is acceptable for screening because the purpose is to identify the important factors, not to build a precise model.

Worked Example: PB Design Size Calculation

You want to screen 11 media components (8 amino acids, 2 trace metals, 1 vitamin) for their effect on mAb titer.

Compare: Full factorial at 2 levels = 211 = 2,048 runs
Efficiency gain: 2,048 / 15 = 137× fewer experiments

Always include 3–5 center-point replicates (all factors at their midpoint). Center points serve three purposes: (1) they estimate pure error for the F-test, (2) they detect curvature (if the center-point mean differs significantly from the factorial mean, at least one factor has a quadratic effect), and (3) they verify that the baseline condition is reproducible.

Media Cost Estimator

Estimate the cost impact of media component changes across scales from shake flask to production bioreactor.

Open Calculator

Step 2: Optimization — Mapping the Response Surface

Once screening identifies 3–5 critical components, optimization DOE maps the curved response surface to find the true optimum concentration for each. The two most common designs are the central composite design (CCD) and the Box–Behnken design (BBD).

Central Composite Design (CCD)

A CCD consists of three parts: the 2k factorial points, 2k axial (star) points at a distance α from the center, and center-point replicates. For k = 3 factors, this means 8 factorial + 6 axial + 5–6 center = 19–20 total runs. The CCD fits a full second-order polynomial model:

Y = β0 + Σ βiXi + Σ βiiXi2 + Σ βijXiXj + ε

Set α = (2k)1/4 for rotatability (equal prediction variance at equal distances from the center in all directions). For 3 factors, α = 1.682. Use face-centered CCD (α = 1) if you cannot run factor levels beyond the original screening range.

Box–Behnken Design (BBD)

A BBD avoids extreme corners of the design space by only testing factor combinations at the midpoints of edges and the center. For 3 factors, it requires only 13 runs (vs. 20 for a CCD). This is advantageous when corner combinations might cause cell death (e.g., simultaneously high copper and low cysteine). The trade-off is that BBD cannot detect effects at the extremes of the design space.

Choosing Between CCD and BBD

Figure 3 — Pareto chart from a PB screening design of 11 CHO media components. Components above the Bonferroni-corrected significance threshold (dashed line) are carried forward to optimization. Glutamine, asparagine, and iron are the top 3 drivers of mAb titer in this example.

Definitive Screening Designs: Screen and Optimize in One Step

A definitive screening design (DSD) combines the efficiency of screening with the ability to detect curvature and some two-factor interactions in a single experimental campaign. Introduced by Jones and Nachtsheim in 2011, DSDs require only 2k+1 runs for k factors (e.g., 17 runs for 8 factors) while providing three levels per factor.

DSDs achieve this efficiency through a conference-matrix construction that ensures all main effects are orthogonal to each other, orthogonal to all two-factor interactions, and orthogonal to all quadratic effects. In practice, this means you can fit main effects + any significant quadratic or interaction terms without confounding, provided the number of active effects is fewer than about half the number of runs.

When to Use DSD vs. Sequential PB + RSM

A 2024 study by Bai et al. compared fractional factorial designs and DSD for CHO cell culture process development in Ambr 15 miniature bioreactors. The DSD was more effective at detecting both main and quadratic effects while using fewer runs, with models that transferred well from the 15 mL to 3 L bioreactor scale.

Table 2 — Design Comparison: Experiments Required for Media Optimization
Approach 8 Factors 11 Factors Detects Curvature? Detects 2FI? Campaigns
Full factorial 256 2,048 Yes Yes (all) 1
PB screen + CCD (3 factors) 12 + 20 = 32 12 + 20 = 32 Yes (stage 2) Yes (stage 2) 2
DSD alone 17 23 Yes Partial 1
DSD + CCD confirmation 17 + 9 = 26 23 + 9 = 32 Yes Yes 2
OFAT (3 levels) 24 33 No No 8–11
Runs include center-point replicates. The DSD approach achieves the best efficiency-to-information ratio for 5–16 factors. OFAT campaigns listed assume sequential, non-parallel single-factor experiments.

Worked Example: CHO mAb Media Optimization

This worked example walks through a realistic CHO mAb media optimization study from screening through verification, showing how to size each stage and interpret the results.

Scenario

A CHO DG44 cell line producing a biosimilar mAb currently achieves 3.2 g/L in a 14-day fed-batch process with a commercial chemically defined medium. Spent media analysis shows depletion of glutamine, asparagine, serine, cysteine, and isoleucine by day 8. Trace metal analysis shows iron dropping to <10% of initial. The goal is to improve titer to ≥5 g/L without changing the basal medium vendor.

Stage 1: PB Screening (12 runs + 3 center points)

Screening Setup

Factors (11 total):

Responses measured: peak VCD (106 cells/mL), day-14 titer (g/L), day-14 viability (%), galactosylation (% G0F)

Design: 12-run PB + 3 center points = 15 experiments in shake flasks

Results (Pareto analysis at p < 0.05):

Stage 2: Path of Steepest Ascent (6 runs)

The steepest ascent method moves along the gradient defined by the screening model coefficients. Starting from the center point, each step increases glutamine, asparagine, and iron in proportion to their estimated effects. After 6 steps, titer peaked at step 4 (4.6 g/L) then declined at step 5, indicating the region near the optimum.

Stage 3: CCD Optimization (20 runs)

CCD Setup and Results

Factors (3): glutamine (2–8 mM), asparagine (1–5 mM), iron (5–25 µM)

Design: CCD with α = 1.682, 8 factorial + 6 axial + 6 center = 20 runs in 250 mL shake flasks

Response surface model (titer):

Titer = 5.24 + 0.41·Gln + 0.28·Asn + 0.18·Fe − 0.35·Gln2 − 0.22·Asn2 + 0.15·Gln·Asn

Model quality: R2 = 0.94, R2adj = 0.91, R2pred = 0.85, adequate precision = 14.2

Predicted optimum: glutamine 5.8 mM, asparagine 3.4 mM, iron 18 µM → predicted titer 5.6 g/L

Stage 4: Verification (5 runs)

Five replicate shake flasks at the predicted optimum yielded titers of 5.3, 5.5, 5.7, 5.4, and 5.6 g/L (mean = 5.5 ± 0.16 g/L, CV = 2.9%). The predicted 5.6 g/L fell within the 95% confidence interval of the observed mean. The optimized medium was then scaled to 3 L bioreactors, achieving 5.8 g/L with improved mixing and DO control.

Titer improvement: 3.2 → 5.5 g/L = 72% increase
Total experiments: 15 + 6 + 20 + 5 = 46 runs over 10 weeks

Figure 4 — mAb titer improvement across the four DOE stages. Each bar represents the best-performing condition at that stage. Error bars show the range of replicate measurements. The baseline 3.2 g/L improved to 5.5 g/L after CCD optimization and verification.

Fed-Batch Calculator

Model fed-batch feeding profiles and predict glucose, lactate, and cell density trajectories for optimized media conditions.

Open Calculator

Common Pitfalls and How to Avoid Them

Media optimization DOE studies fail most often due to avoidable experimental design and execution errors, not statistical complexity. Here are the seven most common pitfalls and their remedies.

1. Factor Ranges Too Narrow

If the low and high levels are too close, the signal is buried in noise. A rule of thumb: the expected response difference between low and high levels should be at least 2–3× the standard deviation of your measurement assay. For mAb titer measured by Protein A HPLC (typical CV = 3–5%), factor levels should produce at least a 10–15% titer difference.

2. Ignoring Center Points

Skipping center-point replicates removes your ability to detect curvature and estimate pure error. Without pure error, you cannot compute a lack-of-fit test. Always run at least 3 center points in any screening or optimization design.

3. Confounding Run Order with Factor Effects

If you run all "high glutamine" conditions on Monday and all "low glutamine" on Friday, passage number, media age, or operator fatigue become confounded with the glutamine effect. Always randomize run order. If complete randomization is impractical (e.g., hard-to-change factors like temperature), use a split-plot design.

4. Measuring Too Few Responses

Optimizing titer alone can inadvertently degrade glycosylation, charge variants, or aggregation. Always measure at least 3 responses: growth (VCD or IVC), productivity (titer), and quality (glycan profile, SEC monomer %, or charge heterogeneity by icIEF).

5. Oversized Optimization Designs

Running a CCD with 5 factors when only 3 were significant in screening wastes resources. Carry forward only factors with p < 0.05 (or p < 0.10 if borderline and biologically plausible). Use the screening model's lack-of-fit and curvature tests to decide whether augmentation is needed before jumping to RSM.

6. Skipping Verification

The predicted optimum from the response surface model is a statistical estimate with confidence intervals. Running 3–5 verification replicates at the predicted optimum is essential to confirm reproducibility and check that the model's prediction error is acceptable (<10% relative error for titer).

7. Not Accounting for Scale Effects

Media optimization in 96-well plates or shake flasks may not transfer perfectly to stirred-tank bioreactors. Oxygen transfer, mixing time, and pH control differ across scales. Run the verification stage at the target scale (or at least at miniature bioreactor scale with controlled DO and pH) to confirm robustness.

Table 3 — Screening Design Selection Guide
Factors Recommended Design Runs (+ center pts) Resolution Key Trade-off
4–5 Half-fraction factorial 8–16 + 3 IV–V High resolution but limited to few factors
6–8 DSD 13–17 + 0 n/a Best for simultaneous screen + curvature
7–11 Plackett–Burman 12 + 3 III Most run-efficient; aliases 2FI with main
12–19 PB (20-run) 20 + 3 III Handles many factors; same aliasing risk
20+ Group screening + staged PB 2 × 12 + 3 III Split factors into logical groups
DSD center points are built into the design matrix. For PB and fractional factorial, add center points explicitly.

Frequently Asked Questions

How many experiments does a Plackett–Burman design require for media screening?

A Plackett–Burman design requires N+1 runs (rounded up to the nearest multiple of 4), where N is the number of factors. For example, screening 11 media components requires only 12 experimental runs, compared to 2,048 runs for a full factorial design at two levels. This makes PB designs extremely efficient for initial media component screening.

What is the difference between screening and optimization DOE for media development?

Screening DOE (e.g., Plackett–Burman, fractional factorial) identifies which of many media components significantly affect your response, using 2 levels per factor and few runs. Optimization DOE (e.g., central composite, Box–Behnken) maps the response surface for the 3–5 critical components identified during screening, using 3–5 levels per factor to find the optimal concentration of each.

Can DOE handle more than 10 media components simultaneously?

Traditional DOE works well for up to 10–15 components. For larger design spaces (20–60 components in chemically defined media), a staged approach is necessary: group related components (amino acids, trace metals, vitamins), screen each group separately, then combine the significant factors into a joint optimization study. Alternatively, definitive screening designs or machine learning-augmented DOE can handle higher-dimensional spaces more efficiently.

What responses should I measure in a media optimization DOE?

Measure multiple responses simultaneously: viable cell density (VCD), viability, product titer, and at least one product quality attribute (e.g., glycosylation profile for mAbs). Include a growth-rate metric (integral of viable cells, IVC) and a productivity metric (specific productivity qP in pg/cell/day). Multiple responses help avoid optimizing titer at the expense of product quality or cell health.

How do I set factor ranges for media optimization DOE?

Set factor ranges based on literature, spent media analysis, and preliminary experiments. For screening, use a wide range (typically 0.5–2× of the current concentration) to ensure you detect real effects. For optimization, narrow the range to ±20–30% around the best levels identified in screening. Always include center points (3–5 replicates) to estimate pure error and detect curvature.

CHO Troubleshooter

Diagnose and fix common CHO cell culture problems including low titer, poor viability, lactate accumulation, and abnormal glycosylation.

Open Tool

Related Tools

References

  1. Torkashvand F. et al. "Designed Amino Acid Feed in Improvement of Production and Quality Targets of a Therapeutic Monoclonal Antibody." PLOS ONE, 10(10): e0140597, 2015. doi: 10.1371/journal.pone.0140597
  2. González-Leal I.J. et al. "Use of a Plackett–Burman statistical design to determine the effect of selected amino acids on monoclonal antibody production in CHO cells." Biotechnology Progress, 27(6): 1709–1717, 2011. doi: 10.1002/btpr.674
  3. Zhou T. et al. "A review of algorithmic approaches for cell culture media optimization." Frontiers in Bioengineering and Biotechnology, 11: 1195294, 2023. doi: 10.3389/fbioe.2023.1195294
  4. Bai Y. et al. "Enhancing early-stage cell culture process development efficiency using an integrated approach of high-throughput miniaturized bioreactors and definitive screening design." Biochemical Engineering Journal, 203: 109217, 2024. doi: 10.1016/j.bej.2024.109217
  5. Narayanan H. et al. "Accelerating cell culture media development using Bayesian optimization-based iterative experimental design." Nature Communications, 16: 4539, 2025. doi: 10.1038/s41467-025-61113-5
Share

Resources & Further Reading