A Deep Learning Approach For Optimizing Monoclonal Antibody Production Process Parameters
A Deep Learning Approach For Optimizing Monoclonal Antibody Production Process Parameters
www.scipublication.com
Keywords Abstract
Monoclonal antibody This study presents a new deep learning method for optimizing monoclonal
production, Deep antibody (mAb) production processes using a hybrid Convolutional Neural
learning, Process Network-Long Short-Term Memory (CNN-LSTM) architecture. The model
optimization, CNN- was developed and validated using industry data from 50 products over 18
LSTM months. The proposed design outperforms statistical models, machine
learning algorithms, and other deep learning models, achieving a root mean
squared error of 0.412 g/L and R^ 2 value of 0.947 for mAb titer prediction.
Feature importance analysis identified temperature, dissolved oxygen, and pH
as the most critical parameters affecting mAb production. In silico
optimization, experiments demonstrated a 28.1% increase in mAb titer and a
27.9% improvement in volumetric productivity. The model's robustness and
generalizability were validated across cell lines and bioreactor scales (50L to
2000L). A novel Dynamic Trajectory Similarity (DTS) score was introduced
to quantify the model's ability to capture process dynamics, yielding a score
of 0.923. This approach offers significant potential for enhancing process
understanding, optimizing production efficiency, and facilitating scale-up in
industrial mAb manufacturing. The study also discusses limitations, including
interpretability challenges and the need for uncertainty quantification in
future work.
Recent studies have explored the use of deep learning Quality by Design (QbD) models are also important in
for various aspects of mAb production, including mAb development. This approach involves identifying
process monitoring, defect detection, and predictive a design environment in which changes in the process
product modeling. The ability of deep learning models do not affect product quality. The implementation of
to handle high-dimensional data and learn different QbD has led to a better understanding of the process
and a more robust production process. Process
Data preprocessing involves several steps to ensure samples, each containing 96-time steps (24 hours * 4
data quality and consistency. Missing values were measurements per hour) for online variables and 1-time
imputed using a combination of linear interpolation for step for daily measurements.
short gaps and a k-nearest neighbors algorithm for
longer gaps. Outliers were detected and removed using 3.2. Deep learning model architecture design
the Interquartile Range (IQR) method. To facilitate
model training, all variables were standardized to zero The proposed deep learning architecture combines
mean and unit variance. Convolutional Neural Networks (CNNs) and Long-
Time-series data was restructured into sliding windows Short-Term Memory (LSTM) networks to capture
of 24 hours, with a stride of 6 hours, to capture spatial and temporal patterns in bioprocess data. Figure
temporal dependencies. This resulted in a total of 7,200 1 illustrates the overall structure of the model.
3.3. Model training and hyperparameter epochs to prevent overfitting and monitor the
optimization validation loss.
Hyperparameter optimization was performed using
The model was trained using the Adam optimizer with Bayesian optimization with a Gaussian process prior.
an initial learning rate of 0.001 and a batch size of 64. The hyperparameters tuned included the number of
The loss function was a mean squared error (MSE) for convolutional filters, LSTM units, dropout rates, and
both mAb titer and quality attribute predictions. Early learning rates. Table 3 shows the hyperparameter
stopping was implemented with a patience of 20 search space and the optimal values found.
Table 3: Hyperparameter optimization results
Hyperparameter Search Range Optimal Value
Conv1D Filters [32, 64, 128] 64
LSTM Units (Layer 1) [64, 128, 256] 128
LSTM Units (Layer 2) [32, 64, 128] 64
Dropout Rate [0.1, 0.3, 0.5] 0.3
Learning Rate [1e-4, 1e-3, 1e-2] 1e-3
Figure 2 visualizes the hyperparameter optimization convergence towards the optimal hyperparameter
process, showing the optimization algorithm's configuration.
Figure 2: Hyperparameter optimization convergence
Figure 3 illustrates the distribution of prediction errors across different process phases using violin plots
.
Figure 3: Distribution of prediction errors across process phases
The proposed CNN-LSTM model outperformed all 94.7% of the variance in mAb titer, demonstrating its
benchmark models across all evaluation metrics. The strong predictive capability. The Dynamic Trajectory
model achieved a root mean squared error (RMSE) of Similarity (DTS) score of 0.923 highlights the model's
0.412 g/L and a mean absolute error (MAE) of 0.318 ability to accurately capture the temporal dynamics of
g/L for mAb titer prediction, representing the bioprocess[23] . Figure 4 illustrates the prediction
improvements of 13.4% and 12.9%, respectively, accuracy of the proposed model compared to actual
compared to the next best performing model (CNN). mAb titer values over the course of a production batch.
The R² value of 0.947 indicates that the model explains
Figure 4: Comparison of predicted and actual mAb titer trajectories
The analysis reveals that temperature, dissolved of precise control of these parameters throughout the
oxygen, and pH are the most critical parameters production process. Figure 5 visualizes the temporal
affecting mAb production, accounting for 48.5% of the importance of key process parameters throughout the
total feature importance. This aligns with existing production batch.
knowledge in the field and underscores the importance
Figure 5: Temporal importance of key process parameters
The optimized process parameters led to a 28.1% adjustments to key process parameters, demonstrating
increase in mAb titer on day 14, from 3.2 g/L to 4.1 the power of data-driven optimization in fine-tuning
g/L. Volumetric productivity improved by 27.9%, complex bioprocesses. Figure 6 illustrates the
while specific productivity increased by 19.8%. These optimization landscape for mAb titer as a function of
improvements were achieved through subtle temperature and dissolved oxygen.
Figure 6: Optimization landscape for mAb titer
The model demonstrated robust performance across In terms of scalability, the model's performance
different cell lines, with only minor degradation in remained strong when applied to data from 50L and
predictive accuracy. The RMSE increased by 6.1%, 2000L bioreactors. The RMSE increased by 4.1% for
9.5%, and 13.6% for cell lines B, C, and D, the 50L scale and 8.0% for the 2000L scale, compared
respectively, compared to the original cell line A. This to the original 200L scale. This indicates that the
suggests that the model can capture generalizable model can effectively capture scale-dependent effects
features of mAb production processes across different and maintain its predictive power across different
cell lines. production scales. Figure 7 visualizes the model's
performance consistency across different cell lines and
scales.
Figure 7: Model performance consistency across cell lines and scales