Next Article in Journal
Vertical Distributions and Bioavailabilities of Heavy Metals in Soil in An-Tea Plantations in Qimen County, China
Previous Article in Journal
Coprocessing Corn Germ Meal for Oil Recovery and Ethanol Production: A Process Model for Lipid-Producing Energy Crops
Previous Article in Special Issue
About Model Validation in Bioprocessing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Holistic Process Models: A Bayesian Predictive Ensemble Method for Single and Coupled Unit Operation Models

1
Digitalization Development Biologicals CMC, Boehringer Ingelheim Pharma GmbH & Co. KG, D-88397 Biberach, Germany
2
Bio Quality Mammalian, Boehringer Ingelheim Pharma GmbH & Co. KG, D-88397 Biberach, Germany
3
Global Computational Biology and Digital Sciences, Boehringer Ingelheim Pharma GmbH & Co. KG, D-88397 Biberach, Germany
4
IT RDM Services, Boehringer Ingelheim GmbH & Co. KG, D-88397 Biberach, Germany
5
IT RDM Services, Boehringer Ingelheim RCV GmbH & Co. KG, A-1121 Vienna, Austria
6
Analytical Development Biologicals, Boehringer Ingelheim Pharma GmbH & Co. KG, D-88397 Biberach, Germany
7
Biberach University of Applied Sciences, D-88400 Biberach, Germany
8
Institute for Computational Physics, University of Stuttgart, D-70569 Stuttgart, Germany
*
Author to whom correspondence should be addressed.
Processes 2022, 10(4), 662; https://doi.org/10.3390/pr10040662
Submission received: 1 March 2022 / Revised: 18 March 2022 / Accepted: 23 March 2022 / Published: 29 March 2022
(This article belongs to the Special Issue Model Validation Procedures)

Abstract

:
The coupling of individual models in terms of end-to-end calculations for unit operations in manufacturing processes is a challenging task. We present a probability distribution-based approach for the combined outcomes of parametric and non-parametric models. With this so-called Bayesian predictive ensemble, the statistical moments such as mean value and standard deviation can be accurately computed without any further approximation. It is shown that the ensemble of different model predictions leads to an uninformed prior distribution, which can be transformed into a predictive posterior distribution using Bayesian inference and numerical Markov Chain Monte Carlo calculations. We demonstrate the advantages of our method using several numerical examples. Our approach is not restricted to certain unit operations, and can also be used for the more robust interpretation and assessment of model predictions in general.

1. Introduction

Bioprocess development and manufacturing for therapeutic drugs is often a challenging task. The parameter settings of individual unit operations at upstream, downstream and formulation stages need to be adjusted so that the final drug product is produced at high concentration with a robust and satisfying quality that meets regulatory and patients requirements [1,2,3]. Due to the coupling of different process steps including cultivation, purification, polishing, filtering and formulation among others, it is known that the dimension of process parameters which form the design space grows significantly with an increasing number of unit operations [1].
Notably, the identification of global optimal process parameter settings for all combined unit operations is not straightforward and provides some challenges for experimental work. Most often, normal operating ranges (NORs) and proven-acceptable ranges (PARs) are defined for robust and local parameter settings of single unit operations. Hence, bioprocess optimization is mainly performed for all process steps individually while the combined global optimization of parameter settings is often ignored due to the enormous amount of experimental work.
In recent years, several computational modeling approaches were introduced to guide experimental settings and to reduce laboratory work [1,3,4]. Standard methods are mechanistic and hybrid models [5,6,7,8,9,10,11,12,13,14] in combination with statistical, data-driven or empirical approaches for upstream and downstream process steps [15,16,17,18]. Parametric models are usually calibrated in terms of certain experimental conditions and are often used to study various process parameter settings. Noteworthy, most parametric modeling approaches are computationally cheap and the calculations can be performed on short time scales, which allows to accelerate development times significantly. In addition, the last years have also seen a significant increase of non-parametric models in terms of machine learning [1,4]. Despite the underlying parametric or non-parametric characteristics, an important concept for model development and application is validation, which is usually conducted in comparison to experimental outcomes [19,20].
The straightforward application of modeling approaches further emerged recent interest in the combined evaluation of process outcomes (Figure 1).
Hence, over the last years, several combined end-to-end process models have been developed, which are often called integrated process models, end-to-end bioprocess models, holistic process models, bioprocess replicas, flowsheet models or bioprocess digital twins [1,4,21,22,23,24,25,26]. Despite the fact that certain differences in the underlying methodology and concepts are evident, all approaches offer the study of process parameter settings across all unit operations in terms of global optimization. However, the coupling of models between the individual unit operations is still a challenging task (Figure 1). A consistent and combined multidimensional approach to estimate parameter ranges for predictions has not been conclusively proposed. Recent approaches suggested a Monte Carlo sampling scheme [22,25], whereas other publications favored a progression of discrete values for further processing steps also in terms of model-predictive control strategies [27,28,29]. In more detail, it was shown that the detailed evaluation of process parameter ranges crucially affects the outcomes [22], such that the detailed consideration of coupled effects for all models is of utmost importance.
Specifically the determination of confidence intervals is a challenging task for parametric and non-parametric models. Notably, standard techniques like bootstrapping [30,31,32] or methods that consider Bayesian inference calculations [33,34] often rely on a reasonable amount of model parameterization or training data. Recent articles already showed that the transformation of outcome ranges along certain unit operations is not trivial and crucially affects the specifications [22,25]. Further articles also focused on model-free identification strategies for optimal parameter settings [35,36]. In terms of coupled unit operation models, it is clear that the resulting design space is high dimensional and optimal parameter settings can mainly be identified with stochastic approaches. In consequence, it can be assumed that probabilistic and stochastic approaches for the calculation and optimization of process outcomes will become even more important in the near future.
In this article, we present a probabilistic approach to predict process outcomes using a Bayesian predictive ensemble approach. We show that an ensemble of different model predictions leads to an uninformed prior distribution, which can be transformed into a predictive posterior distribution using Bayesian inference in combination with Markov Chain Monte Carlo (MCMC) calculations. Such approaches are particularly applicable for the modeling of coupled process variables within the framework of end-to-end or holistic process models. Our approach is not restricted to bioprocess models, and can be independently used for all machine learning calculations as well as parametric models.
For the first time, we present a rigorous mathematical framework to describe unit operation models in terms of time evolution operators. The corresponding approach allows us to define meaningful and multivariate transformation functions for the connection of unit operations in accordance with holistic process models. The corresponding introduction of predictive distributions facilitates Bayesian inference calculations for a meaningful evaluation of confidence intervals. We show that point-like discrete connections or simple Gaussian distributions lead to a strong increase of uncertainty after an increasing number of connected unit operations. In contrast, our suggested approach of using conditional probabilities in terms of posterior distribution functions results in a controllable increase and further takes the process memory into full consideration. As a new concept, we introduce the idea of using a collection of different model predictions. This so-called Bayesian predictive ensemble allows us to overcome drawbacks in terms of limited experimental data for model parametrization.
The article is organized as follows. In Section 2, we introduce a mathematical description for time-dependent models and discuss their influence on the underlying probability distribution functions. The corresponding framework is then used to outline the main properties of the Bayesian predictive ensemble in combination with transformation functions in terms of posterior and prior distributions. In Section 3, we present a simplified numerical example to highlight the implications for the sake of clarity. We conclude and summarize in Section 4.

2. Theoretical Background

Biotechnological or chemical processes are often dominated by time-dependent behavior. Hence, the corresponding process outcomes show a temporal evolution which hinders the straightforward calculation of statistical moments. In the first subsection, we propose a consistent scheme to compute statistical properties at certain time points for processes with temporal evolution. In more detail, we present a time discretization scheme that can be interpreted as the initial and the end time points of individual process steps. The connections between these time points for the individual probability distribution functions are described in terms of model propagation steps. Explicit evaluation of these distribution functions is performed in terms of Bayesian inference. We show that the outcomes of different models can be used as uninformed prior distributions. The application of numerical Markov Chain Monte Carlo (MCMC) calculations allows us to compute the corresponding posterior distribution for the considered process outcomes. The range of this distribution reflects the impact of model or experimental uncertainty regarding the correct prediction. Moreover, the respective posterior distribution can further be used as initial input for the subsequent model in order to define a new prior distribution for the next unit operation. This iterative scheme based on conditional probabilities represents a new approach for the rational combination of unit operation models.

2.1. Non-Stationary Processes

Any non-stationary process is characterized by its explicit time dependency. Thus, the evaluation of mean values A and higher moments like the variance σ 2 ( A ) = ( A A ) 2 for arbitrarily chosen variables or process outcomes A ( t ) with temporal evolution is challenging. As a further complication, the ergodic hypothesis [37]
lim T 1 T 0 T d t A ( Γ , t ) = Ω d ρ ( Γ ) A ( Γ ) ,
is not applicable for time-dependent processes with temporal evolution. The left hand side of Equation (1) denotes a time average whereas the right hand side highlights measurements over different statistical replicas of the system
A ( Γ ) = Ω d ρ ( Γ ) A ( Γ ) ,
where the corresponding parameters are an arbitrarily chosen observable A ( Γ , t ) = A ( Γ ( t ) ) | Γ ( 0 ) = Γ and the normalized probability density Ω d ρ ( Γ ) = 1 , where T denotes the measured time interval, and Ω the parameter space including all relevant parameters Ω = { Γ } in terms of Γ = ( X ) , where X denotes all variables X = ( X 0 , X 1 , , X P ) T in accordance with vector notation. Notably, the ergodic hypothesis holds for most systems as long as the decorrelation times τ from autocorrelation functions
A ( Γ , t = 0 ) A ( Γ , t ) exp ( t / τ ) β
are short in accordance with the Markovian properties including the stretched exponential factor β { R + , 0 } .
Despite all challenges, it is possible to average over the values at certain fixed time points t M , which results in
0 T d t A ( Γ , t ) δ ( t t M ) = Ω d ρ ( Γ , t M ) A ( Γ , t M )
and hence
A ( Γ , t ) t M = Ω d ρ ( Γ , t M ) A ( Γ , t M )
for time dependent values ρ ( Γ , t ) , A ( Γ , t ) , where δ ( t t M ) denotes the delta function, which is δ ( t t M ) = 1 for t = t M and 0 otherwise. Hence, if we assume that t M is the time point of measurement, one can interpret A ( Γ , t ) t M as the corresponding statistical outcome for process values at identical time points for different replicas of the system. In consequence, the implications from Equation (5) provide the usage of standard statistical approaches for non-stationary processes like cultivation steps [15]. Thus, we introduced a framework to project the outcomes of a set of non-stationary processes on fixed time points in order to define distributions with mean values and standard deviations. With regard to such a concept, we can consider time-dependent and stationary processes as equivalent in terms of identical statistical considerations.

2.2. Time Evolution Propagators

In addition to the quasi-stationary approximation for the calculation of mean values that we introduced in the previous subsection, it is also possible to define a similar interpretation for the probability distribution functions in the presence of explicit time evolution. For a constant and time-independent probability density, it follows
d d t ρ ( Γ , t ) = 0 .
which can be written as a continuity equation in accordance with [38]
ρ ( Γ , t ) t = Γ ρ ( Γ ) Γ ˙
where the dot over the symbol denotes the first derivative in time according to Γ ˙ = ( X ˙ i , X ˙ i + 1 , , X ˙ N ) and Γ Γ ˙ = i N X ˙ i / X i . Notably, Equation (7) can also be written as
i ρ ( Γ , t ) t = L ( Γ ) ρ ( Γ , t )
with the Liouville-like operator [38]
L ( Γ ) = i i N X ˙ i X i
providing the Eigenvalue relation
L ( Γ ) ρ ( Γ , t ) = λ j ρ ( Γ , t )
with the eigenvalues λ j . Here, we do not explicitly intend to introduce Hamiltonian and thus deterministic energy conservation dynamics [38], but mainly rely on the idea of conserved time evolution through operator formalisms. Thus, the Liouville-like operator does not depend on the momenta and spatial derivatives of the microscopic variables for our purposes. The complex conjugate of Equation (9) reads L * = L , such that Equation (10) yields
L * ( Γ ) ρ ( Γ , t ) = λ j ρ ( Γ , t )
which is equivalent to a time inversion t t . The formal solution for Equation (8) reads
ρ ( Γ , t ) = ρ 0 ( Γ ) e i L ( Γ ) t .
such that the exponential factor can be further interpreted as a time propagator
P ( Γ , t ) = e i L ( Γ ) t
for the probability distribution with invertible and thus reversible time evolution. In summary, Equation (13) can be considered as an universal time propagator that is used to evaluate model calculations in a mathematical consistent framework. Herewith, we are able to map process parameter values to the corresponing outcomes. The concept allows us to consider models for non-stationary and stationary processes by means of an identical mathematical description.

2.3. Coupled Non-Stationary Processes: Model Propagator Dynamics

In terms of our previous discussions, it was shown that the propagator projects the temporal evolution of the process outcome in terms of statistical quantities onto certain fixed time points. Thus, it becomes immediately clear from Equation (13), that the progression of probability distributions in terms of certain time points t M for a set of parameters Γ M can be written as
ρ t M ( A ( Γ ) ) = P ( Γ , t M ) ρ t 0 ( A ( Γ ) )
with the probability distribution ρ t j ( A ( Γ ) ) for observable A ( Γ ) at time t j and P * ( Γ , t M ) = P ( Γ , t M ) in accordance with Equation (11). It has to be noted that Equation (14) is valid for all times as long as the process parameter settings Γ remain unchanged.
For the following discussion, we replace the exact time evolution operator P with a new model operator M including identical time reversal symmetry conditions. In contrast to P ( Γ ) , the set of parameters for M ( κ ) is incomplete in terms of κ Γ and κ Γ . This can be mainly understood with regard to our incomplete knowledge of underlying correlations and effects for most processes as represented by model artifacts [19]. Hence, any model reflects a certain kind of uncertainty such that the model outcomes after the training or parameterization stage need to be carefully evaluated against experimental data, which can be considered as the ground truth distribution [19]. With regard to a change of parameters from Γ 0 κ 1 at t 0 t 1 according to
ρ t 1 ( A ( κ 1 ) ) = M ( κ 1 , t 1 ) ρ t 0 ( A ( Γ 0 ) ) ,
it follows that ρ t 1 ( A ( κ 1 ) ) also differs from the true distribution ρ t 1 ( A ( Γ 1 ) ) . In order to correct model parameter settings with the corresponding experimental distributions, one can introduce the Bayes theorem [39]
ρ ( z i | y i ) = ρ ( y i | z i ) ρ ( y i ) ρ ( z i )
where ρ ( · | · ) denotes the conditional probability as defined by ρ ( y i | z i ) = ρ ( y i z i ) / p ( z i ) where ρ ( y i z i ) denotes the joint probability for outcomes y i and z i . Statistical independence is only evident for ρ ( y i | z i ) = ρ ( y i ) and ρ ( z i | y i ) = ρ ( z i ) . The Bayes theorem thus strongly relies on the prior uninformed probability distribution ρ ( z i ) in combination with the likelihood ρ ( y i | z i ) for the calculation of the posterior distribution ρ ( z i | y i ) . Applying the Bayes theorem on Equation (15), one can show that the resulting posterior distribution can be written as
ρ t 1 ( A ( κ 1 ) | A ( Γ 1 ) ) = ρ t 1 ( A ( Γ 1 ) | A ( κ 1 ) ) ρ t 1 ( A ( Γ 1 ) ) ρ t 1 ( A ( κ 1 ) )
which now explicity incorporates the true experimental distribution of A in terms of the full set of parameters Γ 1 . Furthermore, the corresponding predictive posterior distribution can be computed by
ρ ( A ^ ( κ 1 ) | A ( κ 1 ) ) = ρ t 1 ( A ^ ( κ 1 ) | A ( κ 1 ) , A ( Γ 1 ) ) ρ t 1 ( A ( κ 1 ) | A ( Γ 1 ) ) d A ( κ 1 )
where A ^ ( Γ 1 ) denotes new data points generated by the model at t 1 . In addition, the posterior distribution ρ t 1 ( A ( κ 1 ) | A ( Γ 1 ) ) can be combined with a new model operator including new parameters κ 2 in order to yield
ρ t 2 ( A ( κ 2 ) ) = M ( κ 2 , t 2 ) ρ t 1 ( A ( κ 1 ) | A ( Γ 1 ) )
which can be interpreted as the prior distribution for the next step
ρ t 2 ( A ( κ 2 ) | A ( Γ 2 ) ) = ρ t 2 ( A ( Γ 2 ) | A ( κ 2 ) ) ρ t 2 ( A ( Γ 2 ) ) . ρ t 2 ( A ( κ 2 ) )
In consequence, if we identify the different time points t 0 , t 1 , t N as initial starting points of N individual unit operations as described by different model propagators M 1 ( κ 1 ) , , M N ( κ N ) , one can interpret the corresponding evolution equation as a Bayesian hierarchical model [1] according to
ρ t j ( A ( κ j ) | A ( Γ j ) ) = = ρ t j ( A ( Γ j ) | A ( κ j ) ) ρ t j ( A ( Γ j ) ) M ( κ j , t j ) ρ t j 1 ( A ( κ j 1 ) | A ( κ j 2 ) , , A ( κ 1 ) , A ( Γ 1 ) )
for j = 4 , , N . With regard to the previous relations, one is thus able to project and to combine distribution functions on certain time points in terms of Bayesian inference for an improved prediction of model outcomes. Moreover, it becomes clear that the explicit history of the process in terms of previous parameter settings and process outcomes is well-defined and taken into consideration through conditional probabilities. In more detail, Equation (21) represents the center of our proposed approach. Herewith, we can combine individual unit operation models for stationary and non-stationary processes at fixed time points in terms of transformation functions as represented by prior and posterior distribution functions. The corresponding single model predictions can be used to define a predictive distribution with well-defined confidence intervals, which can be used as prior distribution for the next unit operation model. The corresponding mathematical framework is applicable without any further restriction. In the next section we will show how this approach can be used to refine predictions from different models in order to define suitable transformation functions between coupled unit operations.

2.4. Bayesian Predictive Ensemble: Combination of Different Models

In the previous subsection, we identified Γ and the corresponding probability distribution function ρ ( A ( Γ ) ) as the experimental ground truth without further restrictions. Such an approach is beneficial if a sufficient amount of experimental data for identical process parameter settings are available. Here, we show how the previous relations can be used to assess the uncertainty of model predictions in terms of Bayesian predictive ensembles with regard to coupled model descriptions. It has often been discussed that the accurate calculation of ranges for process outcomes in terms of parametric and non-parametric models is a challenging task. The condition of time inversion for the model propagator introduces a deterministic behavior such that each parameter setting for a fixed model yields an unique model prediction value. Due to these reasons, it is useful to combine different models with reasonable accuracy in order to assess the uncertainty of correct model predictions.
In principle, any modeling approach can be used for the prediction of singular values for the variable A from the probability distribution function ρ ( A ( Γ ) ) = A ( Γ ) + δ ( x Γ ) d x in accordance with Equation (15), such that
A ¯ j ( κ j ) = M j ( κ j ) A ( Γ 0 )
where M j denotes m individual models with the corresponding model parameters κ j with j [ 0 , m ] . Here, we ignore the explicit time-dependence for the sake of clarity. The consideration of all model predictions A ¯ j ( κ j ) in terms of the distribution ρ ( A ¯ 0 ( κ 0 ) , , A ¯ m ( κ m ) ) can be used to evaluate the mean and standard deviation of model predictions. The corresponding values can be used to define an uninformed prior distribution function ρ ( A ¯ ( κ ¯ ) ) in accordance with
ρ ( A ¯ ( κ ¯ ) | A ¯ 0 ( κ 0 ) , , A ¯ m ( κ m ) ) = ρ ( A ¯ 0 ( κ 0 ) , , A ¯ m ( κ m ) | ρ ( A ¯ ( κ ¯ ) ) ρ ( A ¯ 0 ( κ 0 ) , , A ¯ m ( κ m ) ) ρ ( A ¯ ( κ ¯ ) )
which provides the informed posterior distribution ρ ( A ¯ ( κ ¯ ) | ρ ( A ¯ 0 ( κ 0 ) , , A ¯ m ( κ m ) | ρ ( A ¯ ( κ ¯ ) ) ) as needed for the definition of process parameter ranges for coupled models as well (Equation (21)). The corresponding hierarchy of coupled model predictions and Bayesian inference calculations (Bayesian predictive ensemble) can be summarized as follows:
  • Bayesian predictive ensemble: Application of different models with varying accuracy to estimate the prior distribution for a certain unit operation (mean value and standard deviation) in accordance with Equation (22) based on Equations (5) and (13);
  • Bayesian inference to estimate the posterior probability distribution (MCMC calculations) from the prior distribution of step 1 in accordance with Equation (23) based on Equation (22);
  • Representatives from the posterior distribution: Certain choices from the posterior probability distribution as values for models of the next unit operation in accordance with Equation (22);
  • Repeated application of step 1: Calculation of the prior distribution for the next unit operation with values from step 3.
The corresponding steps 1 to 4 represent the basic implementation of the algorithm. As a prerequisite, one parameterizes a set of models for predictions of a certain unit operation. An alternative approach is to use one model that is trained on different sets of training data in terms of the bootstrapping approach. The corresponding predictions of the models for a certain parameter setting are performed in step 1. The set of predictions can be interpreted as a prior distribution with a certain mean value and standard deviation. The prior distribution is then transformed via MCMC calculations into a posterior distribution in accordance with step 2. The resulting posterior distribution as represented by conditional probabilities is interpreted as transformation function to be used for the next unit operation in terms of the underlying input value distribution. Based on the posterior distribution, certain values are randomly drawn, which are then used as new parameter input values in combination with further process parameter settings as required for the next set of models for the subsequent unit operation (step 3). Finally, steps 1, 2, and 3 are repeated for the considered subsequent unit operations.
It has to be noted that the Bayesian predictive ensemble relies on different parametric or non-parametric models without further restriction in combination with directed Bayesian networks [40] in terms of conditional probability distribution functions. The corresponding variations of the predictions from the different models allow us to draw some conclusions on the uncertainty of the model predictions based on inaccurate parameterization or missing information in the data. Moreover, Bayesian approaches are well-suited to be used for low amounts of predictions [39]. Hence, only a limited number of available models can be used for the Bayesian predictive ensemble. Furthermore, it has to be noted that also Bayesian linear regression approaches [41] can be used as prior and posterior distribution functions. If such models are applicable in terms of reliable model predictions, it becomes clear that the time-consuming step 2 in the previous algorithm is already included in the Bayesian linear regression approach. Finally, it has to be noted that the calculation of Equation (23) is a straightforward task with regard to recent numerical MCMC sampling schemes [42]. Noteworthy, the previous conclusions hold as long as the number of considered variables is low. Although also multidimensional Bayesian predictive ensembles can be defined, it has to be noted that certain convergence issues of MCMC steps for multivariate distributions were often observed.

2.5. Combination of Unit Operation Models: Error Propagation

As already discussed, the Bayesian predictive ensemble provides an unbiased estimate for the uncertainty of different model predictions. Despite certain deviations from the ground truth distribution, one can assume that all models are parameterized or trained to provide the best predictions. In general, the difference between two distributions can be studied in terms of the Kullback–Leibler divergence [43], which is defined as
D ( ρ | | q ) = ρ ( A i ) ln ρ ( A i ) q ( A i ) d A i
where ρ ( A ) denotes the model-predicted and q ( A ) the true distribution, respectively. In more detail, the Kullback–Leibler divergence is closely related to information entropy, such that large values of D ( ρ | | q ) reveal a missing knowledge about the true distribution and thus crucial deviations. For point-like predictions of a model δ ( x ) with vanishing variance, it can be seen that D ( ρ | | q ) such that the largest deviation from the ground truth distribution with finite variance can be assumed. This becomes even more important after combining a set of models in terms of coupled process model predictions where the output parameter δ ( x 1 ) is used as an input parameter for the next unit operation and so forth. Due to the logarithmic properties, the Kullback–Leibler divergence is additive for certain distributions D ( ρ 1 , ρ 2 , , ρ N | | q 1 , q 2 , , q N ) = i = 1 N D ( ρ i | | q i ) , which yields in combination with Equation (24) the following expression for point-like predictions
D ( δ 1 , δ 2 , , δ N | | q 1 , q 2 , , q N ) = i = 1 N δ i ( A i ) ln δ i ( A i ) q i ( A i ) d A i
such that D ( ρ 1 , ρ 2 , , ρ N | | q 1 , q 2 , , q N ) . Due to these reasons, it can be concluded that point-like predictions from single as well as coupled unit operation models provide the highest inaccuracy in comparison to the unknown true distributions.
Corresponding conclusions can also be drawn for the consideration of fixed model predicted mean values A i in combination with pre-defined standard deviations σ ( A i ) in terms of Gaussian noise. Hence, inserting the corresponding distributions ρ i ( A i ) = N i ( A i , σ ( A i ) ) into Equation (24) yields
D ( N 1 , N 2 , , N N | | q 1 , q 2 , , q N ) = = i N N i ( A i , σ ( A i ) ) ln N ( A i , σ ( A i ) ) q i ( A i ) d A i ,
which implies that the Kullback–Leibler divergence monotonously grows with
D ( N 1 , N 2 , , N N | | q 1 , q 2 , , q N ) N ln ( N ( x , σ ( x ) ) ) .
In consequence, the model uncertainty for coupled unit operations also grows linearly with O ( N ) , which results in D ( N 1 , N 2 , , N N | | q 1 , q 2 , , q N ) for N .
In contrast, coupled Bayesian predictive ensemble distributions reveal
D ( ρ 1 , ρ 2 , , ρ N | | q 1 , q 2 , , q N ) ln ρ N ( A N | A N 1 , A N 2 , A 1 )
such that D ( ρ 1 , ρ 2 , , ρ N | | q 1 , q 2 , , q N ) stays finite for N . Hence, the accuracy of coupled model predictions grows for point-like predictions via added stochastic noise predictions to conditional probabilities for the Bayesian predictive ensemble method. It is thus of significant importance to consider conditional probabilities for holistic process models in order to reduce the progression of errors and uncertainties to a reasonable level.
By means of holistic process models, it becomes clear that the coupled model design space Ω of parameter settings for Γ becomes accessible [1]. If we focus on a certain process outcome like the product concentration C ( Γ ) as a function of Γ , the corresponding simulations for different parameter settings can be used to find the global optimum value of C ( Γ max ) in accordance with
C ( Γ max ) = max Γ Ω C ( Γ ) .
Noteworthy, the search of global optima ina high dimensional spaces is a challenging task. Besides the fact that it is not clear if one global maximum value exists or even none or many, the search for these optimal conditions is usually performed via stochastic or deterministic optimization algorithms [44,45,46].

3. Numerical Example

3.1. Numerical Details

The Bayesian predictive ensemble, as introduced in the previous section, is used in combination with process steps for further illustration. We consider two simplified unit operations for the sake of clarity, which can be associated with a filtration and a first order chemical reaction step. All of the source code was written in Python 3.9.1 [47] in combination with the modules NumPy 1.19.5 [48], scikit-learn 1.0.1 [49], and PyMC3 3.11.4 [50]. In addition to linear regression (LR), we use random forest (RF) [51], decision tree (DT) [52], extra tree (ET) [53], gradient boosting (GB) [54], and Gaussian process (GP) [55] approaches as non-parametric models for the prediction of process outcomes. The number of estimators in the RF and GB models were chosen as 100 and we used a Gaussian white noise kernel for the GP approach. All Bayesian inference [39] calculations were performed through pyMC3. A standard MCMC approach with an acceptance criterion of 0.95 was used for all calculations. Before effective calculations, we conducted an equilibration phase of 2000 draws for each of the two chains and hereafter the production run of 6000 draws, also each for the two chains. All values are in dimensionless units for the sake of clarity. We mainly focused on ensemble methods like GB, RF, and ET, which are less prone to overfitting.
The first unit operation can be regarded as a filtration step where a concentration A is filtered to achieve a final concentration B. The following linear relation is used
B = χ A + η 1
where the filter constant was chosen as χ = 0.5 , including the measurement uncertainty η 1 ( t ) = 0 and η 1 ( t ) η 1 ( t ) = 0.2 δ ( t t ) . The second unit operation can be regarded as a chemical reaction step in which the concentration B transforms into the final product concentration C according to the first order chemical reaction
C = B e k τ + η 2
with the rate constant k = 1 , η 2 ( t ) = 0 and η 2 ( t ) η 2 ( t ) = 0.1 δ ( t t ) . The choice of such simple equations for unit operations allows us to calculate values of B and C analytically and to compare the outcomes with the model predictions accordingly. Furthermore, we can discuss basic features of the method without having to focus on specifics of more complex unit operations. In terms of a combined approach, the individual models were coupled to predict the outcomes of a holistic process model for the filtration and the reaction step. The coupled unit operation steps are schematically shown in Figure 2. The first unit operation model predicts the concentration B from a fixed value of the concentration A in accordance with Equation (30). From the corresponding posterior distribution of B as calculated by the Bayesian predictive ensemble, certain values were randomly drawn, which are then used for the calculation of C in accordance with Equation (31). Noteworthy, we limit the calculations to only six training data points for each unit operation in order to mimic realistic settings for parameterization purposes.

3.2. Training Phase: Unit Operation Models

For reliable estimates of single unit operation outcomes, we first studied the predictive accuracy of several non-parametric models for each process step individually. As training data, we chose six values of A t r , which were used for the calculation of B t r in accordance with Equation (30). The values of A t r were uniformly distributed between 49 and 51 to mimic certain designs of the experiment conditions [56]. The low amount of training or parameterization data is a standard problem for biotechnological model development as experimental work is time-consuming and costly. The corresponding six data points were used to train individual LR, GP, GB, RF, DT, and ET models. Hereafter, we again draw six randomly chosen values of A as testing data [19] for the calculation of the corresponding B E x p values (Equation (30)) and for predictions of the models B p r e d . The standard values [19], like the root-mean squared error (RMSE), the normalized root-mean squared error (nRMSE), and the mean absolute error (MSE) between predicted B p r e d and analytical values B E x p for the individual models are summarized in Table 1. It can clearly be seen that the predictive accuracy of all models is rather low. This can be rationalized by the low amount of training data and the corresponding influence of the measurement noise on B E x p from Equation (30). However, it comes out that the GP model reveals the highest predictive accuracy despite certain deviations from the predicted values. The corresponding nRMSE value of 0.54 is quite high, but is still acceptable for the sake of our simple example. Based on these values, we chose the GP model as the corresponding representative approach for the filtration unit operation. For the development of different models in accordance with the Bayesian predictive ensemble, we split the previous training data set into four different data sets with a test:training ratio of 20:80. Based on the individual data, we trained four different GP models with each of these reduced training sets in accordance with the concept of bootstrapping [31,32].
A comparable procedure for estimating the model with the highest predictive accuracy was also performed for the chemical reaction unit operation.
We thus first prepared the training data set by choosing six values of B, which were inserted into Equation (31) for the calculation of C E x p . The values of B were uniformly distributed between 24 and 26 to mimic a certain design of experiment conditions [56] and the values of τ were uniformly distributed between 0.4 and 0.6. These data points were used to train individual LR, GP, GB, RF, DT, and ET models. Hereafter, we repeated the same procedure such that we draw six random values from a uniform distribution of B and inserted them into Equation (31) for the calculation of C E x p and for C p r e d from model predictions. The corresponding RMSE, nRMSE, and MAE values for the model predictions are shown in Table 2.
It can clearly be seen that the LR models reveal the highest predictive accuracy. Nevertheless, the nRMSE values are still rather high, which highlights certain deviations between the predicted and the observed values. For reasons of simplicity, we ignore a further improvement of the models and choose the pre-trained LR model as a corresponding representative for the second unit operation. In agreement with the previous unit operation, we split the corresponding training data set into four different data sets with a test:training ratio of 20:80. Based on the individual data, we trained four reduced LR models that are used for the Bayesian predictive ensemble in accordance with bootstrapping calculations [31,32] in the next subsection.

3.3. Bayesian Predictive Ensemble: Single Unit Operations

The corresponding pre-trained GP and LR models are used for the prediction of process outcomes for the filtration and chemical reaction unit operation steps. Hence, we chose an initial value of A = 50.0 , which was inserted into the bootstrapped GP models and the GP model trained with the full data set. The corresponding predictions of B p r e d from the four bootstrapped GP models are shown as red stars in Figure 3. The prediction of B p r e d from the GP model with the full training set is used as mean value for the prior distribution function in combination with the corresponding standard deviation from the bootstrapped GP model predictions in accordance with step 1 of the algorithm. Furthermore, the values of B p r e d from the boostrapped models were used as observed values for the corresponding MCMC calculations (step 2 of the algorithm). The outcomes of these calculations for the posterior distribution B B P E are shown in Figure 3.
In addition to the predictions from the Bayesian predictive ensemble, we also visualized the corresponding distribution of observed values B E x p as reference and the distribution of predictions B B S from the boostrapping calculations as simplified estimates. One can clearly see that the distribution of B B S is different from B B S E in terms of the mean value and the standard deviation. In comparison to the observed values, one can conclude that the distribution of B B P E shows more similarities than B B S in terms of similar mean values and standard deviations. Although the stochastic nature of the experimental values B E x p from six randomly chosen values of A has to be noted, one can assume that the Bayesian predictive ensemble calculations provides a better description than the simplified bootstrapping estimates. This can be partly understood by the low amount of training data, which restricts the number of bootstrapped models such that these predictions are of minor accuracy.
Corresponding conclusions can also be drawn for the predictions of the chemical reaction unit operation as shown in the right side of Figure 3. We chose initial values of B = 24.87 and τ = 0.5 that were inserted into the bootstrapped GP models and the full GP model.The prediction of the full GP model in combination with the standard deviation from the predictions of the bootstrapped GP models were used as mean value and standard deviation, respectively, for the prior distribution function of C (in accordance with step 1 of the algorithm). The corresponding predictions from the boostrapped LR models were used as observed values for the calculation of the posterior distribution function in accordance with step 2 of the algorithm. The posterior distribution of C B P E from the Bayesian predictive ensemble calculations after the MCMC steps in combination with the bootstrapping distribution C B S are shown as black and red curves, respectively. It can be seen that the mean values C B S and C B P E coincide, but the standard deviations differ. Hence, the standard deviation from the BPE calculations is slightly broader and thus overlaps with the corresponding observed experimental distribution function C E x p . Although the accuracy is not that high when compared to the previous unit operation, it can be concluded that the combined BPE calculations provide reliable estimates for the distribution of process outcomes with higher accuracy than simple Bootstrapping predictions. The corresponding mean values and standard deviations for all distributions and unit operations are shown in Table 3.
Moreover, we included point-like predictions for B and C for the corresponding values from the full GP and the full LR model. As can be seen, the best ML models provide good estimates but no reasonable standard deviation for estimating the model uncertainty. However, their usage as mean value for the prior distribution function allows us to improve the corresponding values for the posterior distribution function. Hence, it can be concluded that the combination of both aspects in terms of the Bayesian predictive ensemble calculations provides good estimates for model outcome ranges and mean values.

3.4. Coupled Unit Operations: Holistic Process Models

In this subsection, we evaluate the corresponding model predictions for the coupled unit operation steps of filtration and chemical reaction. For all predictions, the pre-trained models from the previous subsection were used. The initial starting value for the first unit operation (filtration) was chosen as A = 25.0 . Thus, the calculations as well as the values for the distribution of B p r e d C were identical to the procedure as described in the previous subsection. The corresponding posterior distribution B B P E C when considered as transformation function was used as starting distribution for the second unit operation in accordance with step 2 of the algorithm. We draw six random samples from the posterior distribution B p r e d C in combination with a fixed value of τ = 0.5 (Equation (31)) which are used as input values for the corresponding pre-trained bootstrapped LR models for the chemical reaction step in accordance with step 3 of the algorithm. The outcomes of these bootrapped GP models are used as observed values for the calculation of the posterior distribution function C B P E C . Moreover, we used the six random values of B B P E C as initial values for the full LR model. The resulting distribution is used as prior distribution function for the MCMC calculations in accordance with step 2 for the calculation of C p r e d C . The corresponding posterior distributions for the coupled outcomes B p r e d C and C p r e d C are shown in Figure 4 in combination with the observed coupled values B E x p C and C E x p C .
It can clearly be seen that the distribution for C B P E C is roughly identical when compared to C B P E for single unit operation models (Table 3). For single unit operation models, we already used as input values a comparable value of B = 24.87 in order to estimate the accuracy of the approaches after coupling. Hence, it can be concluded that the coupled approaches reveal a high accuracy with reliable estimates for the predicted distributions. Thus, all corresponding results reveal that the Bayesian predictive ensemble can be regarded as a fast and straightforward approach to estimate the mean outcomes for coupled and individual unit operation models in good agreement with the observed distribution.

4. Summary and Conclusions

In this study, we presented a new approach in accordance with Bayesian inference and various model predictions for the advanced calculation of single unit as well as coupled end-to-end process outcomes in combination with the corresponding uncertainty ranges. The Bayesian predictive ensemble method provides a more robust interpretation of model predictions and parameter ranges when compared to discrete or stochastic connections. Due to the coupling of individual unit operations, a consistent consideration of conditional probability distributions is introduced, which crucially relies on pre-defined parameter uncertainty ranges. Hence, the standard deviations reflect the general model uncertainty in combination with the variability of the observed data which was used for parameterization or training purposes.
It has to be noted that this concept is also applicable for high-dimensional design spaces as well as multivariate parameter distribution functions with certain limitations in terms of MCMC convergence issues. Furthermore, the prior distributions for the Bayesian inference calculations can include as much information as possible, including experimental data as well, which thus increases the accuracy of the posterior distribution functions. In addition to using different model types like neural network- or decision tree-based approaches, it is also possible to focus on a single type of model with slight changes in the hyper parameter settings or in combination with boostrapping concepts. It has to be noted that the Bayesian framework reveals its full advantages for a small number of model predictions. Herewith, we reweigh the corresponding model predictions under the assumption that the majority of models shows a sufficient predictive accuracy.
In summary, we presented a robust, extensible, and consistent approach for single and coupled unit operation models. The main advantages are the probabilistic interpretation of outcome ranges in terms of uncertainty intervals. In consequence, we explicitly use different models for the prediction of process outcomes. In terms of Bayesian inference and the consideration of conditional probabilities, we incorporate the outcomes of previous unit operations as implicit knowledge into the resulting posterior distributions. Thereby, we present a consistent approach to transfer the resulting distributions with meaningful and interpretable ranges between the individual unit operations. In contrast to other approaches, the proposed transfer functions including prior and posterior distributions have well defined and interpretable ranges. Furthermore, we have shown that the deviations from the true distribution deviate only marginally after certain process steps in contrast to other approaches. Hence, our approach provides a controllable and finite error progression for any holistic process description in terms of parametric or non-parametric models. Noteworthy, the method is also applicable for limited experimental data sets in order to obtain reliable distributions from MCMC calculations. Despite all benefits, it has to be noted that the MCMC calculations are rather time consuming and may show some convergence issues for poorly behaving model predictions. The computational costs for these calculations increase linearly with the amount of considered unit operations and the number of dimensions in the model design space. It has to be noted that our concept is not restricted to chemical or biotechnological process models, as it also can be used for other machine learning applications. As an outlook, we assume that the presented approach provides a robust basis for the development of future holistic process models to support the development of optimized and robust manufacturing processes.

Author Contributions

Conceptualization, J.S.; methodology, J.S.; formal analysis, J.S.; validation, J.S., T.E., I.-T.H., B.S., M.M. and E.B.; supervision, J.S., L.M.H., R.G., M.L., E.B. and A.J.; project administration, J.S., R.G., M.L., A.J. and E.B.; funding acqusition: J.S., L.M.H., A.J. and E.B.; writing—original draft preparation: J.S., L.M.H., T.E. and E.B, writing—review and editing, all authors. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Boehringer Ingelheim Pharma GmbH & Co. KG, Development Biologicals and the Digital Innovation Unit.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We thank Hermann Schuchnigg for valuable discussions. We acknowledge funding from the Digital Innovation Unit and Boehringer Ingelheim Pharma GmbH & Co. KG.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
NORNormal operation range
PARProven acceptable range
PATProcess Analytical Technology
BSBootstrapping
MCMCMarkov Chain Monte Carlo
LRLinear Regression
GBGradient Boosting
DTDecision Trees
ETExtra Trees
GPGaussian Processes
RFRandom Forest
RMSERoot-Mean Squared Error
nRMSENormalized Root-Mean Squared Error
MAEMean Absolute Error
BPEBayesian Predictive Ensemble

References

  1. Smiatek, J.; Jung, A.; Bluhmki, E. Towards a Digital Bioprocess Replica: Computational Approaches in Biopharmaceutical Development and Manufacturing. Trends Biotechnol. 2020, 38, 1141–1153. [Google Scholar] [CrossRef] [PubMed]
  2. Brass, J.M.; Hoeks, F.W.J.M.M.; Rohner, M. Application of Modelling Techniques for the Improvement of Industrial Bioprocesses. J. Biotechnol. 1997, 59, 63–72. [Google Scholar] [CrossRef]
  3. Narayanan, H.; Luna, M.F.; von Stosch, M.; Cruz Bournazou, M.N.; Polotti, G.; Morbidelli, M.; Butté, A.; Sokolov, M. Bioprocessing in the Digital Age: The Role of Process Models. Biotechnol. J. 2020, 15, 1900172. [Google Scholar] [CrossRef] [PubMed]
  4. Herwig, C.; Pörtner, R.; Möller, J. (Eds.) Digital Twins: Applications to the Design and Optimization of Bioprocesses; Advances in Biochemical Engineering and Biotechnology; Springer Nature: Berlin/Heidelberg, Germany, 2021; Volume 177. [Google Scholar]
  5. Narayanan, H.; Sokolov, M.; Morbidelli, M.; Butté, A. A New Generation of Predictive Models—The added Value of Hybrid Models for Manufacturing Processes of Therapeutic Proteins. Biotechnol. Bioeng. 2019, 116, 2540–2549. [Google Scholar] [CrossRef]
  6. von Stosch, M.; Davy, S.; Francois, K.; Galvanauskas, V.; Hamelink, J.M.; Luebbert, A.; Mayer, M.; Oliveira, R.; O’Kennedy, R.; Rice, P.; et al. Hybrid Modeling for Quality by Design and PAT—Benefits and Challenges of Applications in Biopharmaceutical Industry. Biotechnol. J. 2014, 9, 719–726. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Bayer, B.; von Stosch, M.; Striedner, G.; Duerkop, M. Comparison of Modeling Methods for DoE-Based Holistic Upstream Process Characterization. Biotechnol. J. 2020, 15, 1900551. [Google Scholar] [CrossRef] [Green Version]
  8. Sokolov, M.; Soos, M.; Neunstoecklin, B.; Morbidelli, M.; Butté, A.; Leardi, R.; Solacroup, T.; Stettler, M.; Broly, H. Fingerprint Detection and Process Prediction by Multivariate Analysis of Fed-Batch Monoclonal Antibody Cell Culture Data. Biotechnol. Prog. 2015, 31, 1633–1644. [Google Scholar] [CrossRef]
  9. Rischawy, F.; Saleh, D.; Hahn, T.; Oelmeier, S.; Spitz, J.; Kluters, S. Good Modeling Practice for Industrial Chromatography: Mechanistic Modeling of Ion Exchange Chromatography of a Bispecific Antibody. Comput. Chem. Eng. 2019, 130, 106532. [Google Scholar] [CrossRef]
  10. Briskot, T.; Stückler, F.; Wittkopp, F.; Williams, C.; Yang, J.; Konrad, S.; Doninger, K.; Griesbach, J.; Bennecke, M.; Hepbildikler, S.; et al. Prediction Uncertainty Assessment of Chromatography Models using Bayesian Inference. J. Chromatogr. A 2019, 1587, 101–110. [Google Scholar] [CrossRef]
  11. Saleh, D.; Wang, G.; Müller, B.; Rischawy, F.; Kluters, S.; Studts, J.; Hubbuch, J. Straightforward Method for Calibration of Mechanistic Cation Exchange Chromatography Models for Industrial Applications. Biotechnol. Prog. 2020, 36, e2984. [Google Scholar] [CrossRef]
  12. Ulonska, S.; Kroll, P.; Fricke, J.; Clemens, C.; Voges, R.; Müller, M.M.; Herwig, C. Workflow for Target-Oriented Parametrization of an Enhanced Mechanistic Cell Culture Model. Biotechnol. J. 2018, 13, 1700395. [Google Scholar] [CrossRef] [PubMed]
  13. Moser, A.; Appl, C.; Brüning, S.; Hass, V.C. Mechanistic Mathematical Models as a Basis for Digital Twins. In Digital Twins; Springer: Berlin/Heidelberg, Germany, 2020; pp. 133–180. [Google Scholar]
  14. Narayanan, H.; Seidler, T.; Luna, M.F.; Sokolov, M.; Morbidelli, M.; Butté, A. Hybrid Models for the Simulation and Prediction of Chromatographic Processes for Protein Capture. J. Chromatogr. A 2021, 1650, 462248. [Google Scholar] [CrossRef]
  15. Smiatek, J.; Clemens, C.; Herrera, L.M.; Arnold, S.; Knapp, B.; Presser, B.; Jung, A.; Wucherpfennig, T.; Bluhmki, E. Generic and Specific Recurrent Neural Network Models: Applications for Large and Small Scale Biopharmaceutical Upstream Processes. Biotechnol. Rep. 2021, 31, e00640. [Google Scholar] [CrossRef]
  16. Narayanan, H.; Dingfelder, F.; Condado Morales, I.; Patel, B.; Heding, K.E.; Bjelke, J.R.; Egebjerg, T.; Butté, A.; Sokolov, M.; Lorenzen, N.; et al. Design of Biopharmaceutical Formulations Accelerated by Machine Learning. Mol. Pharm. 2021, 18, 3843–3853. [Google Scholar] [CrossRef] [PubMed]
  17. Narayanan, H.; Dingfelder, F.; Butté, A.; Lorenzen, N.; Sokolov, M.; Arosio, P. Machine Learning for Biologics: Opportunities for Protein Engineering, Developability, and Formulation. Trends Pharmacol. Sci. 2021, 42, 151–165. [Google Scholar] [CrossRef] [PubMed]
  18. Sokolov, M. Decision Making and Risk Management in Biopharmaceutical Engineering—Opportunities in the Age of Covid-19 and Digitalization. Ind. Eng. Chem. Res. 2020, 59, 17587–17592. [Google Scholar] [CrossRef]
  19. Smiatek, J.; Jung, A.; Bluhmki, E. Validation Is Not Verification: Precise Terminology and Scientific Methods in Bioprocess Modeling. Trends Biotechnol. 2021, 39, 1117–1119. [Google Scholar] [CrossRef]
  20. Rajamanickam, V.; Babel, H.; Montano-Herrera, L.; Ehsani, A.; Stiefel, F.; Haider, S.; Presser, B.; Knapp, B. About Model Validation in Bioprocessing. Processes 2021, 9, 961. [Google Scholar] [CrossRef]
  21. Nargund, S.; Guenther, K.; Mauch, K. The Move toward Biopharma 4.0: Insilico Biotechnology Develops "Smart" Processes that Benefit Biomanufacturing through Digital Twins. Gen. Eng. Biotechnol. 2019, 39, 53–55. [Google Scholar] [CrossRef]
  22. Zahel, T.; Hauer, S.; Mueller, E.; Murphy, P.; Abad, S.; Vasilieva, E.; Maurer, D.; Brocard, C.; Reinisch, D.; Sagmeister, P.; et al. Integrated Process Modeling—A Process Validation Life Cycle Companion. Bioengineering 2017, 4, 86. [Google Scholar] [CrossRef] [Green Version]
  23. Chen, Y.; Yang, O.; Sampat, C.; Bhalode, P.; Ramachandran, R.; Ierapetritou, M. Digital Twins in Pharmaceutical and Biopharmaceutical Manufacturing: A Literature Review. Processes 2020, 8, 1088. [Google Scholar] [CrossRef]
  24. Park, S.Y.; Park, C.H.; Choi, D.H.; Hong, J.K.; Lee, D.Y. Bioprocess Digital Twins of Mammalian Cell Culture for Advanced Biomanufacturing. Curr. Opin. Chem. Eng. 2021, 33, 100702. [Google Scholar] [CrossRef]
  25. Taylor, C.; Marschall, L.; Kunzelmann, M.; Richter, M.; Rudolph, F.; Vajda, J.; Presser, B.; Zahel, T.; Studts, J.; Herwig, C. Integrated Process Model Applications Linking Bioprocess Development to Quality by Design Milestones. Bioengineering 2021, 8, 156. [Google Scholar] [CrossRef] [PubMed]
  26. Zobel-Roos, S.; Schmidt, A.; Uhlenbrock, L.; Ditz, R.; Köster, D.; Strube, J. Digital Twins in Biomanufacturing. In Digital Twins; Springer: Berlin/Heidelberg, Germany, 2021; pp. 181–262. [Google Scholar]
  27. Gomis-Fons, J.; Schwarz, H.; Zhang, L.; Andersson, N.; Nilsson, B.; Castan, A.; Solbrand, A.; Stevenson, J.; Chotteau, V. Model-Based Design and Control of a Small-Scale Integrated Continuous End-to-End mAb Platform. Biotechnol. Prog. 2020, 36, e2995. [Google Scholar] [CrossRef] [PubMed]
  28. Appl, C.; Moser, A.; Baganz, F.; Hass, V.C. Digital Twins for Bioprocess Control Strategy Development and Realisation. In Digital Twins; Springer: Berlin/Heidelberg, Germany, 2020; pp. 63–94. [Google Scholar]
  29. Maloney, A.J.; Icten, E.; Capellades, G.; Beaver, M.G.; Zhu, X.; Graham, L.R.; Brown, D.B.; Griffin, D.J.; Sangodkar, R.; Allian, A.; et al. A Virtual Plant for Integrated Continuous Manufacturing of a Carfilzomib Drug Substance Intermediate, part 3: Manganese-Catalyzed Asymmetric Epoxidation, Crystallization, and Filtration. Organ. Proc. Res. Dev. 2020, 24, 1891–1908. [Google Scholar] [CrossRef]
  30. DiCiccio, T.J.; Efron, B. Bootstrap Confidence Intervals. Statist. Sci. 1996, 11, 189–228. [Google Scholar] [CrossRef]
  31. Efron, B. Bootstrap Methods: Another Look at the Jackknife. In Breakthroughs in Statistics: Methodology and Distribution; Kotz, S., Johnson, N.L., Eds.; Springer: New York, NY, USA, 1992; pp. 569–593. [Google Scholar]
  32. Efron, B. Second Thoughts on the Bootstrap. Stat. Sci. 2003, 18, 135–140. [Google Scholar] [CrossRef]
  33. Efron, B. Bayes, Oracle Bayes and Empirical Bayes. Stat. Sci. 2019, 34, 177–201. [Google Scholar] [CrossRef]
  34. Efron, B. Bayes’ Theorem in the 21st Century. Science 2013, 340, 1177–1178. [Google Scholar] [CrossRef]
  35. Roman, R.C.; Precup, R.E.; Petriu, E.M. Hybrid Data-Driven Fuzzy Active Disturbance Rejection Control for Tower Crane Systems. Eur. J. Control 2021, 58, 373–387. [Google Scholar] [CrossRef]
  36. Zhu, Z.; Pan, Y.; Zhou, Q.; Lu, C. Event-Triggered Adaptive Fuzzy Control for Stochastic Nonlinear Systems with Unmeasured States and Unknown Backlash-Like Hysteresis. IEEE Transact. Fuzzy Sys. 2020, 29, 1273–1283. [Google Scholar] [CrossRef]
  37. Walters, P. An Introduction to Ergodic Theory; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  38. Prigogine, I. Non-Equilibrium Statistical Mechanics; Courier Dover Publications: Mineola, NY, USA, 2017. [Google Scholar]
  39. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006. [Google Scholar]
  40. Jensen, F.V. Bayesian Networks and Decision Graphs; Springer: Berlin/Heidelberg, Germany, 2001. [Google Scholar]
  41. Carlin, B.P.; Louis, T.A. Bayesian Methods for Data Analysis; Chapman and Hall/CRC: London, UK, 2008. [Google Scholar]
  42. Geyer, C.J. Introduction to Markov Chain Monte Carlo. In Chapman & Hall/CRC Handbooks of Modern Statistical Methods; CRC Press: Boca Raton, FL, USA, 2011. [Google Scholar]
  43. Kullback, S.; Leibler, R.A. On Information and Sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  44. Homem-de Mello, T.; Bayraksan, G. Monte Carlo Sampling-Based Methods for Stochastic Optimization. Surv. Oper. Res. Manag. Sci. 2014, 19, 56–85. [Google Scholar] [CrossRef]
  45. Fouskakis, D.; Draper, D. Stochastic Optimization: A Review. Int. Stat. Rev. 2002, 70, 315–349. [Google Scholar] [CrossRef]
  46. Ruder, S. An Overview of Gradient Descent Optimization Algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  47. Van Rossum, G.; Drake, F.L. Python 3 Reference Manual; CreateSpace: Scotts Valley, CA, USA, 2009. [Google Scholar]
  48. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array Programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef]
  49. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  50. Salvatier, J.; Wiecki, T.V.; Fonnesbeck, C. Probabilistic Programming in Python using PyMC3. PeerJ Comput. Sci. 2016, 2, e55. [Google Scholar] [CrossRef] [Green Version]
  51. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  52. Kamiński, B.; Jakubczyk, M.; Szufel, P. A Framework for Sensitivity Analysis of Decision Trees. Cent. Eur. J. Oper. Res. 2018, 26, 135–159. [Google Scholar] [CrossRef]
  53. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely Randomized Trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  54. Mason, L.; Baxter, J.; Bartlett, P.; Frean, M. Boosting Algorithms as Gradient Descent. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 1999; Volume 12, pp. 512–518. [Google Scholar]
  55. Williams, C.K.; Rasmussen, C.E. Gaussian Processes for Machine Learning; MIT Press: Cambridge, MA, USA, 2006. [Google Scholar]
  56. Politis, S.N.; Colombo, P.; Colombo, G.; Rekkas, D.M. Design of Experiments (DoE) in Pharmaceutical Development. Drug Dev. Ind. Pharm. 2017, 43, 889–901. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Top row in blue color: Experimental set-up of three unit operations. The output values of the individual unit operations B and C are used as input values for the next unit operations. In addition, there exist certain CQA values such as A1, B1, D1, and the final process outcome D. The corresponding experimental data is used for the parameterization or the training of models for the individual unit operations (bottom row in green color). Three models represent the process behavior at the unit operation stage and can be used for predictions. The results of the model predictions are used as input values for the next models in combination with further CQAs. The corresponding transfer functions with the model outputs are represented as green horizontal arrows.
Figure 1. Top row in blue color: Experimental set-up of three unit operations. The output values of the individual unit operations B and C are used as input values for the next unit operations. In addition, there exist certain CQA values such as A1, B1, D1, and the final process outcome D. The corresponding experimental data is used for the parameterization or the training of models for the individual unit operations (bottom row in green color). Three models represent the process behavior at the unit operation stage and can be used for predictions. The results of the model predictions are used as input values for the next models in combination with further CQAs. The corresponding transfer functions with the model outputs are represented as green horizontal arrows.
Processes 10 00662 g001
Figure 2. Schematic representation of the coupled unit operations in terms of a Bayesian network [40]: (i) filtering and (ii) chemical reaction. The models predict the concentration B from a fixed value of the concentration A = 25.0 in accordance with Equation (30). From the corresponding posterior distribution of B as calculated by the Bayesian predictive ensemble, 100 values were randomly drawn, which were then used for the calculation of C in accordance with Equation (31). The species B react to species C with a certain time constant κ F .
Figure 2. Schematic representation of the coupled unit operations in terms of a Bayesian network [40]: (i) filtering and (ii) chemical reaction. The models predict the concentration B from a fixed value of the concentration A = 25.0 in accordance with Equation (30). From the corresponding posterior distribution of B as calculated by the Bayesian predictive ensemble, 100 values were randomly drawn, which were then used for the calculation of C in accordance with Equation (31). The species B react to species C with a certain time constant κ F .
Processes 10 00662 g002
Figure 3. (Left side): Individual model predictions of B for a value of A = 50.0 from the bootstrapped GP models (red stars) in combination with the mean value from the observed values (blue circle). The blue curve shows the observed experimental distribution, the red curve the simplified estimate from the boostrapped GP model predictions, and the black curve the corresponding distribution of B from the Bayesian predictive ensemble calculations. (Right side): Individual model predictions of C for values of B = 24.87 and τ = 0.5 from the bootstrapped GP models (red stars) in combination with the mean value from the observed values (blue circle). The blue curve shows the observed experimental distribution, the red curve the simplified estimate from the boostrapped GP model predictions, and the black curve the corresponding distribution of C from the Bayesian predictive ensemble calculations. All units are dimensionless.
Figure 3. (Left side): Individual model predictions of B for a value of A = 50.0 from the bootstrapped GP models (red stars) in combination with the mean value from the observed values (blue circle). The blue curve shows the observed experimental distribution, the red curve the simplified estimate from the boostrapped GP model predictions, and the black curve the corresponding distribution of B from the Bayesian predictive ensemble calculations. (Right side): Individual model predictions of C for values of B = 24.87 and τ = 0.5 from the bootstrapped GP models (red stars) in combination with the mean value from the observed values (blue circle). The blue curve shows the observed experimental distribution, the red curve the simplified estimate from the boostrapped GP model predictions, and the black curve the corresponding distribution of C from the Bayesian predictive ensemble calculations. All units are dimensionless.
Processes 10 00662 g003
Figure 4. Posterior distributions of B p r e d C (green solid curve) and C p r e d C (blue solid curve) from Bayesian predictive ensemble calculations for coupled unit operations. The corresponding observed values are shown as dashed curves. More details can be found in the main text. All units are dimensionless.
Figure 4. Posterior distributions of B p r e d C (green solid curve) and C p r e d C (blue solid curve) from Bayesian predictive ensemble calculations for coupled unit operations. The corresponding observed values are shown as dashed curves. More details can be found in the main text. All units are dimensionless.
Processes 10 00662 g004
Table 1. Root-mean squared errors (RMSE), normalized root-mean squared errors (nRMSE), mean absolute errors (MAE), and Pearson correlation coefficients R 2 for the individual models regarding the test data between predicted B p r e d and observed values B E x p for the filtration step.
Table 1. Root-mean squared errors (RMSE), normalized root-mean squared errors (nRMSE), mean absolute errors (MAE), and Pearson correlation coefficients R 2 for the individual models regarding the test data between predicted B p r e d and observed values B E x p for the filtration step.
ApproachRMSEnRMSEMAE
GP0.190.540.03
LR0.190.570.04
RF0.340.990.11
ET0.361.060.13
DT0.391.140.15
GB0.391.140.15
Table 2. Root-mean squared errors (RMSE), normalized root-mean squared errors (nRMSE), mean absolute errors (MAE), and Pearson correlation coefficients R 2 for the individual models regarding the test data between predicted C p r e d and observed values C E x p for the chemical reaction unit operation step.
Table 2. Root-mean squared errors (RMSE), normalized root-mean squared errors (nRMSE), mean absolute errors (MAE), and Pearson correlation coefficients R 2 for the individual models regarding the test data between predicted C p r e d and observed values C E x p for the chemical reaction unit operation step.
ApproachRMSEnRMSEMAE
LR0.250.250.06
ET0.570.550.32
RF0.710.690.51
GB0.790.770.63
DT0.830.810.69
GP1.271.241.62
Table 3. Mean values B , c , standard deviations σ B , σ C for experimental values (Exp), and calculations from the Bayesian predictive ensemble (BPE) and the simple estimates from the bootstrapped models (BS). The results of the best ML model for a singular prediction of B (full GP model) and C (full LR model) are also shown.
Table 3. Mean values B , c , standard deviations σ B , σ C for experimental values (Exp), and calculations from the Bayesian predictive ensemble (BPE) and the simple estimates from the bootstrapped models (BS). The results of the best ML model for a singular prediction of B (full GP model) and C (full LR model) are also shown.
MethodBσBCσC
Exp24.930.1315.160.11
BS24.850.0615.230.18
BPE24.930.0915.230.21
ML25.040.0015.640.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Montano Herrera, L.; Eilert, T.; Ho, I.-T.; Matysik, M.; Laussegger, M.; Guderlei, R.; Schrantz, B.; Jung, A.; Bluhmki, E.; Smiatek, J. Holistic Process Models: A Bayesian Predictive Ensemble Method for Single and Coupled Unit Operation Models. Processes 2022, 10, 662. https://doi.org/10.3390/pr10040662

AMA Style

Montano Herrera L, Eilert T, Ho I-T, Matysik M, Laussegger M, Guderlei R, Schrantz B, Jung A, Bluhmki E, Smiatek J. Holistic Process Models: A Bayesian Predictive Ensemble Method for Single and Coupled Unit Operation Models. Processes. 2022; 10(4):662. https://doi.org/10.3390/pr10040662

Chicago/Turabian Style

Montano Herrera, Liliana, Tobias Eilert, I-Ting Ho, Milena Matysik, Michael Laussegger, Ralph Guderlei, Bernhard Schrantz, Alexander Jung, Erich Bluhmki, and Jens Smiatek. 2022. "Holistic Process Models: A Bayesian Predictive Ensemble Method for Single and Coupled Unit Operation Models" Processes 10, no. 4: 662. https://doi.org/10.3390/pr10040662

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop