Skip to Content
EntropyEntropy
  • Article
  • Open Access

10 February 2026

Pareto-Front Optimization of Variance-Added Expected Loss with Interrelated Qualities

and
Department of Industrial Engineering, Hanyang University, Seoul 04763, Republic of Korea
*
Author to whom correspondence should be addressed.

Abstract

In industries, particularly in quality optimization, the trade-off between model bias and variance is inevitable, reflecting the tension between accuracy and uncertainty. Traditional methods often address these aspects separately, potentially leading to suboptimal decisions. This study proposes a Pareto-front optimization framework for a variance-added expected loss function within the context of interrelated quality characteristics. By integrating multivariate quadratic loss with a variance term, our approach simultaneously captures deviation from targets (bias) and system uncertainty (variance). Unlike sequential approaches that first minimize bias and then variance—often increasing total risk—our weighted formulation flexibly adjusts for their trade-offs. This enables a more balanced and efficient optimization process that identifies solutions with lower overall risk. Through Pareto-front analysis, we reveal trade-offs between expected loss and variance, allowing users to select optimal quality designs based on their preferred bias–variance balance. Representative examples and a case study adopted from the literature validate the effectiveness and practical applicability of the proposed method.

1. Introduction

In modern industries, achieving optimal performance across multiple interrelated quality attributes is a complex, yet essential, task. Systems often face conflicting objectives, in which improving one quality indicator may deteriorate another. As industries demand higher precision and robustness, efficiently managing these trade-offs becomes increasingly critical. In a machine-learning task, for example, a trade-off between model bias and model variance is inevitable: a low-level model bias from substantial model complexity incurs high-level model variance, and vice versa. Representative quality objectives in conflict are accuracy and variance arising from predefined targets and random observations.
Traditional optimization methods typically focus on minimizing a single objective loss function, but they often fail to account for interdependencies between different quality attributes. Many real-world systems such as manufacturing, engineering, and healthcare exhibit tightly coupled quality characteristics that directly influence overall system performance. In such contexts, a more comprehensive approach is required—one that considers not only deviations from target quality but also uncertainty in system behavior by incorporating variance into overall loss.
This paper proposes a novel approach for globally optimizing interrelated quality attributes by adding variance-induced uncertainty to expected loss. The goal is to minimize not only deviation from a target quality but also uncertainty in performance. For the sake of global optimization, we use Pareto fronts to explore trade-offs between conflicting objectives, enabling decision-makers to select solutions, or decisions, that achieve the desired balance between accuracy and risk. Importantly, our method allows users to explicitly choose the ratio between bias and variance terms, providing greater flexibility to manage the trade-offs arising from specific application requirements.
In this study, we apply an alternative global criterion (GC)-based method for generating Pareto fronts to handle accuracy-induced and variance-induced qualities. This method offers increased computational efficiency compared to traditional optimization techniques and effectively handles the trade-offs between multiple objectives. We illustrate the applicability of this approach through two representative examples and demonstrate its practical use with a dataset adopted from the literature. The experiments reveal the applicability of the proposed method to a complex system with interrelated quality attributes. Our analysis shows that the optimization results outperform existing quality decisions in terms of both quality and flexibility, providing valuable insights into achieving quality optimization in industrial settings.
The proposed method presented in this paper offers a systematic framework for quality optimization in complex systems with interrelated attributes and variance. By applying this approach, decision-making processes in industries requiring simultaneous optimization of multiple quality factors can be significantly improved.

2. Related Works

In multi-objective optimization problems, it is often necessary to simultaneously consider distinct quality characteristics such as mean deviation and variance. As a result, conventional single-objective optimization techniques have limited applicability, and various multi-objective quality optimization methods have been proposed.
A well-known early approach is the two-step strategy proposed by [1], in which variance is minimized first, followed by minimization of mean deviation. This method is computationally simple and easy to interpret, making it appealing in practical settings. However, in real-world applications, most design variables influence both the mean and variance simultaneously. Thus, separating these objectives can lead to inaccurate decisions and reduce the overall predictive accuracy and interpretability of the optimization process.
Other studies have sought to mathematically incorporate interactions between multiple quality characteristics. For example, Shiau proposed a weighted sum of signal-to-noise (SN) ratios to aggregate multiple quality characteristics [2], while Tong et al. developed a multi-response SN ratio that sums weighted expected losses before converting the result to an overall performance measure [3]. More recently, a loss function explicitly considering interactions between variables has been proposed (see, e.g., [4]), offering a more theoretically robust formulation compared to earlier methods. However, these approaches share practical limitations: they rely mainly on numerical computations from existing data, do not support the setting of design variables or the generation of actionable design alternatives, and thus are generally restricted to post hoc analysis rather than active design optimization.
Building on such interaction-based loss functions, Costa and Lourenço [5] proposed a game-theoretic approach to simultaneously optimize process and product quality characteristics in bi-objective settings. By employing Stackelberg’s strategy, their method avoids subjective parameter tuning and seeks efficient trade-offs. Nonetheless, such game-theoretic frameworks often struggle to yield a solution that simultaneously satisfies all quality objectives, especially when multiple interdependent characteristics are involved. This underscores the need for methods that not only capture interactions among characteristics but also offer clear and actionable design recommendations.
To address such multi-objective optimization challenges, especially in design settings, surrogate-based approaches have gained popularity. A representative example is the efficient global optimization (EGO) algorithm [6], which uses the expected improvement (EI) criterion to sequentially explore the design space and identify optima. Ref. [7] extended this framework to multi-objective problems by introducing ParEGO, an algorithm that employs Tchebycheff scalarization to search the Pareto front. Furthermore, ref. [8] proposed an integrated approach combining multiple scalarization techniques (e.g., Tchebycheff and weighted sum) to better approximate diverse trade-offs. However, these methods inherently reduce multi-objective problems to a series of single-objective subproblems and, as a result, cannot fully capture the complex trade-offs among conflicting objectives.
In this study, we propose a multivariate quadratic loss-based Pareto optimization framework that mathematically captures the interactions between design variables and multiple quality characteristics, and simultaneously accounts for overall quality loss arising from both mean deviation and variance. In contrast to scalarization-based methods, the proposed approach directly derives interpretable Pareto-optimal solutions in the multi-objective space without reducing the problem to a single-objective formulation. For reproducibility, we provide a GitHub URL (https://github.com/KIM-sang-won/PFOVELIQ/blob/main/PFOVE__Python_code.ipynb, accessed on 4 February 2026) for the experiments.

3. Methods

3.1. Multivariate Quadratic Loss Function and Total Risk

A multivariate quadratic loss function was introduced by Pignatiello [1] and is defined as follows:
L ( y ) = ( y τ ) T C ( y τ ) ,
where y = ( y 1 , , y p ) T is a p × 1 vector consisting of principal components (PCs) of the ’nominal-the-best’ type, and τ = ( τ 1 , , τ p ) T is a p × 1 vector containing the corresponding target values. The  p × p positive definite cost matrix C assigns weights to deviations among different PCs.
The elements of C are determined by defining the loss in terms of deviations. Let z i = y i τ i denote the deviation of the i-th characteristic from its target. For example, in a bivariate case ( p = 2 ), the loss function is expressed as:
L ( y ) = c 11 z 1 2 + c 22 z 2 2 + 2 c 12 z 1 z 2 ,
in which c i i represents the cost coefficient associated with the squared deviation z i 2 , and  c i j , i j , accounts for the interaction effect between deviations z i and z j . By default, y is of the nominal-the-best type, but it can be adjusted to the lower-the-better or higher-the-better type by modifying τ accordingly.
Furthermore, Pignatiello [1] assumes that y follows a p-dimensional multivariate normal distribution with mean vector μ = ( μ 1 , , μ p ) T and p × p variance–covariance matrix Σ = ( σ i j 2 ) , i.e.,
y N ( μ , Σ ) ,
where μ represents the expected values of the PCs, and  Σ captures their variability and correlations.
The expected quadratic loss can be decomposed into bias and variance contributions:
E [ L ( y ) ] = ( μ τ ) T C ( μ τ ) + trace ( C Σ ) ,
where the first term measures the deviation of the mean from the target, and the second term represents the variance contribution weighted by C .
We define the total risk T by adding the variance contribution weighted by the parameter λ :
T = E [ L ( y ) ] + λ trace ( C Σ ) = ( μ τ ) T C ( μ τ ) + ( 1 + λ ) trace ( C Σ ) ,
where we used E [ L ( y ) ] = ( μ τ ) T C ( μ τ ) + trace ( C Σ ) . The term trace ( C Σ ) derives from the identity
E [ ( y μ ) T C ( y μ ) ] = trace ( E [ ( y μ ) T C ( y μ ) ] ) = trace ( E [ C ( y μ ) ( y μ ) T ] ) = trace ( C Σ ) .
Physically, it represents cost-weighted variability, implying that variance contributes to loss only when associated with non-zero costs in C . Thus, the risk is interpreted in terms of economic consequences rather than variance alone. The parameter λ adjusts this variance contribution, typically satisfying λ 0 but allowing λ 1 .
Mathematically, introducing λ generalizes the standard expected loss by relaxing the rigid 1:1 bias–variance trade-off, allowing for a flexible control and scalarization of conflicting objectives. Practically, minimizing variance is often more critical than centering the mean, as excessive variability can lead to irreversible consequences. For instance, high variance can cause matching failures in precision assembly or explosion risks in chemical processes, whereas mean shifts are often correctable. Thus, λ allows for heavier penalization of variability, reflecting a risk-averse strategy in robust design.

3.2. Risk Derivation for Mixed Quality Characteristics

To accommodate both nominal-the-best ( τ i 0 ) and smaller-the-better ( τ i = 0 ) types of quality characteristics, we partition the vector y into two sub-vectors, y 1 and y 2 :
y = [ y 1 T y 2 T ] T ,
where y 1 = ( y 1 , , y q ) T is a q × 1 vector of nominal-the-best PCs, and  y 2 = ( y q + 1 , , y p ) T is a ( p q ) × 1 vector of smaller-the-better PCs. For simplicity, one may also scale y to remove unit effects in the loss function.
Following this partition, all related vectors and matrices are restructured accordingly. In particular, the mean vector μ and the target vector τ are expressed as
μ = [ μ 1 T μ 2 T ] T , τ = [ τ 1 T τ 2 T ] T , with τ 2 = 0 .
The covariance matrix Σ and the cost matrix C are also expressed in block forms, respectively:
Σ = Σ 11 Σ 12 Σ 12 T Σ 22 , C = C 11 C 12 C 12 T C 22 .
Using this structure, the multivariate quadratic loss function for y is defined as
L ( y ) = ( y τ ) T C ( y τ ) = y 1 τ 1 y 2 T C 11 C 12 C 12 T C 22 y 1 τ 1 y 2 = ( y 1 τ 1 ) T C 11 ( y 1 τ 1 ) + y 2 T C 22 y 2 + 2 ( y 1 τ 1 ) T C 12 y 2 .
The expected multivariate quadratic loss of y can then be written as
E [ L ( y ) ] = E ( y 1 τ 1 ) T C 11 ( y 1 τ 1 ) + E y 2 T C 22 y 2 + 2 E ( y 1 τ 1 ) T C 12 y 2 .
Expanding the first and second terms using the mean vector μ and variance–covariance matrix Σ , we have
E ( y 1 τ 1 ) T C 11 ( y 1 τ 1 ) = trace ( C 11 Σ 11 ) + trace ( C 11 ( μ 1 τ 1 ) ( μ 1 τ 1 ) T ) , E y 2 T C 22 y 2 = trace ( C 22 Σ 22 ) + trace ( C 22 μ 2 μ 2 T ) ,
and the third term can be rewritten as
E ( y 1 τ 1 ) T C 12 y 2 = trace ( C 12 Σ 12 T ) + trace ( μ 2 ( μ 1 τ 1 ) T C 12 ) .
Combining all three terms, the expected loss becomes
E [ L ( y ) ] = trace ( C 11 Σ 11 ) + trace ( C 11 ( μ 1 τ 1 ) ( μ 1 τ 1 ) T ) + trace ( C 22 Σ 22 ) + trace ( C 22 μ 2 μ 2 T ) + 2 trace ( C 12 Σ 12 T ) + 2 trace μ 2 ( μ 1 τ 1 ) T C 12 = trace C ( μ τ ) ( μ τ ) T + trace C Σ .
This can also be expressed in summation form as
E [ L ( y ) ] = i = 1 p j = 1 p c i j ( μ i τ i ) ( μ j τ j ) + i = 1 p j = 1 p c i j σ i j 2 .
Hence, from Equations (5) and (6), we could express the total risk, T, as 
T = E [ L ( y ) ] + λ trace ( C Σ ) = trace C ( μ τ ) ( μ τ ) T + ( 1 + λ ) trace ( C Σ ) = i = 1 p j = 1 p c i j ( μ i τ i ) ( μ j τ j ) + ( 1 + λ ) i = 1 p j = 1 p c i j σ i j 2 .
Assuming that independent input variables x = ( x 1 , x k ) T influence y , the mean vector and covariance matrix are expressed as functions of x , μ ( x ) and Σ ( x ) . The bias and variance components are then defined as
f 1 ( x ) = trace C ( μ ( x ) τ ) ( μ ( x ) τ ) T ,
f 2 ( x ) = trace C Σ ( x ) ,
so that the total risk can be written as T = f 1 ( x ) + ( 1 + λ ) f 2 ( x ) . Naturally, the total risk T 0 leads to the parameter, the variance weight, λ 1 .
The determination of the weighting parameter, 1 + λ , is possible in two ways depending on the decision-maker’s strategy. Subjectively, users may assign a value to 1 + λ reflecting their perceived importance of bias versus variance. Alternatively, for an objective criterion, 1 + λ can be derived from economic cost structures. If we define the total expected cost as E [ Cos t ] = C bias f 1 ( x ) + C var f 2 ( x ) , where C bias and C var are the aggregate cost coefficients for bias and variance respectively, then normalizing the equation by C bias yields 1 + λ = C var / C bias . For example, C bias may represent process adjustment costs, while C var corresponds to the scrap costs associated with defective products. This formulation allows the trade-off to be driven by tangible economic data when available.

3.3. Multi-Objective Optimization via Global Criterion

Ideally, one would minimize the sum of the functions for a given λ . Practically, however, minimizing both functions simultaneously is challenging. For example, field engineers typically determine a few inputs in x to optimize f 1 , then determine the remaining to optimize f 2 sequentially.
To effectively balance both terms of bias and variance and to globally optimize them, we utilize a Pareto front that represents the comprehensive optimal trade-offs. Unlike standard optimization problems targeting a single objective, multi-objective optimization seeks a set of solutions, called a Pareto optimal set, where no solution is strictly better than the others in all objectives. These solutions form a Pareto front when visualized, allowing us to analyze and adjust λ for an appropriate compromise.
In essence, the Global Criterion (GC) method transforms the multi-objective problem into a scalar minimization problem by measuring the relative deviation of each objective function from its ideal value. We apply an alternative global criterion (GC)-based method for generating a well-distributed Pareto front [9]. The GC function minimizes the distance to a target, defined as i = 1 n | y ^ i T i | / ( U i L i ) p i , where y ^ i is the i-th estimated response, T i is the target (here T i = 0 ), and  U i , L i are the bounds.
Adapting this to our problem with two objective functions ( n = 2 ), f 1 ( x ) and f 2 ( x ) , the GC function is formulated as:
min x | f ^ 1 ( x ) | U 1 L 1 p 1 + | f ^ 2 ( x ) | U 2 L 2 p 2 ,
in which the upper ( U i ) and lower ( L i ) bounds are defined as the maximum and minimum values of f ^ i ( x ) , an approximation of f i ( x ) over the dataset X ,
U i = max x X f ^ i ( x ) , L i = min x X f ^ i ( x ) .
We assume the function is well-behaved enough for its maximum and minimum to be attainable and computable over the dataset X . Alternatively, these bounds can be set by technical constraints or economic conditions. Normalization by ( U i L i ) is crucial for the following reasons [10]:
  • It ensures comparability across responses with different scales.
  • It reduces subjectivity in weighting.
  • It prevents a single criterion from dominating the optimization.
  • It improves the spread of Pareto-optimal solutions.
The exponent parameters p 1 and p 2 assign priorities to the objectives, enabling the exploration of trade-offs. While p i can theoretically range from 0 to infinity, Costa et al. [9] suggest that p i > 3 yields diminishing returns. Our experiments confirm that limiting p i 3 is sufficient. By varying p 1 and p 2 numerically in a grid-search manner, we can explore both convex and non-convex regions of the Pareto frontier [11].
In this research, we perform the optimization over x using pairs of p 1 and p 2 from a predefined grid. The resulting diverse set of solutions x numerically forms the Pareto front, spanning various trade-offs to offer a comprehensive view of the optimization landscape.
To solve Equation (10), we adopt the SLSQP algorithm from Python’s scipy.optimize. minimize library. As a gradient-based method, it efficiently handles equality and inequality constraints, making it well-suited for nonlinear, multi-response problems required to construct an effective Pareto front.

3.4. Procedure of Framework

To summarize the operational procedure of our framework, we propose the following four-step workflow. First, appropriate regression models are fitted for the mean, variance, and covariance of the quality characteristics using data from a designed experiment. Second, the bias loss f 1 ( x ) and variance loss f 2 ( x ) are computed as functions of the design variables via the fitted models. Third, the Pareto front is generated by solving the GC minimization problem across a grid of weighting exponents p 1 and p 2 . Finally, the optimal operating condition is selected from the Pareto front by identifying the point that satisfies a desired risk tolerance or a specific bias-to-variance ratio.

4. Experiments

We now present two synthetic cases to demonstrate the applicability of the proposed approach.

4.1. Case 1: Bivariate Quality Characteristics ( p = 2 )

In the first case, we consider one NB-type PC and one SB-type PC, i.e., q = 1 and p = 2 :
y 1 = ( y 1 ) , y 2 = ( y 2 ) .
Then, total risk T = f 1 ( x ) + ( 1 + λ ) f 2 ( x ) in Equations (7) and (10) is defined by the functions f 1 ( x ) and f 2 ( x ) as follows:
f 1 ( x ) = i = 1 2 j = 1 2 c i j ( μ i ( x ) τ i ) ( μ j ( x ) τ j ) , f 2 ( x ) = i = 1 2 j = 1 2 c i j σ i j 2 ( x ) .
For input variables x = ( x 1 , x 2 ) , 1 x i 1 , i = 1 , 2 , let us assume that the estimated regression functions f ^ 1 ( x ) and f ^ 2 ( x ) , corresponding to f 1 ( x ) and f 2 ( x ) , are given as follows:
f ^ 1 ( x ) = ( 3 x 1 + 3 x 2 + 6 ) 2 , f ^ 2 ( x ) = ( ( x 1 + x 2 ) 2 + 2 ) 2 .
Furthermore, we set U i and L i , for each i = 1 , 2 , to be 0 and 100, respectively. The input variables satisfy 1 x i 1 . We increment each parameter of p 1 and p 2 from 0.05 to 3 with step size 0.1. The initial starting point in the SLSQP algorithm is set to ( 0 , 0 ) .
Using the alternative GC-based method, we numerically obtain the Pareto frontier shown in Figure 1. In this graph, increasing the value of λ , which reflects a higher emphasis on f ^ 2 ( x ) , leads to solutions x with higher bias but lower variance. Conversely, decreasing λ emphasizes f ^ 1 ( x ) , resulting in lower bias but higher variance.
Figure 1. Pareto-front solutions highlighting a solution (depicted by the red circle) that minimizes the sum of bias loss and variance loss, λ = 0 , for the first case.
For instance, as shown in Figure 1, when λ = 0 , the minimum total risk T is 17.54, achieved at x = ( 0.57573 , 0.57573 ) , with corresponding objective values f ^ 1 ( x ) = 6.48 and f ^ 2 ( x ) = 11.06 , which is highlighted by the red circle in Figure 1. Notice that for λ = 0 the tangent line with the ratio, x-intercept/y-intercept, being 1 (thereby slope 1 ) is adopted. If λ differs from zero in general, one can easily draw another tangent line of which the ratio, x-intercept/y-intercept, is 1 + λ (thereby slope 1 λ ) in Figure 1, finding corresponding objective values f ^ 1 ( x ) and f ^ 2 ( x ) along with input vector x .
Here, x denotes the input vector that minimizes the GC-based objective function for a given value of p i , where f ^ 1 ( x ) and f ^ 2 ( x ) represent the estimated bias and variance components, respectively, at the optimal point. In Figure 1, the red line indicates the tangent with slope 1 to the Pareto frontier at the point where the total risk T = f ^ 1 ( x ) + f ^ 2 ( x ) , λ = 0 , is minimized, illustrating the trade-off between the two objectives. This suggests that decision-makers can adjust the input variables by finding a tangent line with a slope relating to their specific industry characteristics or strategic preferences.
As shown above, the construction of a Pareto front enables flexible control over the trade-off between bias and variance terms. Furthermore, let us suppose that one wants to set a ratio of the bias-loss term to the variance-loss term to a certain value in practice. Figure 2 shows three Pareto-Front solutions for specific ratios, 2 : 1 , 1 : 1 , and 1 : 2 of the two terms under approximation error 0.1 .
Figure 2. Pareto-front solutions, depicted by the red circles, for specific ratios of bias loss to variance loss, f ^ 1 ( x ) : f ^ 2 ( x ) being 1:2, 1:1, and 2:1, in the first case.
Firstly, if one sets f 1 ( x ) to be twice f 2 ( x ) with the ratio being 2 : 1 , the optimal solution is x = [ 0.387 , 0.387 ] , with function values f 1 ( x ) = 13.52 and f 2 ( x ) = 6.76 . As seen in Figure 2, the solution is easily extracted by drawing a line with slope 1 / 2 passing through the origin. Secondly, when f 1 ( x ) equals f 2 ( x ) , in which the ratio is 1 : 1 , we easily obtain the optimal solution, x = [ 0.5 , 0.5 ] , which yields function values f 1 ( x ) = f 2 ( x ) = 9.00 . Thirdly, when f 1 ( x ) is half of f 2 ( x ) with a ratio of 1 : 2 , the optimal solution is x = [ 0.595 , 0.595 ] with function values f 1 ( x ) = 5.90 and f 2 ( x ) = 11.67 . These results illustrate how different bias-variance preferences can be effectively represented by specific functional ratios along the Pareto front. We also notice that the extraction of both bias risk f 1 ( x ) and variance risk f 2 ( x ) with a desired ratio is hardly possible without the use of the proposed approach.

4.2. Case 2: Four-Variable Quality Characteristics ( p = 4 )

In the second case, we consider the situation where there are two NB-type PCs and two SB-type PCs, i.e., q = 2 and p = 4 :
y 1 = ( y 1 , y 2 ) , y 2 = ( y 3 , y 4 ) .
The total risk T = f 1 ( x ) + f 2 ( x ) , λ = 0 , in Equations (7) and (10) is defined by functions f 1 ( x ) and f 2 ( x ) as follows:
f 1 ( x ) = i = 1 4 j = 1 4 c i j ( μ i ( x ) τ i ) ( μ j ( x ) τ j ) , f 2 ( x ) = i = 1 4 j = 1 4 c i j σ i j 2 ( x ) .
For input variables x = ( x 1 , x 2 , x 3 ) , let us assume that f ^ 1 ( x ) and f ^ 2 ( x ) are expressed as follows:
f ^ 1 ( x ) = ( 0.5 x 1 2 0.5 x 2 2 + 0.5 x 3 2 + x 1 x 3 4 ) 2 , f ^ 2 ( x ) = ( x 1 2 x 2 2 + 3 x 3 ) 2 .
Similarly to the first case, the upper bounds of f ^ 1 ( x ) and f ^ 2 ( x ) are set to 100, and the lower bounds are set to 0. The initial starting point is ( 1 , 1 , 1 ) , and the input variables are constrained within the range 1 x i 1 . The exponent parameters p i are incremented from 0.05 to 3 with step size 0.05 . Under the setting, we obtain a Pareto front as shown in Figure 3.
Figure 3. Pareto-front solutions highlighting a solution (depicted by the red circle) that minimizes the sum of bias loss and variance loss, λ = 0 , for the second case.
When λ = 0 , we achieve the minimum value of the total risk T at 13.23314, which is highlighted by the red circle (by slope 1 ) in Figure 3, with
f ^ 1 ( x ) = 12.11403 , f ^ 2 ( x ) = 1.11911 ,
and the corresponding input vector is x = ( 1 , 0.00209 , 0.01929 ) .
By allowing approximation error 0.1 , we similarly extract optimal settings under desired trade-off between bias loss and variance loss as shown in Figure 4.
Figure 4. Pareto-front solutions depicted by circles for specific ratios of bias loss to variance loss, f ^ 1 ( x ) : f ^ 2 ( x ) being 1:2, 1:1, and 2:1, in the second case.
Similarly, we extract the solutions from the Pareto front by drawing a line with the appropriate slope passing through the origin. Firstly, when the ratio f 1 ( x ) / f 2 ( x ) is 2, we find the optimal solution, x = [ 1.00 , 1.00 , 0.71 ] , which yields function values f 1 ( x ) = 9.19 , f 2 ( x ) = 4.59 . Secondly, when the ratio f 1 ( x ) / f 2 ( x ) is 1, the optimal solution becomes x = [ 1.00 , 1.00 , 0.90 ] with function values, f 1 ( x ) = 7.28 , f 2 ( x ) = 7.27 . Thirdly, when the ratio f 1 ( x ) / f 2 ( x ) is 1 / 2 , the optimal solution is x = [ 1.00 , 0.83 , 1.00 ] , yielding function values, f 1 ( x ) = 5.49 , f 2 ( x ) = 10.99 .
We notice that the Pareto frontier obtained using the GC-based method contains visual gaps as shown in Figure 3 and Figure 4. Similar discontinuities have been observed in previous studies [9], which demonstrated that achieving a perfectly connected and comprehensive Pareto front is often difficult, especially in the presence of complex or nonconvex response surfaces.

5. Application to Simulation Data

In this section, we apply our approach to a simulation dataset introduced by Pignatiello [1]. This problem serves as a standard test case in multi-response optimization literature, involving two response variables, y 1 and y 2 , and three input variables, x 1 , x 2 , and x 3 .

5.1. Simulation Data and Regression Modeling

The dataset simulates a replicated factorial experiment, enabling the estimation of mean and covariance structures essential for joint optimization. By capturing simultaneous shifts in mean and variance, the Pareto front reveals risk trade-offs that sequential approaches often overlook.
The simulation data is presented in Table 1.
Table 1. Simulation Data.
Using the values in each row in Table 1, we obtain the sample mean, sample variance, and sample covariance as shown in Table 2. For example, the first μ ^ 1 = 103.54875 is the sample average of the four y 1 observations, 104.454 , 105.029 , 99.786 , 104.923 , in Table 1.
Table 2. Statistical Summary of Simulation Data.
Using regression analysis, we derive the following predictive equations:
μ ^ 1 ( x ) = 104.9 3.14 x 1 + 2.38 x 1 x 2 0.35 x 1 x 3 , σ ^ 11 2 ( x ) = 1.613 0.254 x 2 + 1.25 x 3 2 , μ ^ 2 ( x ) = 70.45 0.349 x 1 + 3.59 x 2 0.449 x 1 x 3 + 0.614 x 2 x 3 , σ ^ 22 2 ( x ) = 1.729 + 0.417 x 1 + 1.21 x 3 + 0.263 x 1 x 3 2 , σ ^ 12 2 ( x ) = 3.712 + 1.203 x 1 + 3.555 x 3 + 1.161 x 1 x 3 .
To derive the response surface models, we utilized the replicate data provided in the simulation dataset. For the mean ( μ ^ i ) and variance ( σ ^ i i 2 ) functions, we adopted the models established by Kazemzadeh et al. [12]. However, for the covariance structure ( σ ^ 12 2 ), which is central to our study, we explicitly calculated the sample covariance for each design point using the four replicates and subsequently fitted a second-order regression model to these computed values.
To ensure the reliability of the optimization framework, the derived regression models must meet specific statistical criteria established in quality engineering literature. Following the guidelines suggested by Joglekar and May [13] and Montgomery [14], we adopted a significance level of p < 0.05 and a minimum coefficient of determination R 2 > 0.80 as the thresholds for model adequacy. An R 2 value above 0.80 generally indicates that the regression model explains a substantial proportion of the variability in the response, which is essential for accurate prediction in robust design. If the regression models fail to meet these thresholds, it implies insufficient statistical power or excessive experimental noise. In such cases, the number of experimental replicates should be increased to reduce the standard error.
In this study, we utilized the dataset from Pignatiello [1] ( n = 4 ). To validate whether this sample size was adequate under our proposed criteria, we evaluated the statistical significance of the derived models. As shown in Table 3, all regression models achieved high coefficients of determination ( R 2 > 0.91 ) and were statistically significant ( p < 0.05 ). Specifically, the covariance model ( σ ^ 12 2 ), a critical component of our framework, demonstrated an R 2 of 0.9726 with a p-value of 0.001. These results confirm that the sample size of n = 4 provided sufficient statistical power to capture the underlying covariance structure accurately.
Table 3. Statistical verification of the regression models ( n = 4 ).

5.2. Formulation of Objective Functions

In this dataset, the target vector is τ = ( 103 , 73 ) . Consistent with the definition in Section 3.1 where cost coefficients are determined by the economic impact of deviations, we adopt the specific values established in the study by Pignatiello [1]:
C = 1.0 0.5 0.5 2.5 .
Based on these target values and cost values, the following functions are defined:
f ^ 1 ( x ) = i = 1 2 j = 1 2 c i j μ ^ i ( x ) τ i μ ^ j ( x ) τ j , f ^ 2 ( x ) = i = 1 2 j = 1 2 c i j σ ^ i j 2 ( x ) .
By evaluating these functions across the experimental design space, we obtain the following statistical ranges:
U 1 = 192.28 , L 1 = 3.643 , U 2 = 63.87 , L 2 = 0.306 .
To generate a Pareto frontier, we employ an alternative GC-based method with the following objective function:
min x | f ^ 1 ( x ) | 192.28 3.643 p 1 + | f ^ 2 ( x ) | 63.87 0.306 p 2 .

5.3. Pareto Frontier Analysis

Figure 5 shows a Pareto frontier generated by constraining the input variables to the range 1 x i 1 , i = 1 , 2 , 3 , and varying the parameters p 1 and p 2 from 0.05 to 3 in increments of 0.05. Optimization is initialized from starting point ( 0 , 0 , 0 ) .
Figure 5. Pareto-front solutions for various variance-risk weights, λ = 0 , 4 , 6 , (depicted by the three circles coming from the tangent lines) in the simulation experiment.
When λ = 0 , we find the minimum total risk as follows:
T = 1.95304 , f ^ 1 ( x ) = 0.60067 , f ^ 2 ( x ) = 1.35237 , x = ( 1 , 0.71001 , 1 ) .
For λ = 4 , similarly using the front, we extract minimum total risk as follows:
T = 7.02391 , f ^ 1 ( x ) = 1.52971 , f ^ 2 ( x ) = 1.09884 , x = ( 0.52104 , 0.70273 , 1 ) .
Finally, when λ = 6 , minimum total risk becomes
T = 9.05211 , f ^ 1 ( x ) = 2.52681 , f ^ 2 ( x ) = 0.93219 , x = ( 0.17905 , 0.72327 , 1 ) .
The three dashed lines in green, yellow, and red shown in Figure 5 represent the tangents to the Pareto frontier at each solution that minimizes the weighted sum T = f ^ 1 ( x ) + ( 1 + λ ) f ^ 2 ( x ) , for λ = 0 , 4 , 6 , respectively.

5.4. Comparative Analysis with Sequential Strategies

A common practical approach to quality optimization is to adopt a sequential decision process. For instance, a field engineer may first minimize the bias term f ^ 1 ( x ) , and then reduce the variance term f ^ 2 ( x ) among the solutions where bias is already minimized.
To evaluate this practical strategy, we compare it against our proposed approach under the setting λ = 0 , where bias and variance are treated as equally important. The optimal solution minimizing the total risk, T = f ^ 1 ( x ) + f ^ 2 ( x ) , is indicated by the green dot in Figure 6, yielding the minimum value T = 1.95 .
Figure 6. Comparison of decision strategies between sequential decisions depicted by two stars and Pareto-based decisions depicted by the green dot for the total risk, λ = 0 , and by the three red circles for various trade-off ratios of bias loss to variance loss, f ^ 1 ( x ) : f ^ 2 ( x ) being 1:2, 1:1, and 2:1.
In contrast, when one adopts a sequential decision-making strategy, the outcomes are significantly suboptimal compared to the joint optimization approach. For instance, if one first minimizes the bias term f ^ 1 ( x ) and then evaluates the variance term at that point, the results are f ^ 1 ( x ) = 0.30 and f ^ 2 ( x ) = 14.72 , yielding a total risk of T = f ^ 1 ( x ) + f ^ 2 ( x ) = 15.02 for λ = 0 . This point is highlighted by the red star, located at the upper-left side in Figure 6. On the other hand, if one first minimizes the variance term f ^ 2 ( x ) and then evaluates the bias at that point, the corresponding values are f ^ 1 ( x ) = 6.25 and f ^ 2 ( x ) = 0.45 , resulting in a total risk of T = 6.70 . This point is highlighted by the yellow star, located at the lower-right side in Figure 6. Notice that the total risk, T = 1.95 , represented by the green dot in Figure 6, is far smaller than the values of T = 15.02 and T = 6.70 resulting from the two sequential decisions.
This comparison highlights that such a sequential optimization approach—whether prioritizing bias or variance—can lead to significantly worse solutions than those obtained by balancing both objectives jointly.
Furthermore, our approach enables flexible decision-making by allowing users to control the trade-offs between bias and variance via ratio settings. For example, we consider three target ratios: 1, 2, and 0.5. The corresponding solutions and function values are highlighted by the red circles in Figure 6 for direct comparison.
When the ratio f ^ 1 ( x ) / f ^ 2 ( x ) is approximately 1, the optimal solution is x = [ 0.673 , 0.702 , 1 ] , which yields f ^ 1 = f ^ 2 = 1.18 . When the ratio is approximately 2, the solution becomes x = [ 0.346 , 0.709 , 1 ] , resulting in f ^ 1 = 2.01 and f ^ 2 = 1.01 . Lastly, for a ratio of 0.5, the solution is x = [ 0.956 , 0.710 , 1 ] , with f ^ 1 = 0.66 and f ^ 2 = 1.33 .
Consequently, our method simultaneously incorporates both bias and variance through a weighted formulation, enabling a more balanced and efficient optimization process. This approach identifies solutions with lower total risk, clearly demonstrating its effectiveness in achieving a superior trade-off between bias and variance. Additionally, our method provides the flexibility to choose mixing ratios between bias and variance terms, allowing for optimization that can be customized according to specific needs.

6. Conclusions

In this study, we proposed a novel approach for optimizing interrelated quality attributes by incorporating variance into the expected loss function. Leveraging an alternative GC-based method, we successfully constructed Pareto front to visualize the trade-offs between conflicting objectives. The effectiveness and practical utility of the proposed method were demonstrated through both simplified examples and a simulation study.
Nonetheless, constructing a comprehensive Pareto frontier requires the exploration of multiple starting points, and identifying all possible nondominated solutions remains a significant challenge. Future research should focus on developing systematic strategies for selecting initial points and benchmarking the proposed approach against existing multi-objective optimization techniques. Furthermore, as the complexity of the expected loss and variance functions increases, the associated computational burden also grows. Investigating simplified formulations or alternative structures to enhance computational efficiency would be a valuable direction for future work.
In addition, it is important to explore how this method can be extended to more realistic scenarios where only a subset of input variables—such as two or three out of many—can be actively controlled. Adapting our approach to such constrained optimization settings would further broaden its applicability.
Overall, this study contributes to the field of quality optimization by offering a new perspective on managing trade-offs in complex systems. The proposed method has the potential for application across various industries, and future extensions could include higher-dimensional problems and the integration of machine learning techniques for dynamic and adaptive optimization.

Author Contributions

Conceptualization, S.K. and K.L.; methodology, S.K.; software, S.K.; validation, S.K. and K.L.; writing—original draft preparation, S.K.; writing—review and editing, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of the Republic of Korea (RS-2025-23524280).

Data Availability Statement

The data presented in this study are available on GitHub at https://github.com/KIM-sang-won/PFOVELIQ accessed on 4 February 2026.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pignatiello, J.J. Strategies for robust multiresponse quality engineering. IIE Trans. 1993, 25, 5–15. [Google Scholar] [CrossRef]
  2. Shiau, G.H. A study of the sintering properties of iron ores using the Taguchi’s parameter design. J. Chin. Stat. Assoc. 1990, 28, 253–275. [Google Scholar]
  3. Tong, L.I.; Su, C.T.; Wang, C.H. The optimization of multi-response problems in the Taguchi method. Int. J. Qual. Reliab. Manag. 1997, 14, 367–380. [Google Scholar] [CrossRef]
  4. Soh, W.; Kim, H.; Yum, B.-J. A multivariate loss function approach to robust design of systems with multiple performance characteristics. Qual. Reliab. Eng. Int. 2016, 32, 2685–2700. [Google Scholar] [CrossRef]
  5. Costa, N.; Lourenço, J. Bi-Objective Optimization Problems—A Game Theory Perspective to Improve Process and Product. Sustainability 2022, 14, 14910. [Google Scholar] [CrossRef]
  6. Jones, D.R.; Schonlau, M.; Welch, W.J. Efficient global optimization of expensive black-box functions. J. Glob. Optim. 1998, 13, 455–492. [Google Scholar] [CrossRef]
  7. Knowles, J. ParEGO: A hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems. In Proceedings of the 2005 Conference on Genetic and Evolutionary Computation, Washington, DC, USA, 25–29 June 2005; pp. 775–782. [Google Scholar]
  8. Zhang, Q.; Li, H. A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  9. Costa, N.R.; Lourenço, J.; Pereira, Z.L. Multiresponse Optimization and Pareto Frontiers. Qual. Reliab. Eng. Int. 2012, 28, 701–712. [Google Scholar] [CrossRef]
  10. Marler, R.T.; Arora, J.S. Function-transformation methods for multi-objective optimization. Eng. Optim. 2005, 37, 551–570. [Google Scholar] [CrossRef]
  11. Messac, A.; Sundararaj, G.; Tappeta, R.; Renaud, J. Ability of objective functions to generate points on non-convex Pareto frontiers. AIAA J. 2000, 38, 1084–1091. [Google Scholar] [CrossRef]
  12. Kazemzadeh, R.B.; Bashiri, M.; Atkinson, A.C.; Noorossana, R. A general framework for multiresponse optimization problems based on goal programming. Eur. J. Oper. Res. 2008, 189, 421–429. [Google Scholar] [CrossRef]
  13. Joglekar, A.M.; May, A.T. Product excellence through design of experiments. Cereal Foods World 1987, 32, 857–868. [Google Scholar]
  14. Montgomery, D.C. Design and Analysis of Experiments, 9th ed.; John Wiley & Sons: Hoboken, NJ, USA, 2017. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.