Next Article in Journal
An Explicit Shifted Legendre Petrov–Galerkin Technique for the Time Fractional Cable Problem
Next Article in Special Issue
Two-Stage Analysis for Supply Chain Disruptions Considering the Trade-Off Between Profit Maximization and Adaptability
Previous Article in Journal
Robustness of Identifying Item–Trait Relationships Under Non-Normality in MIRT Models
Previous Article in Special Issue
Parallel Simulation Multi-Sample Task Scheduling Approach Based on Deep Reinforcement Learning in Cloud Computing Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gaussian Learning-Based Pareto Evolutionary Algorithm for Parallel Machine Planning in Industrial Silicon Production

1
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
2
School of Logistics and Management Engineering, Yunnan University of Finance and Economics, Kunming 650221, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(23), 3860; https://doi.org/10.3390/math13233860
Submission received: 21 October 2025 / Revised: 19 November 2025 / Accepted: 27 November 2025 / Published: 2 December 2025

Abstract

This study focuses on a multi-objective heterogeneous parallel machine planning problem for industrial silicon smelting. Specifically, under the conflicting objectives of minimizing carbon emissions, rollover penalty costs, and load imbalance, the total production demand of industrial silicon is allocated monthly across multiple machines. We first establish the mathematical model of the problem accounting for real-life management requirements. To solve the model, a Gaussian learning-based Pareto evolutionary algorithm (GLPEA) is proposed. The algorithm is developed based on a nondominated sorting framework and incorporates two key innovations: (1) a generation-wise dynamic Gaussian mixture component selection strategy that adaptively fits the multimodal distribution of elite solutions, and (2) a hybrid offspring generation mechanism that integrates traditional evolutionary operators with a Gaussian sampling strategy trained on perturbed solution sets, thereby enhancing exploration capability while maintaining convergence. The effectiveness of GLPEA is validated on 40 problem instances of varying scales. Compared with NSGA-II and MOEA/D, GLPEA achieves average improvements of 5.78% and 89.23% in IGD, and 1.03% and 264.43% in HV, respectively. We make the source codes of GLPEA publicly available to facilitate future research on practical applications.

1. Introduction

Industrial silicon plays an irreplaceable role in strategic emerging industries, such as photovoltaics, semiconductors, and aluminum [1]. Over the past two years, the authors conducted a practical project for a cooperative industrial silicon factory in Yunnan, China. Our task was to make production planning for heterogeneous parallel furnaces. As illustrated in Figure 1, the smelting process involves raw materials including silica, cleaned coal, charcoal, semi-coke, and petroleum coke, yielding molten industrial silicon while substantial CO 2 . Figure 2 provides an overview of the production planning framework in the factory. The factory has a fixed monthly delivery demand, and the planning task consists of allocating the output of each furnace to meet this demand. Each furnace follows a maintenance planning, and its monthly output must not exceed its rated capacity minus downtime. Furnace load is quantified as the total assigned production during the planning period, and the combined output of all furnaces must equal the total demand. Unfinished tasks are rolled over to subsequent months, incurring a penalty cost, while excess production is treated as early output for the following month. The carbon emission factor, which varies by furnace and month, reflects the CO 2 emissions per ton of silicon produced. The rollover penalty factor is the cost per ton of delayed delivery. Reducing rollover penalties requires assigning more tasks to higher-capacity furnaces. This may worsen load imbalance and increase emissions. Conversely, minimizing emissions or balancing loads could raise penalties. Therefore, the three objective functions in nature conflict with each other.
Given the above real-life applications, we propose introducing a specific heterogeneous parallel machine planning problem (HPMPP). However, no single algorithm can optimize the three conflicting objective functions simultaneously. Therefore, multi-objective evolutionary algorithm (MOEA) can be a suitable solution method. MOEAs use a parallel search framework to find approximate Pareto front (APF) within an acceptable CPU time frame [2].
However, as will be analyzed in detail in Section 2, classical multi-objective evolutionary algorithms (MOEAs) such as NSGA-II, although widely applied, rely heavily on “blind” stochastic search operators. Often, the nondominated sorting model is short for finding promising solutions in certain search regions when there are more than two objective functions. It may also suffer from reduced convergence precision [3,4,5]. To overcome these limitations, machine learning has shown potential to improve MOEAs. Gaussian learning has attracted increasing attention due to its strong capability in distribution fitting. A single-component GMM only offers a unimodal approximation of the primary elite cluster, whereas a multi-component GMM can better represent multimodal distributions of multiple high-quality solution clusters [6]. Therefore, employing different GMM components at different search stages of MOEAs can help yield improved results. In addition, the performance of GMM highly depends on the quality of its training data. If the GMM is trained solely on the current Pareto elite solutions, it tends to perform “local exploitation” around the discovered elites. This leads to a rapid loss of global exploration ability, causing the population to prematurely converge to a local Pareto front. This paper proposes GLPEA for solving HPMPP in industrial silicon smelting. GLPEA employs a Gaussian distribution to discover high-quality offspring solutions and a dedicated GMM training strategy to enhance global exploration. We additionally design a problem-specific repair operator to handle infeasible solutions encountered during the search process. The contributions of this work are summarized as follows.
First, from the modeling perspective, HPMPP aims to optimize rollover penalty costs, carbon emissions, and load imbalance. To our knowledge, this is the first work to model the industrial silicon production planning problem within such a framework. This work provides valuable guidance for managers who desire to balance these objectives. We use 40 problem instances of varying sizes and make the source code of the algorithm publicly available. These resources would be useful to facilitate future research on algorithm benchmarking and applications.
Second, from the algorithm perspective, we introduce a generational dynamic GMM component selection strategy into the nondominated sorting framework. This differs from most studies that use fixed-component GMMs. Our strategy adapts the number of components at each generation, leading to improved model flexibility and sampling effectiveness across different search stages. The perturbed solution set GMM training strategy further enhances exploration during the search. The proposed GLPEA provides useful insights for researchers and practitioners working on complex heterogeneous parallel machine planning problems.
Third, our GMM strategy is of general interest. It can guide the design of new operators within any nondominated sorting framework. Integrating this strategy into other MOEAs could improve their convergence and exploration. Therefore, our findings offer a valuable reference for advancing MOEAs.
The rest of this paper is organized as follows. Section 2 provides a review of the related works. Section 3 presents the mathematical model of HPMPP. Section 4 details the proposed GLPEA. Section 5 reports experimental results and comparisons. In Section 6, we draw concluding comments.

2. Related Work

2.1. Classical MOEAs

With the growing consensus on sustainable manufacturing, scheduling objectives have substantially evolved. Optimization now extends beyond the traditional two-dimensional “efficiency–cost” trade-off to higher-dimensional objectives encompassing “efficiency–cost–environment.” MOEAs have become mainstream tools for addressing such complexity. Generally, MOEAs can be categorized into: (a) dominance-based MOEAs, such as NSGA-II [7] and SPEA-II [8], which rank and select solutions based on Pareto dominance levels and crowding distance (or similar diversity indicators); (b) decomposition-based MOEAs, such as MOEA/D [9] and MOGLS [10], which decompose a multi-objective problem into a set of single-objective subproblems and exploit neighborhood relationships among subproblems for cooperative optimization, and (c) indicator-based MOEAs, such as SMS-EMO [11,12], which directly employ performance indicators (e.g., hypervolume) as selection criteria to drive the population toward a better Pareto front. In particular, the NSGA-II model has shown powerful search performance for solving the parallel machine planning problem [13,14]. Its search process relies on nondominated sorting and crowding distance mechanisms to discover promising search regions. Motivated by these studies, we adopt the nondominated sorting search model of NSGA-II in this work.
Despite their remarkable success, classical MOEAs still face inherent limitations. Their evolutionary operators are often inherently “blind,” relying purely on stochastic perturbations and heuristic selection. This traditional approach lacks the capability to perceive the topological structure of the search space or discern the distribution characteristics of high-quality solutions [15,16]. Such “blind” exploration often leads to limited search efficiency and consequently slow convergence. To overcome these limitations, researchers have introduced machine learning techniques to enhance the performance of MOEAs.

2.2. Surrogate-Assisted MOEAs

The core idea of Surrogate-Assisted MOEAs (SA-MOEAs) is to employ a computationally inexpensive machine learning model (surrogate) to approximate the true objective function, primarily targeting scenarios where evaluations are computationally expensive. Studies by Balekelayi N [17], Rossmann J [18], and Hebbal A [19] have successfully demonstrated that optimization based on Gaussian processes (GP) can achieve faster convergence to the optimal Pareto front than NSGA-II. Wu H [20] and Ravi K [21] further explored sparse and multi-fidelity GPs to improve efficiency on medium-scale problems. Gaussian processes are particularly valuable as they not only provide predictive means but also quantify predictive uncertainty, making them ideal models for Bayesian Optimization (BO) [22], which intelligently balances exploration and exploitation through acquisition functions. Other Surrogate-Assisted MOEAs have also been widely used. For instance, research by Zhang H [23] demonstrated that the Inverse Gaussian Process (IGP)-based MOEA significantly improves performance in dynamic optimization. Research by Yang Z [24] and Niu Y [25] introduced surrogate models based on Radial Basis Functions (RBF), effectively enhancing prediction accuracy. Studies by Sonoda T [26] and Xu D [27] developed surrogate-assisted MOEAs using Support Vector Machines (SVMs), achieving high computational efficiency. Moreover, works by Zhu E [28] and De Moraes M. B [29] demonstrated that surrogate-assisted MOEAs using Random Forests (RF) exhibit outstanding optimization performance.
However, the primary challenge for the HPMPP considered in this study differs significantly from the expensive-evaluation scenario. In this industrial context, production plans are typically established at the beginning of each planning cycle and executed strictly thereafter. The objective function evaluation frequency is thus relatively low. Consequently, the main challenge lies not in computationally expensive evaluations but in the inherently low search efficiency, reflected by slow convergence and limited solution quality improvement. Under such circumstances, the advantages of SA-MOEAs become less pronounced, as their designs are tailored for expensive-evaluation problems and may introduce additional computational overhead, making them less suitable for improving search efficiency in our HPMPP context.

2.3. Estimation of Distribution Algorithms (EDAs)

A distinct paradigm for enhancing MOEAs is the Estimation of Distribution Algorithms (EDAs), which replace traditional genetic operators (crossover and mutation) with a statistical learning or probabilistic generative model. The core principle involves learning the probability distribution of high-quality solutions in the decision space and subsequently sampling new offspring based on this learned model to generate potentially superior solutions [30].
Empirical studies by Wang F [31], Li G [32], and Zou J [33] have demonstrated that GMM-based MOEAs exhibit excellent performance across a wide range of benchmark and real-world problems. Lu C [34] and Aggarwal S [35] proposed MOEAs based on the Covariance Matrix Adaptation Evolution Strategy (CMA-ES), which feature lightweight design and faster execution. Zhang W [36] improved the Regularity Model-based Estimation of Distribution Algorithms (RM-MEDA), enhancing convergence performance. Kalita K [37] introduced the Multi-objective exponential distribution optimizer (MOEDO), which achieves robust performance in balancing diversity and convergence efficiency.

2.4. Analysis and Proposed Approach

For the HPMPP addressed in this study, we aim to solve planning problems that are computationally challenging yet required to be completed within an acceptable CPU time frame. Among existing methods, GMM-based MOEAs are well suited for this purpose. Compared with other probabilistic models, the GMM has been widely validated as effective—particularly in industrial applications—by researchers such as Alghamdi A. S [38], Guerrero-Peña E [39] and Yazdi F [40]. Moreover, GMMs are data-efficient, capable of robustly fitting multimodal distributions even with small to medium sample sizes. Considering practical industrial applicability and algorithmic stability, we adopt the GMM as the probabilistic generative model in this work. In recent years, many studies have sought to enhance the performance of GMM-based sampling strategies. Zhang J [41] decomposed the overall problem into several subproblems and employed multi-component GMMs to capture diverse solution patterns, thereby improving population diversity. Abdulghani A. M [42] proposed the NA-GMM, which integrates adaptive weighting of features and node importance weighting to further enhance diversity.
Most existing studies employ a fixed number of Gaussian components throughout the evolutionary process, which is theoretically suboptimal. During the early evolution stage, the population distribution is broad and dispersed; a simple GMM can adequately capture the global trend, while using an excessive number of components may lead to overfitting. In later stages, when the population converges to a refined and complex Pareto front with multiple distinct elite clusters, a low-component GMM fails to distinguish these clusters, resulting in underfitting. Therefore, we introduce a generation-wise dynamic Gaussian mixture component selection strategy. Furthermore, the training dataset of the GMM has a significant impact on its sampling performance. A training set with greater diversity enables more diverse sampling results. To this end, we propose a GMM training strategy based on perturbed solution sets to enhance the diversity of the solution space and improve global exploration capability.

3. Mathematical Model of HPMPP

HPMPP involves allocating predetermined monthly production targets to a set of heterogeneous parallel machines (furnaces). Given the monthly rollover penalty factors, machine maintenance schedules, capacity limits, and deep learning-predicted carbon emission factors, the objective is to determine the monthly capacity allocation across machines that optimizes the three objective functions detailed below.

3.1. Assumptions

The proposed HPMPP model is based on the following assumptions.
(a) The maximum monthly production capacity of each machine is fixed and time-invariant.
(b) No unexpected interruptions occur during the production process, allowing the analysis to focus on the inherent trade-offs among the optimization objectives.

3.2. Parameters

To clearly construct the model and algorithm, this section classifies all relevant parameters into two categories: Table 1 defines the model parameters of the industrial silicon scheduling problem, while Table 2 defines the algorithm parameters of the proposed GLPEA.

3.3. Decision Variables

X i j Production volume of month i I on machine j M ;
V i Cumulative amount of unfulfilled demand in month i;
U i Unfulfilled demand in month i;
L j Load of machine j.

3.4. Objective Functions

The objective of HPMPP is to minimize carbon emissions (f1), minimize rollover penalty costs (f2), and minimize load imbalance (f3). The carbon emissions (f1) are obtained by multiplying the production of each machine in each month by the corresponding unit carbon emission coefficient and summing over all machines and months. The monthly rollover penalty (f2) is calculated based on unfinished demand: for each month i, the actual total production is compared with the demand for that month, including any rollover from the previous month. If production is insufficient, an unfinished amount U i is generated; if production exceeds demand, the surplus offsets the following month’s demand. The unfinished amount for each month is multiplied by the monthly rollover penalty coefficient D i , and the sum over all months gives the total penalty. Load imbalance (f3) is calculated by first computing the total production load L j of each machine over the entire scheduling period. The squared deviations of each machine’s total load from the average load are then summed to represent the load imbalance.
Min f 1 = i I j M C i j X i j
Min f 2 = i I D i U i
Min f 3 = j M L j 1 M j M L j 2
Subject to:
X i j Cap j 1 R i j H i , i I , j M
i I j M X i j = i I P i
L j = i I X i j , j M
V i = P i + V i 1 j M X i j , i I , V 0 = 0
U i = max 0 , V i , i I , V 0 = 0

3.5. Carbon Emission Factor Prediction

In the proposed mathematical model, the carbon emission factor denotes the carbon emissions per ton of silicon product. Although emissions generally scale with output, the factor is also influenced by operational efficiency, energy consumption, and climatic conditions (e.g., temperature, humidity, rainfall), leading to monthly variations. Therefore, it can vary from month to month. Accurate prediction of this factor is important for emission management. Hochreiter et al. [43] showed that long short-term memory (LSTM) networks perform well on nonlinear time-series prediction.

3.5.1. Datasets and Preprocessing

We collected five years (2020–2024) of production data from an industrial silicon smelting plant (The data set is available at https://gitee.com/zhangjinsi_p/glpea/blob/master/dataset.xlsx, accessed on 30 September 2024). It covers seven furnaces (#1 to #7), with monthly records for each furnace. In total, there are 60 months, comprising 420 entries of monthly production, carbon dioxide emissions, and carbon emission factors. The monthly production is measured in tons and represents the output of each furnace for the corresponding month. The carbon dioxide emissions are also measured in tons, indicating the CO 2 output of each furnace per month. The carbon emission factor represents the ratio of CO 2 emissions to production for each furnace in the corresponding month, i.e., the amount of CO 2 emitted per ton of industrial silicon produced.
As the dataset is relatively small but of high quality, we performed a thorough inspection and found no missing values or obvious outliers. For data preprocessing, we applied the RobustScaler method for standardization. Bachechi et al. [44] showed that this method reduces the impact of outliers and noise.
The dataset was split into training and test sets at an 80:20 ratio. It is worth noting that when the dataset is limited in size, setting aside a separate validation set can significantly reduce the valuable data available for model training. Therefore, we adopted cross validation for time series [45]. Specifically, we set a validation window corresponding to a complete seasonal cycle (window size = 12) and rolled it forward within the training set, ensuring that each validation set only contains data that the model has not seen during training and occurs later in chronological order.

3.5.2. Feature Engineering

Inspired by Box [46], Ma [47], and Dong [48], we construct multidimensional features to capture periodicity, lag effects, and local statistical properties in time series.
(a) Periodic features. We map the month index to a periodic variable. We encode the 12-month cycle using sine and cosine functions as follows:
s i n m o n t h = sin 2 π Φ t P , c o s m o n t h = cos 2 π Φ t P
where Φ t = ( t mod 12 ) + 1 represents the month index in the cycle. The cycle length P = 12 months. This encoding captures seasonal variation and the periodic effects of temperature, humidity, and rainfall.
(b) Lag features and padding. To reflect the influence of past values, we use the observations from the three most recent time steps as lag features. We denote them x t 1 , x t 2 , and x t 3 .
(c) Sliding-window statistical features. To capture local trends and volatility, we compute statistics over a fixed window of size w = 12 (including the current time t). We have
μ t = 1 w i = t w + 1 t x i , σ t = 1 w i = t w + 1 t ( x i μ t ) 2
We additionally calculate the ratio of the standard deviation to the mean to capture local volatility as follows:
γ t = σ t μ t + ϵ
Toward that end, at each time step t, we construct a feature vector X t as follows:
X t = x t , sin 2 π Φ t P , cos 2 π Φ t P , x t 1 , x t 2 , x t 3 , μ t , σ t , γ t

3.5.3. Model Structure and Loss Function

We employed a single-layer LSTM encoder, followed by a fully connected (FC) layer, to predict the estimated carbon emission factor for the next time step, Y ^ t + 1 , which approximates the true value Y t + 1 . The Mean Squared Error (MSE) [49] was utilized as the loss function:
M S E = 1 N i = 1 N ( Y i Y ^ i ) 2
where N represents the number of samples, Y i denotes the true (or actual) value, and Y ^ i is the predicted value from the model.

3.5.4. Model Parameters

The number of LSTM hidden units (lstm_unit) and the Dropout rate (dropout_rate) significantly influence the model’s generalization ability and fitting performance. We conducted an experimental grid search involving nine different combinations of these parameters. The results are presented in Table 3. As shown in the table, the combination of lstm_unit = 48 and dropout_rate = 0.3 yielded the lowest MSE score. Consequently, this optimal hyperparameter set was selected for the training of the final model.

3.5.5. Evaluation and Prediction Results

The model’s performance was evaluated using the Mean Squared Error (MSE), Root Mean Squared Error (RMSE) [50], and Mean Absolute Error (MAE) [51]. The resulting metrics were M S E = 0.0067 , R M S E = 0.0794 , and M A E = 0.0643 . Overall, the model achieved low prediction errors, demonstrating its excellent predictive performance.
The trained model is used to forecast monthly carbon emission factors for each furnace over a 12-month horizon. The results are reported in Table 4.

4. Proposed GLPEA for HPMPP

In GLPEA, we encode a solution with a two-dimensional matrix X. The matrix has I rows and M columns. Each element X i j Cap j 1 R i j H i ,   i I ,   j M . represents the production assigned to machine j in month i. Based on this solution representation method, we detail the search components of GLPEA below.

4.1. Proposed Problem-Specific Repair Operator

To ensure all solutions satisfy the problem constraints during the search process of GLPEA, a dedicated repair operator is proposed in Algorithm 1. The repair operator first constructs a capacity upper-bound matrix to represent the available capacity of machine j in month i accounting for maintenance (lines 3–8). Next, the infeasible solution matrix and the capacity matrix are vectorized (lines 9–10). Thereafter, a bisection search determines the shift parameter θ such that the projection x k = min { max ( z k θ , 0 ) , u k } results a solution satisfying the total production demand with a tolerance 1 × 10 8 (lines 13–23), thereby enforcing both bound constraints and demand satisfaction. The repaired vector is then reshaped into the final feasible matrix (Line 24). Note that in determining θ , we set N b i s = 1000 . Based on our preliminary observations, such an N b i s is sufficient to find a suitable θ value.
Algorithm 1 Proposed problem-specific repair operator
  1:
Input: X
  2:
Output: X r e p a i r e d
  3:
cap_matrix Matrix ( zeros , I , M ) ;
  4:
for i = 1 to I do
  5:
     for  j = 1 to M do
  6:
          cap_matrix[i,j] Cap ( 1 R [ i ] / H [ i ] ) ;
  7:
     end for
  8:
end for
  9:
z flatten ( X ) ;
10:
u flatten (cap_matrix);
11:
θ l o min ( z u ) ;
12:
θ h i max ( z u ) ;
13:
for  i = 1 to N b i s  do
14:
      θ ( θ l o + θ h i ) / 2 ;
15:
     if  ( min ( max ( z θ , 0 ) , u ) ) i I P i 1 e 8  then
16:
           b r e a k ;
17:
     end if
18:
     if  ( min ( max ( z θ , 0 ) , u ) ) > i I P i  then
19:
           θ l o θ ;
20:
     else
21:
           θ h i θ ;
22:
     end if
23:
end for
24:
X r e p a i r e d reshape ( min ( max ( z θ , 0 ) , u ) , I , M ) ;
25:
return X r e p a i r e d

4.2. Gaussian Mixture Model

4.2.1. Univariate Gaussian Distribution

The univariate Gaussian distribution (normal distribution) is one of the most common continuous probability distributions. Its probability density function is defined as:
p ( x | μ , σ 2 ) = 1 2 π σ 2 exp ( x μ ) 2 2 σ 2
where x R is the random variable, μ is the mean, and σ 2 is the variance. Given a set of independent and identically distributed samples { x 1 , x 2 , , x N } , the parameters μ and σ 2 are calculated as follows:
μ = 1 N i = 1 N x i , σ 2 = 1 N i = 1 N ( x i μ ) 2

4.2.2. Multivariate Gaussian Distribution

The multivariate Gaussian distribution generalizes the univariate case. It describes the joint distribution of a D-dimensional continuous random vector x. Its probability density function is:
p ( x | μ , Σ ) = 1 ( 2 π ) D / 2 | Σ | 1 / 2 exp 1 2 ( x μ ) T Σ 1 ( x μ )
where μ R is the mean vector. Σ R D × D is the covariance matrix, which describes the correlations among the dimensions. | Σ | denotes the determinant of the covariance matrix.

4.2.3. Adopted Gaussian Mixture Model

GMM assumes that each data point x is generated from a weighted combination of K Gaussian distributions [52]. Its probability density function is expressed as follows:
p ( x ) = k = 1 K π k N ( x | μ k , Σ k )
where π k denotes the mixing weight of the k-th Gaussian component, satisfying k = 1 K π k = 1 . μ k and Σ k represent the mean vector and covariance matrix of the k-th Gaussian distribution, respectively. The parameters of GMM are usually estimated using the Expectation-Maximization (EM) algorithm [53]. The EM algorithm consists of two steps:
Expectation step (E-step): In this step, the posterior probability that each data point belongs to each Gaussian component, namely the responsibility, is calculated. γ i k represents the probability that the i-th data point is generated by the k-th Gaussian component, and t denotes the current iteration. The formula is given as follows:
γ i k ( t ) = π k ( t ) N x i | μ k ( t ) , Σ k ( t ) j = 1 K π j ( t ) N x i | μ j ( t ) , Σ j ( t )
Maximization step (M-step): In this step, the parameters are updated using the posterior probabilities obtained from the E-step. The new parameters are estimated by maximizing the log-likelihood function. The mixing weights are updated as follows:
π k ( t + 1 ) = 1 N i = 1 N γ i k ( t )
where i = 1 N γ i k ( t ) represents the total responsibility of all data points for the k-th Gaussian component. Note that N is the total number of data points.
Thereby, the mean vector and covariance matrix are updated as follows:
μ k ( t + 1 ) = i = 1 N γ i k ( t ) x i i = 1 N γ i k ( t )
Σ k ( t + 1 ) = i = 1 N γ i k ( t ) ( x i μ k ( t + 1 ) ) ( x i μ k ( t + 1 ) ) T i = 1 N γ i k ( t )
Toward that end, the EM algorithm can alternates between the E-step and the M-step. It increases the log-likelihood at each iteration until it converges to a local optimum.

4.3. Main Framework of GLPEA

As mentioned earlier, GLPEA employs a Gaussian mixture model-based non-dominated sorting framework. The general framework of the algorithm is given in Algorithm 2. Initially, a population of size P S is randomly generated. The objective function calculation module evaluates three objective values F = { f 1 , f 2 , f 3 } for each individual. At each generation, the population first undergoes non-dominated sorting. Subsequently, a perturbed solution set is constructed for Gaussian model training (Section 4.4). Offspring are then generated using a Gaussian mixture model-based evolutionary method (Section 4.5). The parent and offspring populations are merged to form a temporary population, which then undergoes non-dominated sorting and crowding distance calculation. Finally, an elitism preservation strategy selects the top P S optimal individuals to form the new-generation population. This process continues until the maximum generation number G E N S is reached, ultimately outputting an approximate Pareto-optimal solution set.
Algorithm 2 Main framework of GLPEA
  1:
Input: initial_population, GENS, and the other parameters of the algorithm
  2:
Output: final_population
  3:
evaluate objectives F = { f 1 , f 2 , f 3 } for all individuals;
  4:
for  g e n = 1 to G E N S  do
  5:
     perform non-dominated sorting;
  6:
     hybrid data augmented gaussian model training strategy; /∗ Section 4.4 ∗/
  7:
     generate offspring by gaussian mixture model combined evolution method; /∗ Section 4.5 ∗/
  8:
     merge parent and child populations;
  9:
     perform non-dominated sorting;
10:
     selecting the next generation population based on crowding distance;
11:
end for
12:
return fianl_population

4.4. Proposed Gaussian Model Training Strategy Based on Perturbed Solutions

The proposed training strategy is implemented by constructing a specific GMM training dataset, denoted as training_data. The procedure is presented in Algorithm 3. First, the indices of perturbed points are randomly selected from the current Pareto front solutions, denoted as pareto_solutions. The number of perturbed solutions is determined by the product of the number of current Pareto solutions (n) and the disturbance ratio of Pareto solutions ( D R ) (Lines 1–3). Then, inspired by the mutation operator in differential evolution (DE), multiple diversity-enhanced new solutions are generated for each selected Pareto solution (Lines 4–7). After repairing infeasible solutions, the feasible ones are added to the new solution set (Lines 8–9). Finally, the GMM training dataset is constructed by merging the current Pareto front set with the new solution set (Line 11). The core mutation formula of DE [54] is given as follows:
X = X + B E T A ( X a X b )
In Equation (22), relevant parameters are defined as follows: X is the original Pareto solution; X a ,   X b represent a randomly selected pair of Pareto solutions; and B E T A [ 0 , 1 ] serves as the perturbation intensity factor.
Algorithm 3 Gaussian model training strategy based on perturbed solutions
  1:
Input: pareto_solutions, DR, BETA
  2:
Output: training_data
  3:
p e r t u r b e d _ p o i n t s ;
  4:
n | p a r e t o _ s o l u t i o n s | ; / * C a l c u l a t e   t h e   n u m b e r   o f   P a r e t o   f r o n t i e r   s o l u t i o n s */
  5:
s e l e c t e d _ i n d i c e s R a n d o m S a m p l e ( { 0 , 1 , , n 1 } , n · D R ) ;
  6:
for  e a c h i d x in s e l e c t e d _ i n d i c e s  do
  7:
      X p a r e t o _ s o l u t i o n s [ i d x ] ;
  8:
      ( i 1 , i 2 ) R a n d o m S a m p l e ( { 0 , 1 , , | p a r e t o _ s o l u t i o n s | 1 } , 2 ) ;
  9:
      X X + B E T A · ( p a r e t o _ s o l u t i o n s [ i 1 ] p a r e t o _ s o l u t i o n s [ i 2 ] ) ;
10:
      X r e p a i r _ s o l u t i o n ( X ) ;  /∗ Section 4.1 ∗/
11:
      p e r t u r b e d _ p o i n t s p e r t u r b e d _ p o i n t s { X } ;
12:
end for
13:
t r a i n i n g _ d a t a p a r e t o _ s o l u t i o n s p e r t u r b e d _ p o i n t s ;
14:
return t r a i n i n g _ d a t a

4.5. Gaussian Mixture Model Combined Evolutionary Method

In GLPEA, a GMM strategy with dynamically selected component numbers per generation is employed to choose the optimal number of components for each generation. The procedure is presented in Algorithm 4. The hybrid reproduction strategy generates offspring using both GMM sampling and traditional genetic operations. Let P S denote the number of offspring generated per generation and G S R the GMM sampling rate. t r a i n i n g _ d a t a denotes the GMM training dataset constructed in Section 4.4, and p o p u l a t i o n represents the population. If the sampling rate criterion is satisfied, GMM sampling is performed; otherwise, traditional genetic operations are applied. During GMM sampling, all candidate points are flattened and normalized to a standard normal distribution (Lines 3–4). The number of components is traversed from 1 to the maximum component number ( c o m p o n e n t s M a x ), and the Bayesian Information Criterion (BIC) [55] is used to evaluate the models. The model with the smallest BIC corresponds to the optimal number of components, and this GMM is then used for sampling (Lines 5–17). After repairing infeasible solutions, the sampled offspring are added to the offspring solution set (Lines 18–20). Note that c o m p o n e n t s M a x = 5 is set when determining the optimal number of components. Based on our preliminary observations, this value is sufficient to find an appropriate component number. For the traditional genetic operation, parents are selected by binary tournament selection (Lines 23–24; see Section 4.6), followed by crossover and mutation to generate two offspring, which are then added to the offspring solution set (Lines 25–31).

4.6. Binary Tournament Parent Selection

In the traditional genetic algorithm, binary tournament selection is used to choose the parent individuals for crossover and mutation. We give the binary tournament selection is described in Algorithm 5. Let f r o n t s be the positions of individuals in each non-dominated front within the p o p u l a t i o n , and r a n k the front level to which each individual belongs.
As Algorithm 5 shows, a zero vector d i s t a n c e s with a length equal to the population size is first initialized to store the crowding distance of each individual (Line 1). Then, each non-dominated front is traversed (Line 2), and the crowding distance of all individuals in the current front is calculated (Line 3). The calculated distances are assigned to the corresponding positions in the d i s t a n c e s vector (Lines 4–5). Next, two distinct indices are randomly sampled from the set of population indices (Line 8). The selected candidate solutions are compared according to their non-dominated sorting ranks, and the solution with a lower rank (i.e., higher quality) is selected (Lines 9–12). If the two candidates have the same rank, the crowding distance is used as the criterion, and the one with a larger crowding distance is selected to maintain population diversity (Lines 14–19).
Algorithm 4 Proposed GMM combined evolutionary method
  1:
Input: training_data, population, GSR
  2:
Output: offspring, offspring_obj
  3:
o f f s p r i n g ;
  4:
o f f s p r i n g _ o b j ;
  5:
d a t a { f l a t t e n ( p o i n t ) p o i n t t r a i n i n g _ d a t a } ;
  6:
d a t a _ s c a l e d S t a n d a r d S c a l e r ( ) . f i t _ t r a n s f o r m ( d a t a ) ;
  7:
b i c _ v a l u e s [ ] ;
  8:
for  n = 1 to c o m p o n e n t s M a x  do
  9:
    g m m _ n G a u s s i a n M i x t u r e ( n _ c o m p o n e n t s = n ) ; / Section 4.2 ∗/
10:
    g m m _ n . f i t ( d a t a _ s c a l e d ) ;
11:
    b i c _ v a l u e s . a p p e n d ( g m m _ n . B I C ( d a t a _ s c a l e d ) ) ;
12:
end for
13:
g m m _ c o m p o n e n t s argmin ( b i c _ v a l u e s ) + 1 ;
14:
g m _ m o d e l G a u s s i a n M i x t u r e ( n _ c o m p o n e n t s = g m m _ c o m p o n e n t s ) ; / Section 4.2 ∗/
15:
g m _ m o d e l . f i t ( d a t a _ s c a l e d ) ;
16:
while | p o p u l a t i o n | < P S do
17:
   if  U n i f o r m R a n d o m ( 0 , 1 ) < G S R  then
18:
      / * G e n e r a t i n g   o f f s p r i n g   u s i n g   G a u s s i a n   l e a r n i n g */
19:
      c h i l d S t a n d a r d S c a l e r ( ) . i n v e r s e _ t r a n s f o r m ( g m _ m o d e l . S a m p l e ( ) ) . r e s h a p e ( I , M ) ;
20:
      c h i l d r e p a i r _ s o l u t i o n ( c h i l d ) ; / Section 4.1 ∗/
21:
      o f f s p r i n g o f f s p r i n g { c h i l d } ;
22:
      o f f s p r i n g _ o b j o f f s p r i n g _ o b j { c a l c u l a t e _ o b j e c t i v e s ( c h i l d ) } ;
23:
   else
24:
      / * G e n e r a t e   o f f s p r i n g   u s i n g   t r a d i t i o n a l   c r o s s o v e r   a n d   m u t a t i o n */
25:
      p a r e n t 1 B i n a r y T o u r n a m e n t ( p o p u l a t i o n ) ; / Section 4.6 ∗/
26:
      p a r e n t 2 B i n a r y T o u r n a m e n t ( p o p u l a t i o n ) ; / Section 4.6 ∗/
27:
      ( c h i l d 1 , c h i l d 2 ) C r o s s o v e r ( C l o n e ( p a r e n t 1 ) , C l o n e ( p a r e n t 2 ) ) ;
28:
      c h i l d 1 m u t a t i o n ( c h i l d 1 ) ;
29:
      c h i l d 2 m u t a t i o n ( c h i l d 2 ) ;
30:
      o f f s p r i n g o f f s p r i n g { c h i l d 1 } ;
31:
      o f f s p r i n g _ o b j o f f s p r i n g _ o b j { c a l c u l a t e _ o b j e c t i v e s ( c h i l d 1 ) } ;
32:
      o f f s p r i n g o f f s p r i n g { c h i l d 2 } ;
33:
      o f f s p r i n g _ o b j o f f s p r i n g _ o b j { c a l c u l a t e _ o b j e c t i v e s ( c h i l d 2 ) } ;
34:
   end if
35:
end while
36:
return ( o f f s p r i n g , o f f s p r i n g _ o b j )
Algorithm 5 Adopted binary tournament selection operator
  1:
Input: population
  2:
Output: parent
  3:
d i s t a n c e s Z e r o V e c t o r ( l e n g t h = | p o p u l a t i o n | ) ;
  4:
for  e a c h f r o n t in f r o n t s  do
  5:
    f r o n t _ d i s t c r o w d i n g _ d i s t a n c e ( f r o n t ) ;
  6:
   for  e a c h ( i d x , d ) in z i p ( f r o n t , f r o n t _ d i s t )  do
  7:
      d i s t a n c e s [ i d x ] d ;
  8:
   end for
  9:
end for
10:
( i 1 , i 2 ) R a n d o m S a m p l e ( { 0 , 1 , , | p o p u l a t i o n | 1 } , 2 ) ;
11:
if  r a n k [ i 1 ] < r a n k [ i 2 ]  then
12:
    s p o p u l a t i o n [ i 1 ] ;
13:
else if  r a n k [ i 2 ] < r a n k [ i 1 ]  then
14:
    s p o p u l a t i o n [ i 2 ] ;
15:
else
16:
   if  d i s t a n c e [ i 1 ] > d i s t a n c e [ i 2 ]  then
17:
      s p o p u l a t i o n [ i 1 ] ;
18:
   else
19:
      s p o p u l a t i o n [ i 2 ] ;
20:
   end if
21:
end if
22:
return  p a r e n t

5. Experiment Results and Comparisons

5.1. Test Instances and Experimental Design

We consider the combinations M × I = { 3 , 5 , 7 , 9 , 11 , 13 , 15 , 17 } × { 4 , 6 , 8 , 10 , 12 } , yielding 40 instances. For each instance, GLPEA and the comparison algorithms are each run independently 20 times. Experimental results are reported as Best, Worst, and Mean, representing the best, worst, and average values of the performance metrics. All algorithms are implemented in Python 3.8 and executed on a PC running macOS with an Intel Core i7 2.7 GHz quad-core processor and 16 GB of RAM (The source code, test instances, and results are available at https://gitee.com/zhangjinsi_p/glpea.git, accessed on 30 September 2024).

5.2. Performance Metrics

To evaluate the solution quality achieved by GLPEA and compared algorithms, two well-known MOEA performance metrics are used in this work [56].
IGD quantifies the proximity between the solution sets generated by an algorithm and the reference Pareto front (or true Pareto front). It calculates the average Euclidean distance from each point in the reference Pareto front to its nearest neighbor in the algorithm’s solution set. The metric is formulated as follows:
I G D ( P * , A ) = 1 | P * | v P * min u A v u
where P * denotes the reference Pareto front (or true Pareto front), and A represents the solution set obtained by a given compared algorithm. Note that we combine the solutions obtained by all compared algorithm and then identify P * from these solutions. A smaller IGD value indicates closer proximity to the true Pareto front, reflecting superior convergence performance.
HV measures the multidimensional space between a reference point r and the solution set, reflecting the breadth and diversity of the distribution. HV is defined as follows:
H V ( A ) = V o l u m e x A [ x 1 , r 1 ] × [ x 2 , r 2 ] × × [ x m , r m ]
where r = ( r 1 , r 2 , , r m ) represents a predefined reference point in the objective space. A larger HV value signifies stronger exploration capability, with solutions exhibiting a more comprehensive and balanced distribution in the objective space.

5.3. Parameter Settings

GLPEA has six parameters: population size ( P S ), crossover probability ( P c ), mutation probability ( P m ), Gaussian sampling rate ( G S R ), disturbance scaling factor ( B E T A ), and disturbed ratio of the Pareto solutions ( D R ). To find a suitable parameter combination of these parameters, the Taguchi method is employed using the instance “m7i12”. In this approach, the search performance of GLPEA under different parameter settings is evaluated using the two performance metrics. A design-of-experiments (DOE) procedure is adopted to generate candidate parameter combinations. As shown in Table 5, four levels are assigned to each parameter, and an L 4 6 orthogonal array with 25 experimental runs is used. For each parameter combination, GLPEA is executed independently 20 times with G E N S = 1000 . The corresponding performance metrics are given (see Table 6). A proper parameter combination should minimize IGD while maximizing HV. To facilitate comparison, both IGD and HV values are normalized, and the signal-to-noise ratio ( S / N ) is used as the response variable. Our goal is to identify the parameter combination that yields the highest S / N ratio, which is calculated as follows:
S / N = 10 log 1 m k = 1 m 1 sum k 2
where m is the number of GLPEA executions, and s u m k 2 represents the composite performance score obtained from the normalized IGD and HV values in the k-th run.
Based on the results in Table 6, the level trends of the parameters are illustrated in Figure 3. The six parameters are set as follows: P S = 200 , P c = 0.9 , P m = 0.2 , G S R = 0.3 , β = 0.2 , and D R = 0.3 . To validate these settings, additional parameter tuning experiments are conducted on the instances “m7i4" and “m11i4", which confirms the above parameter settings.

5.4. Effectiveness of GLPEA’s Search Components

We evaluate the effectiveness of the search components of GLPEA by comparing it with two variant algorithms: GLPEA_V1 and GLPEA_V2. GLPEA_V1 is a variant in which the generational dynamic component-selection strategy of GMM is replaced with a single-component Gaussian model. In GLPEA_V2, the Gaussian model training based on disturbed solutions is removed, and only the GMM sampling mechanism is retained for offspring generation. For fair comparisons, all algorithms are run with G E N S = 1000 . The comparative results are given in Table 7 and Table 8.
As Table 7 and Table 8 show, GLPEA outperforms the two variant algorithms in both IGD and HV, indicating its superior solution quality. To verify the statistical significance of these improvements, we performed the Wilcoxon signed-rank test [57]. Statistical analysis shows that the p-values (presented at the bottom of the table) for all two algorithm pairs are all less than 0.05, confirming the significant improvement in convergence and exploration capabilities achieved by GLPEA. These results validate the effectiveness of the generation-wise dynamic component-selection strategy and the disturbed-solution-based GMM training method, combined with conventional genetic operations for offspring generation, within the nondominated sorting framework. The proposed approach demonstrates considerable potential for applications in other multi-objective evolutionary algorithms.

5.5. Comparisons with State-of-the-Art MOEAs

We evaluate the performance of GLPEA by comparing it with two state-of-the-art MOEAs: NSGA-II and MOEA/D. Both reference algorithms are configured using the parameter settings reported in their original studies, and the same stopping condition described in Section 5.4 is applied. To facilitate a quantitative comparison, we introduce the metric N B , which denotes the number of instances in which an algorithm achieves the best value for a given quality indicator. The comparative results are reoprted in Table 9 and Table 10, and the distributions of N B values are shown in Figure 4 and Figure 5.
From Table 9 and Table 10, one observes that GLPEA outperforms the two reference algorithms in terms of worst, best, and average for almost all these problem instances. The p-values obtained from the Wilcoxon signed-rank test (presented at the bottom of the table) were all substantially less than 0.05 . This indicates that the observed improvements of GLPEA in both convergence and diversity are statistically significant. Furthermore, as shown in Figure 4 and Figure 5, GLPEA achieves higher N B values than those of the reference algorithms, indicating its superior consistency in obtaining the best search performance. The statistical distributions of the results are further analyzed through the box plots for selected instances (“m11i4”, “m13i4”, and “m15i4”) in Figure 6, and violin plots summarizing the IGD and HV results are provided in Figure 7. These distributions collectively confirm the competitive search capability of GLPEA. Based on the comprehensive experimental analysis, GLPEA can be regarded as an effective solution approach for HPMPP. Given the practical relevance of HPMPP, the proposed algorithm holds promise for extension to other production planning problems in industrial silicon manufacturing.

5.6. Computational Overhead and Limitations Analysis

5.6.1. Analysis of Computational Time

This section analyzes the computational time of the GLPEA. All experiments were executed on the hardware described in Section 5.1. Notably, the GLPEA does not require specialized hardware resources, such as GPUs, for execution. For each instance, GLPEA and the compared algorithms are executed independently 20 times with G E N S = 1000 . To quantitatively compare the execution time, Table 11 presents the average execution time (in seconds) of the three algorithms across 40 instances.
As detailed in Table 9, we observed that the computational time generally increases with the size of the problem instance for most cases. The execution time of GLPEA is significantly higher than NSGA-II but lower than MOEA/D. This finding aligns with the higher computational complexity inherent in the Gaussian Mixture Model (GMM) training integrated into GLPEA. This increase in computational overhead is anticipated; unlike the lightweight stochastic operators of NSGA-II, the primary computational burden of GLPEA stems from the GMM training.

5.6.2. Analysis of Computational Complexity

To theoretically analyze the computational overhead of GLPEA, we base our analysis on the HPMPP problem studied in this research, where P S is the population size and I × M is the dimension of the problem’s decision variables. We analyze the main computational complexity per generation of the algorithm. (1) GLPEA adopts the non-dominated sorting framework of NSGA-II. Since the number of objective functions is a constant (three), the computational complexity of non-dominated sorting can be simplified to O ( ( P S ) 2 ) ; the complexity of calculating the crowding distance is O ( ( P S ) log ( P S ) ) . Thus, the complexity of the traditional evolutionary mechanism is O ( ( P S ) 2 ) . (2) Based on the analysis presented in Section 4.2.3, the computational complexity of the GMM training is primarily determined by the calculation of the covariance matrix, which is approximately O ( t · k · ( P S ) · ( I × M ) 2 ) , where t is the number of EM iterations and k is the number of components. Since t and k are constants, the complexity of GMM training is simplified to O ( ( P S ) · ( I × M ) 2 ) . In summary, the total computational complexity of GLPEA per generation is O ( ( P S ) 2 + ( P S ) · ( I × M ) 2 ) .

5.6.3. Method Limitations and Industrial Applicability

As discussed above, the complexity scales with O ( ( P S ) · ( I × M ) 2 ) due to the GMM training as the problem size I × M increases, leading to a rapid increase in the algorithm’s computational time. This observation is consistent with the computational time trend of GLPEA shown in Table 11. Regarding parameter sensitivity, GLPEA introduces new parameters such as G S R , B E T A , and D R . Although this study determined a suitable combination of parameters in Section 5.3, the algorithm’s performance might be sensitive to these parameters. Re-tuning the parameters may be required when applying the algorithm to new problems, which can be time-consuming.
Despite these limitations, in the context of HPMPP, planning is typically executed periodically and offline. Therefore, obtaining a superior Pareto front, as shown in Table 9 and Table 10, within the computational time demonstrated in Table 11 on modern computing hardware, is fully acceptable. Thus, GLPEA retains strong industrial applicability.

5.7. Discussions

We further evaluate the solution quality of GLPEA concerning the following issues: (a) the trends of IGD and HV values for GLPEA, NSGA-II, and MOEA/D; and (b) a comparison of APFs obtained by the three algorithms. These analyses help verify the search behavior of GLPEA and the distribution of its final solutions.

5.7.1. Trends of IGD and HV Achieved by GLPEA, NSGA-II, and MOEA/D

Using the instance “m7i4”, we execute GLPEA, NSGA-II, and MOEA/D with a stopping condition of G E N S = 1000 . At each generation, the IGD and HV values are calculated and recorded based on the nondominated solutions identified from the populations of the three algorithms. The resulting trends of IGD and HV of GLPEA and the two reference approaches are given in Figure 8 and Figure 9.
As shown in Figure 8 and Figure 9, GLPEA consistently achieves lower IGD values and higher HV values than NSGA-II and MOEA/D throughout the whole iterative process. This indicates a smaller distance to the reference Pareto front and a broader coverage of the objective space, collectively confirming GLPEA’s superior capability in exploring diverse regions of the Pareto front.

5.7.2. APFs Solved by GLPEA, NSGA-II, and MOEA/D

To analyze the distributions of APFs obtained by GLPEA and the two state-of-the-art solution methods, we draw the APFs of the three methods in Figure 10. We note that this analysis is again based on the instance “m7i4” and under the stopping condition of G E N S = 1000 .
Based on the distributions of obtained APFs shown in Figure 10, GLPEA achieves superior uniformity and coverage across all three objectives in the Pareto front, outperforming NSGA-II (which ranks second) and MOEA/D (which shows the weakest performance in both coverage and distribution uniformity). Moreover, these results reflect inherent conflicts among the three optimization objectives: emission control, rollover penalty costs, and load balancing. As a result, a significant improvement in one objective generally leads to deterioration in the others, resulting in a dispersed Pareto front rather than a single optimal solution.

6. Conclusions

We investigate a multi-objective heterogeneous parallel machine scheduling problem arising from industrial silicon smelting, with the aim of simultaneously minimizing carbon emissions, rollover penalty costs, and load imbalance. The model incorporates practical constraints including machine maintenance schedules and monthly production demands, while carbon emission factors are predicted using a machine learning approach for a 12-month horizon. From the algorithmic perspective, the proposed GLPEA introduces a generation-wise dynamic component-selection strategy within a GMM framework, enhanced by a disturbed-solution-set-based training scheme, and integrated with conventional genetic operators for offspring generation. Extensive computational experiments demonstrate that GLPEA achieves superior performance compared to state-of-the-art MOEAs in terms of both IGD and HV metrics.
For future research, carbon emission forecasting could be improved through advanced techniques such as multimodal fusion prediction. Further enhancement of the algorithm may also be achieved by developing more adaptive mechanisms for determining the number of Gaussian mixture components during evolution.

Author Contributions

Conceptualization: R.L. and Z.L.; Methodology: J.Z., R.L. and Z.L.; Software: J.Z.; Validation: J.Z. and Z.L.; Investigation: J.Z., R.L. and Z.L.; Data Curation: J.Z.; Writing—Original Draft: J.Z.; Writing—Review & Editing: R.L. and Z.L.; Supervision: R.L. and Z.L.; Project Administration: R.L.; Funding Acquisition: Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (Grant Numbers: 72202202 and 72362026), the Yunnan Fundamental Research Project (Grant Numbers: 202301AU070086, 202301AT070458, and 202301BE070001-003), the Yunnan Philosophy and Social Sciences Planning Project (Grant Number: YB202589), the “AI +” Special Research Project of Humanities and Social Sciences at Kunming University of Science and Technology (Grant Number: RZZX202502), and the Academic Excellence Cultivation Project of Kunming University of Science and Technology (Grant Number: JPSC2025010).

Data Availability Statement

The data presented in this study are openly available in the source code, test instances, and results at https://gitee.com/zhangjinsi_p/glpea.git, accessed on 30 September 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, Z.; Ma, W.; Wu, J.; Wei, K.; Lei, Y.; Lv, G. A study of the performance of submerged arc furnace smelting of industrial silicon. Silicon 2018, 10, 1121–1127. [Google Scholar] [CrossRef]
  2. Li, Z.C.; Qian, B.; Hu, R.; Chang, L.L.; Yang, J.B. An elitist nondominated sorting hybrid algorithm for multi-objective flexible job-shop scheduling problem with sequence-dependent setups. Knowl.-Based Syst. 2019, 173, 83–112. [Google Scholar] [CrossRef]
  3. Fallahi, A.; Shahidi-Zadeh, B.; Niaki, S.T.A. Unrelated parallel batch processing machine scheduling for production systems under carbon reduction policies: NSGA-II and MOGWO metaheuristics. Soft Comput. 2023, 27, 17063–17091. [Google Scholar] [CrossRef]
  4. Seada, H.; Deb, K. Effect of selection operator on NSGA-III in single, multi, and many-objective optimization. In Proceedings of the 2015 IEEE Congress on Evolutionary Computation (CEC), Sendai, Japan, 25–28 May 2015; pp. 2915–2922. [Google Scholar] [CrossRef]
  5. Tian, Y.; Cheng, R.; Zhang, X.; Su, Y.; Jin, Y. A Strengthened Dominance Relation Considering Convergence and Diversity for Evolutionary Many-Objective Optimization. IEEE Trans. Evol. Comput. 2018, 23, 331–345. [Google Scholar] [CrossRef]
  6. Li, J.; Nehorai, A. Gaussian mixture learning via adaptive hierarchical clustering. Signal Process. 2018, 150, 116–121. [Google Scholar] [CrossRef]
  7. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
  8. Zitzler, E.; Laumanns, M.; Thiele, L. SPEA2: Improving the strength Pareto evolutionary algorithm. TIK Rep. 2001, 103. [Google Scholar] [CrossRef]
  9. Zhang, Q.; Li, H. MOEA/D: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2008, 11, 712–731. [Google Scholar] [CrossRef]
  10. Jaszkiewicz, A. On the performance of multiple-objective genetic local search on the 0/1 knapsack problem: A comparative experiment. IEEE Trans. Evol. Comput. 2002, 6, 402–412. [Google Scholar] [CrossRef]
  11. Zitzler, E.; Künzli, S. Indicator-based selection in multi-objective search. In International Conference on Parallel Problem Solving from Nature; Springer: Berlin/Heidelberg, Germany, 2004; pp. 832–842. [Google Scholar]
  12. Beume, N.; Naujoks, B.; Emmerich, M. SMS-EMOA: Multi-objective selection based on dominated hypervolume. Eur. J. Oper. Res. 2007, 181, 1653–1669. [Google Scholar] [CrossRef]
  13. Bandyopadhyay, S.; Bhattacharya, R. Solving multi-objective parallel machine scheduling problem by a modified NSGA-II. Appl. Math. Model. 2013, 37, 6718–6729. [Google Scholar] [CrossRef]
  14. Rego, M.F.; Pinto, J.C.E.; Cota, L.P.; Souza, M.J. A mathematical formulation and an NSGA-II algorithm for minimizing the makespan and energy cost under time-of-use electricity price in an unrelated parallel machine scheduling. PeerJ Comput. Sci. 2022, 8, e844. [Google Scholar] [CrossRef]
  15. Wang, Z.; Hong, H.; Ye, K.; Zhang, G.E.; Jiang, M.; Tan, K.C. Manifold interpolation for large-scale multiobjective optimization via generative adversarial networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 4631–4645. [Google Scholar] [CrossRef] [PubMed]
  16. Zheng, W.; Doerr, B. Runtime analysis for the NSGA-II: Proving, quantifying, and explaining the inefficiency for many objectives. IEEE Trans. Evol. Comput. 2023, 28, 1442–1454. [Google Scholar] [CrossRef]
  17. Balekelayi, N.; Woldesellasse, H.; Tesfamariam, S. Comparison of the performance of a surrogate based Gaussian process, NSGA2 and PSO multi-objective optimization of the operation and fuzzy structural reliability of water distribution system: Case study for the City of Asmara, Eritrea. Water Res. Manag. 2022, 36, 6169–6185. [Google Scholar] [CrossRef]
  18. Rossmann, J.; Kamper, M.J.; Hackl, C.M. A Global Multi-Objective Bayesian Optimization Framework for Generic Machine Design Using Gaussian Process Regression. IEEE Trans. Energy Convers. 2025, 40, 2384–2398. [Google Scholar] [CrossRef]
  19. Hebbal, A.; Balesdent, M.; Brevault, L.; Melab, N.; Talbi, E.G. Deep Gaussian process for multi-objective Bayesian optimization. Optim. Eng. 2023, 24, 1809–1848. [Google Scholar] [CrossRef]
  20. Wu, H.; Jin, Y.; Gao, K.; Ding, J.; Cheng, R. Surrogate-assisted evolutionary multi-objective optimization of medium-scale problems by random grouping and sparse Gaussian modeling. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 3263–3278. [Google Scholar] [CrossRef]
  21. Ravi, K.; Fediukov, V.; Dietrich, F.; Neckel, T.; Buse, F.; Bergmann, M.; Bungartz, H.J. Multi-fidelity Gaussian process surrogate modeling for regression problems in physics. Mach. Learn. Sci. Technol. 2024, 5, 045015. [Google Scholar] [CrossRef]
  22. Chen, N.; Digel, C.; Doppelbauer, M. Uncertainty Quantification-Based Multi-Objective Optimization Design of Electrical Machines Using Probabilistic Metamodels. IEEE Trans. Energy Convers. 2024, 40, 860–872. [Google Scholar] [CrossRef]
  23. Zhang, H.; Ding, J.; Jiang, M.; Tan, K.C.; Chai, T. Inverse gaussian process modeling for evolutionary dynamic multiobjective optimization. IEEE Trans. Cybern. 2021, 52, 11240–11253. [Google Scholar] [CrossRef]
  24. Yang, Z.; Qiu, H.; Gao, L.; Chen, L.; Liu, J. Surrogate-assisted MOEA/D for expensive constrained multi-objective optimization. Inf. Sci. 2023, 639, 119016. [Google Scholar] [CrossRef]
  25. Niu, Y.; Shao, J.; Xiao, J.; Song, W.; Cao, Z. Multi-objective evolutionary algorithm based on RBF network for solving the stochastic vehicle routing problem. Inf. Sci. 2022, 609, 387–410. [Google Scholar] [CrossRef]
  26. Sonoda, T.; Nakata, M. Multiple classifiers-assisted evolutionary algorithm based on decomposition for high-dimensional multiobjective problems. IEEE Trans. Evol. Comput. 2022, 26, 1581–1595. [Google Scholar] [CrossRef]
  27. Xu, D.; Jiang, M.; Hu, W.; Li, S.; Pan, R.; Yen, G.G. An online prediction approach based on incremental support vector machine for dynamic multiobjective optimization. IEEE Trans. Evol. Comput. 2021, 26, 690–703. [Google Scholar] [CrossRef]
  28. Zhu, E.; Chen, Z.; Cui, J.; Zhong, H. MOE/RF: A novel phishing detection model based on revised multiobjective evolution optimization algorithm and random forest. IEEE Trans. Netw. Serv. Manag. 2022, 19, 4461–4478. [Google Scholar] [CrossRef]
  29. De Moraes, M.B.; Coelho, G.P. Effects of the random forests hyper-parameters in surrogate models for multi-objective combinatorial optimization: A case study using moea/d-rfts. IEEE Lat. Am. Trans. 2023, 21, 621–627. [Google Scholar] [CrossRef]
  30. Hauschild, M.; Pelikan, M. An introduction and survey of estimation of distribution algorithms. Swarm Evol. Comput. 2011, 1, 111–128. [Google Scholar] [CrossRef]
  31. Wang, F.; Liao, F.; Li, Y.; Wang, H. A new prediction strategy for dynamic multi-objective optimization using Gaussian Mixture Model. Inf. Sci. 2021, 580, 331–351. [Google Scholar] [CrossRef]
  32. Li, G.; Wang, Z.; Zhang, Q.; Sun, J. Offline and online objective reduction via Gaussian mixture model clustering. IEEE Trans. Evol. Comput. 2022, 27, 341–354. [Google Scholar] [CrossRef]
  33. Zou, J.; Hou, Z.; Jiang, S.; Yang, S.; Ruan, G.; Xia, Y.; Liu, Y. Knowledge Transfer With Mixture Model in Dynamic Multi-Objective Optimization. IEEE Trans. Evol. Comput. 2025, 29, 1517–1530. [Google Scholar] [CrossRef]
  34. Lu, C.; Liu, Y.; Zhang, Q. MOEA/D-CMA Made Better with (l+l)-CMA-ES. In Proceedings of the 2024 IEEE Congress on Evolutionary Computation (CEC), Yokohama, Japan, 30 June–5 July 2024; pp. 1–8. [Google Scholar]
  35. Aggarwal, S.; Tripathi, S. MODE/CMA-ES: Integrated multi-operator differential evolution technique with CMA-ES. Appl. Soft Comput. 2025, 176, 113177. [Google Scholar] [CrossRef]
  36. Zhang, W.; Wang, S.; Zhou, A.; Zhang, H. A practical regularity model based evolutionary algorithm for multiobjective optimization. Appl. Soft Comput. 2022, 129, 109614. [Google Scholar] [CrossRef]
  37. Kalita, K.; Ramesh, J.V.N.; Cepova, L.; Pandya, S.B.; Jangir, P.; Abualigah, L. Multi-objective exponential distribution optimizer (MOEDO): A novel math-inspired multi-objective algorithm for global optimization and real-world engineering design problems. Sci. Rep. 2024, 14, 1816. [Google Scholar] [CrossRef] [PubMed]
  38. Alghamdi, A.S.; Zohdy, M.A. Boosting cuckoo optimization algorithm via Gaussian mixture model for optimal power flow problem in a hybrid power system with solar and wind renewable energies. Heliyon 2024, 10, e31755. [Google Scholar] [CrossRef] [PubMed]
  39. Guerrero-Peña, E.; Araújo, A.F.R. A new dynamic multi-objective evolutionary algorithm without change detector. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 635–640. [Google Scholar]
  40. Yazdi, F.; Asadi, S. Enhancing Cardiovascular Disease Diagnosis: The Power of Optimized Ensemble Learning. IEEE Access 2025, 13, 46747–46762. [Google Scholar] [CrossRef]
  41. Zhang, J.; Shi, X. Many-objective evolutionary optimization algorithm based on the Gaussian mixture models. In Proceedings of the Fifth International Conference on Applied Mathematics, Modelling, and Intelligent Computing (CAMMIC 2025), Shanghai, China, 21–23 March 2025; SPIE: Bellingham, WA, USA, 2025; Volume 13644, pp. 457–462. [Google Scholar]
  42. Abdulghani, A.M.; Abdullah, A.; Rahiman, A.R.; Abdul Hamid, N.A.W.; Akram, B.O. Network-Aware Gaussian Mixture Models for Multi-Objective SD-WAN Controller Placement. Electronics 2025, 14, 3044. [Google Scholar] [CrossRef]
  43. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  44. Bachechi, C.; Rollo, F.; Po, L. Detection and classification of sensor anomalies for simulating urban traffic scenarios. Clust. Comput. 2022, 25, 2793–2817. [Google Scholar] [CrossRef]
  45. Bergmeir, C.; Benítez, J.M. On the use of cross-validation for time series predictor evaluation. Inf. Sci. 2012, 191, 192–213. [Google Scholar] [CrossRef]
  46. Box, G. Box and Jenkins: Time series analysis, forecasting and control. In A Very British Affair: Six Britons and the Development of Time Series Analysis During the 20th Century; Palgrave Macmillan UK: London, UK, 2013; pp. 161–215. [Google Scholar]
  47. Ma, J.; Ding, Y.; Cheng, J.C.; Jiang, F.; Gan, V.J.; Xu, Z. A Lag-FLSTM deep learning network based on Bayesian Optimization for multi-sequential-variant PM2.5 prediction. Sustain. Cities Soc. 2020, 60, 102237. [Google Scholar] [CrossRef]
  48. Dong, L.; Fang, D.; Wang, X.; Wei, W.; Damaševičius, R.; Scherer, R.; Woźniak, M. Prediction of streamflow based on dynamic sliding window LSTM. Water 2020, 12, 3032. [Google Scholar] [CrossRef]
  49. Köksoy, O. Multiresponse robust design: Mean square error (MSE) criterion. Appl. Math. Comput. 2006, 175, 1716–1729. [Google Scholar] [CrossRef]
  50. Applegate, R.A.; Ballentine, C.; Gross, H.; Sarver, E.J.; Sarver, C.A. Visual acuity as a function of Zernike mode and level of root mean square error. Optom. Vis. Sci. 2003, 80, 97–105. [Google Scholar] [CrossRef] [PubMed]
  51. Sammut, C.; Webb, G.I. Mean absolute error. Encycl. Mach. Learn. 2010, 652, 14. [Google Scholar]
  52. Hansen, L.P. Generalized Method of Moments Estimation. In The New Palgrave Dictionary of Economics; Palgrave Macmillan: London, UK, 2018. [Google Scholar] [CrossRef]
  53. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1977, 39, 1–22. [Google Scholar] [CrossRef]
  54. Babu, B.V.; Jehan, M.M.L. Differential evolution for multi-objective optimization. In Proceedings of the 2003 Congress on Evolutionary Computation, Canberra, Australia, 8–12 December 2003; IEEE: Piscataway, NJ, USA, 2003; Volume 4, pp. 2696–2703. [Google Scholar]
  55. Schwarz, G. Estimating the dimension of a model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  56. Riquelme, N.; Von Lücken, C.; Baran, B. Performance metrics in multi-objective optimization. In Proceedings of the 2015 Latin American Computing Conference (CLEI), Arequipa, Peru, 19–23 October 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–11. [Google Scholar]
  57. Woolson, R.F. Wilcoxon signed-rank test. Wiley Encycl. Clin. Trials 2007, 1–3. [Google Scholar] [CrossRef]
Figure 1. Illustration of industrial silicon smelting process.
Figure 1. Illustration of industrial silicon smelting process.
Mathematics 13 03860 g001
Figure 2. Overview of planning for industrial silicon production.
Figure 2. Overview of planning for industrial silicon production.
Mathematics 13 03860 g002
Figure 3. Level trends of parameters ( S / N ).
Figure 3. Level trends of parameters ( S / N ).
Mathematics 13 03860 g003
Figure 4. Distributions of NB values obtianed by NSGA-II, MOEA/D, and GLPEA (IGD).
Figure 4. Distributions of NB values obtianed by NSGA-II, MOEA/D, and GLPEA (IGD).
Mathematics 13 03860 g004
Figure 5. Distributions of NB values obtianed by NSGA-II, MOEA/D, and GLPEA (HV).
Figure 5. Distributions of NB values obtianed by NSGA-II, MOEA/D, and GLPEA (HV).
Mathematics 13 03860 g005
Figure 6. Box plots for NSGA-II, MOEA/D, and GLPEA.
Figure 6. Box plots for NSGA-II, MOEA/D, and GLPEA.
Mathematics 13 03860 g006
Figure 7. Violin plots for MOEA/D, NSGA-II, and GLPEA (Mean).
Figure 7. Violin plots for MOEA/D, NSGA-II, and GLPEA (Mean).
Mathematics 13 03860 g007
Figure 8. Trends of IGD obtained by GLPEA, NSGA-II, and MOEA/D.
Figure 8. Trends of IGD obtained by GLPEA, NSGA-II, and MOEA/D.
Mathematics 13 03860 g008
Figure 9. Trends of HV obtained by GLPEA, NSGA-II, and MOEA/D.
Figure 9. Trends of HV obtained by GLPEA, NSGA-II, and MOEA/D.
Mathematics 13 03860 g009
Figure 10. APFs obtained by GLPEA, NSGA-II, MOEA/D.
Figure 10. APFs obtained by GLPEA, NSGA-II, MOEA/D.
Mathematics 13 03860 g010
Table 1. The model parameters of the industrial silicon scheduling problem.
Table 1. The model parameters of the industrial silicon scheduling problem.
SymbolRecommended Range of ValuesDescription
M { 1 , 2 , , 20 } Set of machines and it also denotes the cardinality of the set.
I { 1 , 2 , , 12 } Set of months and it also denotes the cardinality of the set.
H i { 28 , 29 , 30 , 31 } Number of days in month i I . Defines the number of days within the production planning period for each month.
P i ( 0 , ) Production demand for month i. Used to constrain the total production of all machines to be equal to the overall demand.
C i j ( 0 , ) Carbon emission factor of machine j M in month i I . The result value is obtained through Section 3.5.5.
D i ( 0 , ) Rollover penalty coefficient for month i. Used to calculate the monthly rollover penalty, defined as the unfinished production of each month multiplied by the monthly rollover penalty coefficient.
R i j { 1 , 2 , , 31 } Scheduled maintenance days for machine j in month i. Used to calculate the remaining monthly rated capacity of machines after subtracting maintenance days from the number of days in the production planning period of each month.
C a p j ( 0 , ) Maximum monthly production capacity of machine j. Used to impose machine load constraints.
Note: Parameter values must also comply with the constraints specified in Section 3.4.
Table 2. The algorithm parameters of the proposed GLPEA.
Table 2. The algorithm parameters of the proposed GLPEA.
SymbolRecommended Range of ValuesDescription
P S { 50 , 100 , 150 , 200 } Population size. Determines the total number of solutions (individuals) maintained by the algorithm in each generation.
P c { 0.6 , 0.7 , 0.8 , 0.9 } Crossover probability. Controls the probability of generating new offspring through the crossover process in traditional genetic operations.
P m { 0.1 , 0.2 , 0.3 , 0.4 } Mutation probability. Controls the probability of random mutation occurring in offspring individuals during traditional genetic operations.
G S R { 0.3 , 0.5 , 0.7 , 0.9 } Gaussian sampling rate. Determines the proportion of individuals in the new generation that are generated through GMM sampling rather than traditional genetic operations.
B E T A { 0.2 , 0.4 , 0.6 , 0.8 } Disturbance scaling factor. Controls the scaling factor used when generating perturbed solutions.
D R { 0.1 , 0.3 , 0.5 , 0.7 } Disturbed ratio of the Pareto solutions. Determines the proportion of Pareto solutions used to generate “perturbed solutions” in the perturbed solution set training strategy.
Table 3. Experimental results of model parameters.
Table 3. Experimental results of model parameters.
lstm_unitdropout_rateMSE
320.10.0087
320.30.0078
320.50.0091
480.10.0074
480.30.0067
480.50.0080
640.10.0088
640.30.0079
640.50.0081
Table 4. Predicted carbon emission factors for furnaces #1–#7 over the next 12 months.
Table 4. Predicted carbon emission factors for furnaces #1–#7 over the next 12 months.
Month#1#2#3#4#5#6#7
113.24912.79712.30512.05811.82713.08413.247
211.95911.92412.48411.80111.57112.35312.636
311.35611.61712.14611.44511.72911.75011.708
412.82312.48012.51112.18412.17912.76512.600
512.43512.28912.02511.85012.27612.48412.299
613.38212.82412.10712.16712.74913.12412.960
711.27011.53811.04711.03312.04411.66311.322
811.82711.86611.17611.20312.34412.03411.843
912.20612.18711.19511.39313.01912.50411.744
1010.30410.67910.13111.07811.27710.84410.626
1110.71411.06511.21010.68611.79311.21011.028
1212.20312.12511.64011.55213.49212.56412.662
Table 5. Parameter levels.
Table 5. Parameter levels.
ParameterLevel 1Level 2Level 3Level 4
P S 50100150200
P c 0.60.70.80.9
P m 0.10.20.30.4
G S R 0.30.50.70.9
B E T A 0.20.40.60.8
D R 0.10.30.50.7
Table 6. Orthogonal array and response values.
Table 6. Orthogonal array and response values.
No.Combination S / N
PS Pc Pm GSR BETA DR
13334113.075
24231414.106
3111111−3.595
4113334−8.018
54443214.511
64111124.061
73411443.433
8111241−3.975
92221311.692
10121422−5.032
112144431.597
122113111.502
13142113−1.217
143122212.902
15113123−5.308
164124143.643
172432122.415
183213133.059
19134111−5.297
20141431−5.052
21124214−5.454
224312334.153
233141323.031
242311241.142
25132342−0.608
Table 7. Comparisons of GLPEA, GLPEA_V1, and GLPEA_V2 (IGD).
Table 7. Comparisons of GLPEA, GLPEA_V1, and GLPEA_V2 (IGD).
InstanceGLPEAGLPEA_V1GLPEA_V2
Best Mean Best Mean Best Mean
m3i40.0310.0350.0330.0370.4060.494
m3i60.0350.0380.0360.0400.3470.415
m3i80.0460.0490.0450.0500.3520.387
m3i100.0450.0500.0460.0520.3420.373
m3i120.0420.0500.0450.0530.3850.429
m5i40.0320.0360.0360.0420.4820.589
m5i60.0360.0560.0430.0640.3950.438
m5i80.0440.0490.0450.0520.3400.407
m5i100.0440.0490.0400.0520.3510.386
m5i120.0430.0530.0510.0580.3880.434
m7i40.0400.0490.0450.0530.4490.500
m7i60.0360.0450.0480.0540.4400.526
m7i80.0480.0560.0460.0550.3510.407
m7i100.0510.0580.0460.0580.3730.416
m7i120.0490.0610.0530.0640.4580.505
m9i40.0440.0500.0460.0530.3850.525
m9i60.0430.0550.0470.0600.3780.520
m9i80.0460.0550.0480.0570.4210.467
m9i100.0520.0640.0530.0640.4000.439
m9i120.0510.0660.0560.0690.5290.560
m11i40.0430.0560.0510.0720.5910.660
m11i60.0450.0550.0470.0630.4770.506
m11i80.0440.0550.0460.0560.3670.388
m11i100.0550.0630.0600.0690.3730.404
m11i120.0520.0640.0550.0660.4580.484
m13i40.0560.0630.0570.0740.5300.574
m13i60.0390.0550.0450.0600.4730.509
m13i80.0490.0600.0570.0660.4140.441
m13i100.0550.0670.0540.0670.3600.391
m13i120.0500.0700.0600.0760.4830.509
m15i40.0410.0480.0470.0530.5400.598
m15i60.0480.0620.0580.0750.5410.570
m15i80.0520.0590.0540.0650.3990.426
m15i100.0460.0590.0550.0650.3900.421
m15i120.0480.0580.0490.0640.4720.497
m17i40.0470.0580.0470.0640.5830.621
m17i60.0490.0630.0500.0710.5380.562
m17i80.0460.0570.0540.0650.4050.430
m17i100.0530.0630.0540.0670.3710.400
m17i120.0560.0680.0560.0770.4510.468
p-value0.0000.0320.0000.000
Table 8. Comparisons of GLPEA, GLPEA_V1, and GLPEA_V2 (HV).
Table 8. Comparisons of GLPEA, GLPEA_V1, and GLPEA_V2 (HV).
InstanceGLPEAGLPEA_V1GLPEA_V2
Best Mean Best Mean Best Mean
m3i40.7300.7240.7230.7190.4180.239
m3i60.8220.8150.8140.8100.3930.329
m3i80.8480.8320.8420.8310.4290.402
m3i100.8310.8170.8280.8120.4350.394
m3i120.8640.8380.8450.8310.3980.358
m5i40.7760.7670.7690.7620.3100.162
m5i60.8330.8230.8260.8100.3510.292
m5i80.8380.8160.8320.8110.3810.314
m5i100.8640.8480.8680.8390.4300.390
m5i120.8590.8450.8440.8320.3810.332
m7i40.8000.7910.7890.7790.3250.242
m7i60.7810.7660.7510.7400.2430.151
m7i80.8310.8050.8190.8020.3760.325
m7i100.8300.8160.8320.8150.3990.362
m7i120.8390.8020.8120.7940.2660.216
m9i40.7800.7670.7710.7620.3460.235
m9i60.7470.7230.7290.7140.3530.236
m9i80.8200.8030.8190.7930.3080.268
m9i100.7960.7760.8040.7800.3050.281
m9i120.8340.8040.8270.8020.2220.194
m11i40.7100.7000.6960.6850.1300.062
m11i60.7950.7780.7740.7460.1920.161
m11i80.8150.7970.8140.7880.3470.322
m11i100.7860.7490.7710.7410.3420.309
m11i120.8150.7860.8150.7840.2740.252
m13i40.7470.7350.7360.7250.2190.160
m13i60.8190.7930.7930.7670.2210.181
m13i80.7900.7690.7810.7550.2650.232
m13i100.8110.7740.7960.7710.3430.313
m13i120.8280.7880.8130.7830.2280.203
m15i40.7700.7530.7570.7420.1670.126
m15i60.7900.7610.7500.7310.1330.102
m15i80.8250.8000.8110.7800.2920.263
m15i100.7980.7720.7890.7590.2830.262
m15i120.8370.8140.8410.8000.2560.233
m17i40.7480.7260.7370.7180.1460.110
m17i60.7990.7740.7790.7590.2130.171
m17i80.7880.7570.7560.7350.2590.237
m17i100.7860.7620.7740.7500.3270.299
m17i120.8310.8030.8430.7830.2790.263
p-value0.0000.0100.0000.000
Table 9. Comparisons of NSGA-II, MOEA/D, and GLPEA (IGD).
Table 9. Comparisons of NSGA-II, MOEA/D, and GLPEA (IGD).
InstanceNSGA-IIMOEA/DGLPEA
Worst Best Mean Worst Best Mean Worst Best Mean
m3i40.0370.0240.0280.3970.2860.3160.0280.0220.026
m3i60.0630.0370.0430.5050.3590.4400.0470.0360.040
m3i80.0530.0410.0460.4200.3310.3880.0510.0410.045
m3i100.0510.0410.0470.4500.3280.3950.0500.0400.046
m3i120.0740.0420.0500.5700.4160.5080.0550.0400.049
m5i40.0490.0330.0400.6780.4720.6000.0440.0320.036
m5i60.0830.0450.0560.5850.4270.5220.0890.0350.054
m5i80.0530.0380.0440.4970.4000.4380.0490.0380.043
m5i100.0550.0390.0450.4960.3770.4300.0510.0370.043
m5i120.0580.0420.0480.5760.4550.5030.0540.0370.047
m7i40.0650.0430.0530.6460.4640.5710.0590.0420.048
m7i60.0550.0350.0430.6330.4960.5530.0500.0340.040
m7i80.0590.0410.0490.5310.4060.4660.0570.0410.049
m7i100.0630.0420.0490.4730.3690.4130.0590.0400.048
m7i120.0740.0460.0570.6160.4700.5520.0690.0430.055
m9i40.0590.0440.0510.6640.5020.5970.0550.0420.047
m9i60.0760.0410.0560.5700.4170.5190.0740.0390.054
m9i80.0620.0450.0510.6270.5010.5450.0570.0430.050
m9i100.0620.0440.0550.5690.3850.4480.0620.0420.052
m9i120.0760.0480.0600.5950.4630.5120.0740.0430.059
m11i40.1060.0520.0740.6780.5230.6050.0880.0420.055
m11i60.0840.0450.0620.5350.4130.4810.0810.0420.056
m11i80.0590.0410.0490.4310.3380.3910.0550.0390.045
m11i100.0690.0480.0550.4740.3600.4120.0660.0450.054
m11i120.0660.0490.0560.5070.3960.4510.0710.0430.055
m13i40.0890.0600.0730.7080.5110.6000.0830.0520.065
m13i60.0980.0390.0580.4640.3850.4330.0650.0370.054
m13i80.0620.0440.0530.4770.4090.4380.0620.0410.052
m13i100.0700.0450.0580.4430.3490.3930.0720.0430.055
m13i120.0880.0450.0610.5160.3960.4300.0860.0470.059
m15i40.0530.0410.0470.6550.4970.5860.0530.0350.043
m15i60.0940.0460.0650.5450.4140.4820.0960.0390.063
m15i80.0620.0440.0520.4900.4060.4470.0590.0440.050
m15i100.0700.0420.0570.4160.3260.3770.0670.0420.050
m15i120.0740.0450.0570.5470.4090.4800.0700.0420.052
m17i40.0870.0450.0590.7100.5390.6410.0830.0450.058
m17i60.0930.0610.0720.5900.4880.5210.0890.0470.064
m17i80.0670.0480.0550.4860.3810.4250.0650.0400.051
m17i100.0770.0510.0620.4560.3690.4130.0790.0470.057
m17i120.0910.0500.0660.5060.4320.4680.0800.0490.062
N B 871000353940
p-value0.0000.0000.0000.0000.0000.000
Note: For each instance, the superior value among the NSGA-II, MOEA/D, and GLPEA algorithms for the worst, best, and mean metrics is shown in bold. For example, for instance m3i4. Under the Worst case, the values for NSGA-II, MOEA/D, and GLPEA are 0.037, 0.397, and 0.028, respectively; 0.028 is superior. Under the Best case, the values are 0.024, 0.286, and 0.022, respectively; 0.022 is superior. Under the Mean case, the values are 0.028, 0.316, and 0.026, respectively; 0.026 is superior.
Table 10. Comparisons of NSGA-II, MOEA/D, and GLPEA (HV).
Table 10. Comparisons of NSGA-II, MOEA/D, and GLPEA (HV).
InstanceNSGA-IIMOEA/DGLPEA
Worst Best Mean Worst Best Mean Worst Best Mean
m3i40.7340.7480.7430.2930.5200.4400.7420.7530.747
m3i60.7770.7950.7870.1930.3670.2700.7830.8010.793
m3i80.8200.8450.8310.3070.4120.3540.8240.8480.832
m3i100.8100.8460.8330.3560.4540.4010.8190.8480.836
m3i120.8030.8410.8180.1780.3170.2540.7920.8490.820
m5i40.7790.7920.7870.1640.4010.2550.7820.7980.789
m5i60.7620.8040.7920.0980.3020.1940.7860.8170.805
m5i80.8070.8450.8270.2280.3480.2980.8080.8530.834
m5i100.8180.8550.8380.2270.3380.2810.8110.8600.844
m5i120.8210.8580.8430.2350.3410.2840.8380.8660.852
m7i40.7930.8140.8040.2130.2940.2480.8000.8230.815
m7i60.7680.7980.7820.1350.2240.1790.7790.8090.796
m7i80.7980.8410.8170.2300.3560.2780.8030.8330.819
m7i100.7970.8490.8180.2750.3570.3160.8000.8340.820
m7i120.7950.8400.8200.1860.2800.2420.7950.8570.825
m9i40.7660.7890.7780.1590.3200.2300.7710.7970.786
m9i60.7750.8110.7930.1170.2660.1900.7770.8190.802
m9i80.7730.8120.7930.1440.2280.1830.7730.8180.800
m9i100.7700.8150.7960.2160.3180.2740.7690.8310.801
m9i120.8160.8620.8380.2520.3070.2810.8200.8710.839
m11i40.7500.7690.7610.1110.2630.1820.7600.7790.772
m11i60.7290.7840.7650.0790.2330.1270.7710.8070.791
m11i80.7700.8120.7940.1940.3190.2380.7730.8380.800
m11i100.7320.7660.7520.2060.2810.2470.7130.7900.754
m11i120.7820.8310.8050.2180.3250.2690.7920.8300.812
m13i40.7430.7620.7550.1140.2040.1680.7400.7720.761
m13i60.7750.8120.7980.1680.2440.2070.7960.8390.815
m13i80.7500.8150.7810.1520.2410.2000.7570.8020.782
m13i100.7380.8110.7750.2180.2940.2520.7430.8200.786
m13i120.7670.8490.8140.2100.2680.2440.7990.8420.822
m15i40.7710.8070.7900.1300.2260.1890.7890.8140.800
m15i60.7380.7840.7610.0550.1650.1090.7500.8070.780
m15i80.7750.8200.7970.1640.2320.1970.7900.8370.811
m15i100.7400.8100.7720.2210.3190.2510.7490.8120.788
m15i120.7650.8590.8100.1640.2790.2110.7740.8450.823
m17i40.7250.7600.7380.0670.2280.1160.7350.7670.747
m17i60.7470.7940.7750.1010.1940.1450.7580.8090.785
m17i80.7310.7960.7630.1770.2830.2280.7340.8010.772
m17i100.7300.7790.7560.2260.3050.2670.7430.7840.760
m17i120.7700.8480.8080.1940.2880.2380.7750.8450.816
N B 770000353340
p-value0.0000.0000.0000.0000.0000.000
Note: For each instance, the superior value among the NSGA-II, MOEA/D, and GLPEA algorithms for the worst, best, and mean metrics is shown in bold. For example, for instance m3i4. Under the Worst case, the values for NSGA-II, MOEA/D, and GLPEA are 0.734, 0.293, and 0.742, respectively; 0.742 is superior. Under the Best case, the values are 0.748, 0.520, and 0.753, respectively; 0.753 is superior. Under the Mean case, the values are 0.743, 0.440, and 0.747, respectively; 0.747 is superior.
Table 11. The average execution time (in seconds) of the three algorithms.
Table 11. The average execution time (in seconds) of the three algorithms.
InstanceNSGA-IIMOEA/DGLPEA
m3i4779.7262000.1481283.671
m3i6794.3074657.9091330.478
m3i8828.9509622.0021377.809
m3i10792.81911,121.1781377.129
m3i12808.0407133.2461679.763
m5i4814.0612907.5411334.321
m5i6799.3672415.9971341.155
m5i8791.8598448.0231749.291
m5i10776.2018627.9501927.886
m5i12795.7556869.1321065.323
m7i4804.5284796.0241030.131
m7i6817.0144019.8911077.614
m7i8812.6546771.5981096.161
m7i10803.4606438.1961585.894
m7i12816.0468097.1052110.503
m9i4801.1675500.9981646.787
m9i6807.6053516.8912005.195
m9i8794.8487504.6072040.184
m9i10827.7519542.7242134.090
m9i12829.9507998.8942240.898
m11i4814.9275035.5242095.210
m11i6790.5283404.5472029.272
m11i8824.76110,279.5482074.627
m11i10820.2508103.2751128.435
m11i12836.55610,684.3742177.091
m13i4792.3004672.9101996.058
m13i6831.3164324.2662287.419
m13i8764.2657660.3142050.911
m13i10875.11811,367.9582316.989
m13i12840.69210,505.0992359.349
m15i4766.2185112.3261992.162
m15i6763.7244651.1582042.017
m15i8784.0777785.5382024.366
m15i10806.5799968.6312206.082
m15i12729.1058401.7492323.825
m17i4831.6562554.8791744.554
m17i6842.6031958.1532911.123
m17i8826.6419448.5042146.167
m17i10832.81310,989.4282377.331
m17i12849.4499877.9212873.884
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Luo, R.; Li, Z. Gaussian Learning-Based Pareto Evolutionary Algorithm for Parallel Machine Planning in Industrial Silicon Production. Mathematics 2025, 13, 3860. https://doi.org/10.3390/math13233860

AMA Style

Zhang J, Luo R, Li Z. Gaussian Learning-Based Pareto Evolutionary Algorithm for Parallel Machine Planning in Industrial Silicon Production. Mathematics. 2025; 13(23):3860. https://doi.org/10.3390/math13233860

Chicago/Turabian Style

Zhang, Jinsi, Rongjuan Luo, and Zuocheng Li. 2025. "Gaussian Learning-Based Pareto Evolutionary Algorithm for Parallel Machine Planning in Industrial Silicon Production" Mathematics 13, no. 23: 3860. https://doi.org/10.3390/math13233860

APA Style

Zhang, J., Luo, R., & Li, Z. (2025). Gaussian Learning-Based Pareto Evolutionary Algorithm for Parallel Machine Planning in Industrial Silicon Production. Mathematics, 13(23), 3860. https://doi.org/10.3390/math13233860

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop