1. Introduction
Dynamic Multi-objective Optimization Problems (DMOPs) have attracted the attention of scientists and developers for decades due to their common appearances in various fields, e.g., logistics [
1], disaster management [
2], nuclear physics [
3], and chemical reaction equipment management [
4]. Solving DMOPs is challenging work due to the uncertain change of the actual environmental chaos. The Dynamic Multi-Objective Evolutionary Algorithm (DMOEA) is one of the most efficient ways to overcome this challenge. A DMOEA employs a global search strategy of an evolutionary algorithm (EA) that is inspired by a natural or social law to solve DMOP instances (multi-objective optimization problems) in every steady environmental state (i.e., static multi-objective problems), and tunes searching directions to adapt to changed environments by some reaction/prediction approaches.
It is challenging to design a DMOEA for solving DMOPs. This is mainly because the real world can bring rapid changes to DMOPs, which requires the DMOEA to not only capture the change fast and accurately but also respond to the change in a reasonable way [
5]. Therefore, several works have studied responsive approaches for DMOEA to upgrade the search area (the population) based on the predicted movement of the optimal solution set (PS) or the Pareto optimal front (PF) when the environment (the current DMOP instance) is changed. But there are still some issues that existing related works have ignored or not addressed. First, the majority of DMOEAs assume the PS/PF is moving linearly, which falls short of reality. Second, most DMOEAs predict PS/PF movements by only mining the relationship for every single independent variable. These methods infer the new value of a variable from its historical values, and thus do not consider correlations between different variables, even though the phenomenon is common in the real world. Last but not least, to our best knowledge, no existing DMOEA considers the moving interconnectedness of different solutions of PS. Due to the above issues, the current related works can result in inaccurate predicted results, which may lead to bad population updates for the DMOP and degrade the performance of the solution solving.
To overcome the above issues, in this paper, we design a prediction approach based on linear regression (LR) technology to capture correlations between variables and between optimal solutions to accurately learn the movement pattern of PS/PF. To be specific, for the DMOP, by using historical values of all variables jointly, we establish an LR model where the variables in the current environment are the regressand and the variables in the last environment are the regressors. With the LR-based prediction model, we can calculate the variable values in the next environment by the current ones and update the current population based on the predicted variable values to adapt to the new environment. In brief, the contributions of this paper include the following three aspects:
We propose a PS/PF movement prediction model based on LR to learn correlations between variables and between optimal solutions, which predicts key points of PS in the new environment based on ones of the current PS.
We design a DMOEA based on the above LR-based prediction model, which updates the population based on the predicted key points of PS when the environment is changed and employs the Non-dominated Sorting Genetic Algorithm II (NSGA-II) to evolve the population when the environment is stable.
We conduct experiments based on CEC 2018 benchmark DMOPs. The results show that our DMOEA has the best overall performance in solving DMOP, compared with six other up-to-date DMOEAs.
The following contents are organized as follows. The second section illustrates the terminologies that are related to the DMOP and helpful for understanding this paper. The third section presents our proposed DMOEA based on the LR-based prediction model. The fourth section describes the experiment environment for the performance evaluation of the proposed DMOEA. The fifth section shows and analyzes the experiment results. The sixth discusses the existing works related to the DMOEA. The last section concludes the work of this paper.
2. Prerequisite Knowledge
Definition 1 (DMOP (Dynamic Multi-optimization Problem)). DMOP is the optimization problem consisting of a series of Multi-optimization Problems (MOPs) or DMOP instances that have multiple objectives. The DMOP instance is changed with the time/environment.
The DMOP can be formulated into Equation (
1).
is the objectives to be optimized at time
t.
represents the number of optimization objectives at
t. Both the objective functions and their number can be changed with the time or the environments. Equations (2) and (3) are equality and inequality constraints, respectively. The constraint functions,
and
, may be varied with time.
includes the independent variables/arguments of the DMOP (objective and constraint functions), with the domain of
. The domain and the number of arguments (
) are also changed with the time.
t represents the time that the environment is changed. At this moment, DMOP is changed from one MOP into another.
T is the number of environments that DMOP is concerned about. If the DMOP is concerned about the long-term optimization,
.
subject to
Definition 2 (Dominating). For an MOP or a DMOP instance, an available solution dominates another available solution if and only if the first solution does not have a worse value of all objectives than the latter one and has at least one better objective value.
It can be formulated as Equation (
6) that the available solution
dominates the available solution
, where
F represents the optimization objective functions of the MOP.
Definition 3 (Pareto optimal solution set (PS) and Pareto optimal fronts (PFs)). For an MOP or a DMOP instance, its PS is the set consisting of all available solutions (Pareto optimal solutions) that cannot be dominated by any other solutions. The PF is the set consisting of the objective function values of all Pareto optimal solutions.
PS and PF can be formulated as Equations (
7) and (
8), respectively, for the MOP with the domain
D and the objective function set
F.
The DMOEA is to find the
and
of all DMOP instances for the DMOP.
3. Linear Regression Prediction-Based DMOEA
In this section, to capture various environment change patterns and exploit the inter-correlations of
points (optimal solutions) of the DMOP, we design an LR-based method to predict positions of some points of
at the new environment based on their historical positions. In this paper, we consider that the position of each
point at the next time is decided by that of all
points in history, as exhibited in
Figure 1. Due to the high complexity of predicting all
points (the entire
), we are concerned about a few key points on
. To be specific, we set key points as 11 statistical points (
-th percentile,
) and one center point of the
, in this paper. Our proposed method is also compatible with other kinds of key points, e.g., knee points, that we will be concerned about for a better prediction. Below we detail our proposed LR-based method.
At the beginning of time
, we have historical
and
at all times before
t, and need to predict the new
and
at
. Because the manifold of
is ever-changing for DMOPs in the real world, we focus on the prediction of key points on the
in this paper, which is a powerful guide in updating the population for the DMOEA to adapt to the environmental changes and is the basic idea employed by almost all existing prediction approaches of DMOEAs. We use
(
) to represent the
-th percentile of the
at
t, and
(
) to represent the center of the
at
t, i.e.,
. By LR technology, we can establish the relational model of key points between two continuous environments, as Equation (
9).
(
) and
B are the parameters that will be learned by LR, where
is an
matrix and
B is a
n dimensional vector.
n represents the dimension of the search space of the concerned DMOP.
represents the estimated/predicted position of
.
The samples used for learning the LR model at time
include
,
, which can be easily achieved from the historical
s and
s, and where
and
correspond to
and
(
) in Equation (
9), respectively.
Based on the least-squares method, LR is to learn parameters
and
to minimize the square error (SE) between predicted values and real values for samples. Thus, learning the LR model is solving the minimization problem (
10) for the
k-th key point. This optimization problem can be solved by setting the partial derivatives for the parameters to 0. But in such a way, the derived prediction model is easily overfitting. Therefore, L2 regularization is applied to raise the ability to extend the model, and the optimization objective function (loss function) is transformed into Equation (
12).
is the regularization coefficient that is generally set to a number less than 0.01 [
6,
7].
Based on the loss function (
12), we can achieve the parameters (
(
) and
B) by the gradient descent method that updates these parameters step by step by iterative Formulas (13) and (14).
is the learning rate, which is tuned by experiments and is usually set between 0.1 and 0.001 [
8].
and
are the partial derivatives of the loss function with respect to the parameters, and can be calculated by Equations (15) and (16).
Now, based on the above formula, our proposed LR-based DMOEA can be illustrated by Algorithm 1. At first, the algorithm initializes a population and the environment (lines 1–2 in Algorithm 1). Then, the algorithm evaluates all objective function values for each individual (line 3 in Algorithm 1) and finds
and
, which consist of individuals that are not dominated by any other individual and their objective function values, respectively (line 4 in Algorithm 1). After this, the algorithm randomly initializes the parameters of the LR model, which will be updated during the DMOP solving process (line 5 in Algorithm 1).
After the above initialization, the algorithm begins to search for solutions for DMOPs by iteratively updating/evolving the population (lines 6–20 in Algorithm 1). There are mainly two kinds of population update methods used in our proposed DMOEA. When the environment is stable (line 15 in Algorithm 1), the population is updated by an evolutionary strategy. In this paper, we focus on the prediction method to adapt to the environmental changes of DMOPs, thus, we simply employ the evolutionary strategy of NSGA-II that is widely used for solving multi-objective problems (lines 15 in Algorithm 1) in this paper. In the future, we will exploit other evolutionary strategies with excellent cooperation with our proposed prediction method to improve the efficiency in solving DMOPs. When the environment is changed (lines 8–13 in Algorithm 1), the DMOEA re-initializes a population with the help of the LR-based prediction method to respond to the change, as described below.
Algorithm 1 LR-based DMOEA |
Input:
Objective functions F; domain/solution space D; |
Output: and ; |
- 1:
Randomly initialize a population; - 2:
Initialize the time or environment; - 3:
Evaluate the objective function values; - 4:
Find the current and ; - 5:
Randomly initialize parameters of LR model in Equation ( 9), , B; - 6:
while not reach termination condition do - 7:
Store the last key points, , ; - 8:
if environment is changed then - 9:
Update LR model with stored key points by decreasing gradient algorithm (Algorithm 2); - 10:
Predict new positions of key points by the LR model, ; - 11:
Build a sub-population with ; - 12:
Generate a sub-population with the idea of DNSGA-II-B - 13:
Update the population with random replacements - 14:
else - 15:
Evolve the population and update and by NSGA-II; - 16:
end if - 17:
Re-evaluate the objective function values - 18:
Update the current and - 19:
Go to the next time tick; - 20:
end while - 21:
return and ;
|
Within the environment of a DMOP, the DMOEA first updates parameters of the LR-based prediction model based on the stored locations of
key points (line 9 in Algorithm 1). The parameter update method for the LR model is outlined in Algorithm 2, and will be talked about later. Then, with the updated LR model, the DMOEA predicts the new locations of key points (
) from their last locations (
) by Equation (
9) (line 10 in Algorithm 1). Given the predicted new locations, we generate a set of new individuals with the quantity of a certain proportion of the population size, which constitutes a sub-population (line 11 in Algorithm 1). These new individuals are generated by introducing a Gaussian noise into the predicted key points in every dimension, as formulated by Equation (
17).
P represents the location of a newly generated individual.
is the Gaussian distribution with zero mean and the standard deviation of
.
By the introduction of Gaussian noises, the generalization of the prediction method can be improved, as the generated new individuals are more likely to be close to the
of the new environment, when the predicted locations of key points are near the new
instead of on it. This is a frequent occurrence because the prediction method is not perfect for precisely predicting the
movements in the real world.
Next, for improving the population diversity, we employ the population re-initialization method of DNSGA-II-B to generate another sub-population (line 12 in Algorithm 1). When the environment is changed, DNSGA-II-B randomly selects a certain percentage of individuals from the current population, and performs the mutation operator on them to generate new individuals.
Given the above two generated sub-populations, our proposed DMOEA re-initializes the population by randomly replacing part of its individuals with generated ones (line 13 in Algorithm 1).
For new individuals that are generated above by the evolutionary strategy, the prediction method, and the mutation operator of DNSGA-II-B, the proposed DMOEA evaluates their fitness function values (line 17 in Algorithm 1), and updates the current and (line 18 in Algorithm 1) as some individuals in may be dominated by new individuals and there may be new individuals that cannot be dominated by the current ’s members. After this, the proposed DMOEA enters the next tick for the next population re-initialization or evolution (line 19 in Algorithm 1).
The parameters of the LR model used for location predictions of key points are updated at the beginning of every environment change (line 9 in Algorithm 1). In this paper, we employ the decreasing gradient algorithm to learn the LR model based on the historical key points stored in line 7 of Algorithm 1. As illustrated in Algorithm 2, the process of the decreasing gradient algorithm for training the LR model is repeatedly updating the parameters of the LR model by the iterative Formulas (13)–(16) (line 3 in Algorithm 2). This process is repeated until the predefined maximum iteration time is reached or the parameters almost are not updated several times (line 1 in Algorithm 2).
Algorithm 2 LR model training by decreasing gradient algorithm |
Input:
Historical key points, , , ;
|
initial parameters, , , and B;
|
Output: Updated parameters, , B; |
- 1:
while not reach termination condition do - 2:
for ; do - 3:
Update and B by Equations (13)–(16) with ; - 4:
end for - 5:
end while - 6:
return, , and B;
|
4. Experiment Environment
For evaluating the effectiveness and efficiency of our proposed DMOEA, we conducted extensive experiments to compare the performance with six other DMOEAs on the CEC 2018 benchmark test suite [
9]. The six benchmark DMOEAs are as follows:
FL (Forward-Looking) [
10] is one of the simplest and most widely used methods for the environmental response of DMOPs. FL assumes that
moves with a uniform velocity and predicts that the new
is moving from the current
with the same step as the last movement (from the last
to the current
). For the
center point and every dimension, the moving velocity is
, and the new position of each
point is
.
FLV (Velocity Forward-Looking) assumes that moves with changeable velocity and the velocity is changed with a constant acceleration. Thus, the new position of a key point in a dimension is . is the last moving velocity of the center point. represents the last moving acceleration of the center point, which is calculated by .
KF (Kalman Filtering-based DMOEA) [
11,
12] employs the Kalman filtering technology to model the movement of the
center point.
SVR (Support Vector Regression-based DMOEA) [
13,
14] learns a regression model by support vector regression for the
center point in every dimension based on its historical locations, and predicts the new location by the learned regression model.
XGB (eXtreme Gradient Boosting-based DMOEA) [
15] is similar to SVR except that the regression model is learned by XGB, which is a classical and widely used ensemble learning with boosting technologies.
RF (Random Forest Regression-based DMOEA) is similar to the above two methods except that the regression model is learned by RF, which is an ensemble learning exploiting the bagging technology.
Due to our focus on the prediction method to adapt to the environmental changes for DMOPs, all of the above DMOEAs exploit NSGA-II as the evolutionary strategy when the environment is stable.
The CEC 2018 benchmark test suite includes 14 DMOPs with varied
and
movement patterns. The details of these 14 DMOPs are shown in
Table 1,
Table 2 and
Table 3. In our experiment, we set the environment change to severe by setting the frequency of change and the severity of change to 10 for each DMOP. The number of changes was 30. The number of variables was 10. All of these parameters of DMOPs were set referring to the report that proposes the CEC 2018 test suite [
9].
The performance metrics used for evaluating DMOEAs include Mean Inverted Generational Distance (MIGD) [
16,
17]. Mean Hypervolume Difference (MHVD) [
18], and Mean Average Crowding Distance (MACD). MIGD is the average of Inverted Generational Distance (IGD) in all environment cases for solving DMOPs. The calculation of IGD is conducted first by uniformly sampling points from the ideal
(
). Then, for each point on the
, it identifies the closest point on the solved
(obtained by the algorithm), sums all these minimum distances, and then computes their average. A smaller IGD value reflects better alignment between the algorithm’s results and
. In summary, MIGD can be calculated by Equations (18) and (19).
Due to the common unknown ideal
for real DMOPs, some works use MHVD as a quality metric to quantify the performance of DMOEAs. MHVD is the average value of HVD between
and solved
for all environments. HVD is the difference between the hypervolumes of
and the solved
. The hypervolume is the Lebesgue measure of the volume dominated by a set of points with respect to a reference point. The reference point can be set to the point that the value is the upper bound of the corresponding objective function in every dimension. Thus, the calculation of MHVD for a solved
is formulated as Equations (18) and (19), where
represents the hypervolume of a point set
S.
MACD is the time-average of the Average Crowding Distance (ACD) of PF, where the calculation of ACD can refer to [
19].
For each of these three above performance metrics, a smaller value indicates better overall algorithm performance in both convergence and distribution characteristics.
In this paper, all experiments were implemented by Python 3.12.2 and Numpy 1.26.4, and run on a personal computer with the Windows 11 HOME operating system, a 14th Gen Intel Core i7-14700(F) CPU, and 16GB of DDR5 memory. Each experiment was repeated 20 times, and the average result is reported below.
5. Experiment Results
Table 4 and
Table 5 present the values of various performance metrics achieved by different DMOEAs in solving 14 DMOPs of the CEC 2018 benchmark suite. As shown in the table, we can see that our proposed DMOEA (LR) achieves the best MIGD in six DMOPs, the best MHVD in five DMOPs, and the best MACD in seven DMOPs, and these three numbers are larger than other DMOEAs. This confirms the superior performance of our proposed method for solving DMOPs. The main advantage of our proposed DMOEA is twofold. First, our proposed DMOEA considers predicting the positions of multiple key points instead of only the center point for
. This can adapt to more generalized
shape changes. Secondly, we try to learn the relationship among different key points of
by LR, instead of only exploiting the correlation within one signal point (e.g., the center point). Thus, our proposed method achieves the best overall performance.
There are mainly two situations in which our proposed method achieves a relatively poor performance. The first one is when has almost a parallel motion overall, such as DF1. In such a case, DMOEAs that predict the based on only its center point may achieve better performance than our proposed method that employs the correlation of multiple key points. This is because correlation learning introduces more complex patterns of environmental changes, which is helpful for DMOPs with irregular changes, but may not accurately capture simple rules.
Another situation where our proposed method does not achieve the best performance is when the environment changes are too complex to predict. In this case, no prediction method achieves a good performance, while the simplest method may achieve the best performance. For example, FL achieves the best MIGD for DF11 and the best MHVD for DF14.
Next, we illustrate the stability of our proposed method based on the box plots of performance metric values achieved by various algorithms. To avoid overly lengthy content, we took the MIGD metric and DF1-DF3 as case studies, as shown in
Figure 2, for the following two reasons. First, the results are similar between different performance metrics. Second, these three cases include three scenes where the ranks of our algorithms are low, the best, and high, respectively. As we can see from
Figure 2, the interquartile range achieved by our algorithm has a similar rank to the average value in each case. Additionally, the box plots achieved by our algorithm have no outliers. Therefore, our algorithm has a stable performance in solving DMOPs.
6. Related Works
The basic idea of a DMOEA is predicting the change patterns of DMOP environments ( and ) and generating new individuals for population re-initialization to adapt to the new environment when it is changed. Therefore, there are several related works aiming at designing accurate prediction methods for DMOEAs.
Zhang et al. [
20] first divided the population into several sub-populations, and calculated the accelerations of the center points for these sub-populations. Then, they calculated the velocity for each sub-population based on its center point’s acceleration, and re-initialized a population by adding the velocity and each individual for every sub-population. Wang et al. [
21] first detected the intensity of the environmental change based on the difference between the objective function values of a small proportion of the population in two consecutive environments. Then, based on the detection result, they predicted new individuals by a multi-objective particle swarm optimization algorithm [
22] with adapted particle velocity updates and the strength Pareto evolutionary algorithm [
23]. Different from most works that employed the first-order difference strategy, where the trajectory of moving solutions is decreased by the centroid of
, Zhang et al. [
24] proposed to use the second-order difference method that considered both
and
. The new
center is decided by the joint trajectories of
and
by mapping them into a new two-dimensional space.
To overcome the limitation of most DMOEAs that only rely on predictions from adjacent environments, resulting in a new population adapting to the new environment not well, Zhang et al. [
25] used the Fuzzy C-Means clustering algorithm to cluster all historical PSs, and then divided the clusters into high-quality and low-quality sets based on their center points. With the cluster result, a Support Vector Machine (SVM) classifier was trained to predict whether a new solution is high quality. The re-initialized population consists of randomly generated individuals that are predicted to be high quality by the SVM classifier and Gaussian-mutated initial individuals. Ge et al. [
26] employed correlation alignment and a probabilistic annotation classifier to get high-quality solutions from the current population and randomly generated individuals in the new environment. Based on the dynamic change pattern of elite individuals, which is learned by correlation analysis and a denoising autoencoder, this work identified the parents of current elite solutions, and then used a denoising autoencoder to transfer the change pattern to new environments. This step predicted a set of individuals based on current elitist solutions and the last
. Individuals predicted by these two steps form the re-initialized population for the new environment.
In this paper, to make better predictions by capturing different moving styles of for the DMOP, we try to establish a relation model of different points on . This is the first attempt to introduce the inter-correlations between the movement of different points into the DMOEA, to our best knowledge.