Next Article in Journal
A Space Telescope Scheduling Approach Combining Observation Priority Coding with Problem Decomposition Strategies
Previous Article in Journal
Sodium Alginate–Starch Capsules for Enhanced Stability of Metformin in Simulated Gastrointestinal Fluids
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Objective Optimization Framework That Incorporates Interpretable CatBoost and Modified Slime Mould Algorithm to Resolve Boiler Combustion Optimization Problem

School of Information Engineering, Tianjin University of Commerce, Beichen, Tianjin 300134, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(11), 717; https://doi.org/10.3390/biomimetics9110717
Submission received: 8 October 2024 / Revised: 9 November 2024 / Accepted: 18 November 2024 / Published: 20 November 2024

Abstract

:
The combustion optimization problem of the circulation fluidized bed boiler is regarded as a difficult multi-objective optimization problem that requires simultaneously improving the boiler thermal efficiency and reducing the NOx emissions concentration. In order to solve the above-mentioned problem, a new multi-objective optimization framework that incorporates an interpretable CatBoost model and modified slime mould algorithm is proposed. Firstly, the interpretable CatBoost model combined with TreeSHAP is applied to model the boiler thermal efficiency and NOx emissions concentration. Simultaneously, data correlation analysis is conducted based on the established models. Finally, a kind of modified slime mould algorithm is proposed and used to optimize the adjustable operation parameters of one 330 MW circulation fluidized bed boiler. The experimental results show that the proposed framework can effectively improve the boiler thermal efficiency and reduce the NOx emissions concentration, where the average optimization ratio for thermal efficiency reaches +0.68%, the average optimization ratio for NOx emission concentration reaches −37.55%, and the average optimization time is 6.40 s. In addition, the superiority of the proposed method is demonstrated by ten benchmark testing functions and two constrained optimization problems. Therefore, the proposed framework is an effective artificial intelligence approach for the modeling and optimization of complex systems.

1. Introduction

1.1. Background

At present, thermal power generation still dominates the supply of electric energy, and coal is the main thermal power supply resource. However, as a kind of non-renewable energy source, coal is becoming more and more scarce with the massive consumption of thermal power generation. And the combustion process will be accompanied by large amounts of pollutants discharged into the atmosphere, such as NOx and SO2, which seriously affect the atmospheric environment and human living. Therefore, it is an urgent problem to improve boiler’s thermal efficiency and reduce NOx emissions for thermal power generation.
The above-mentioned problem is regarded as a multi-objective optimization problem, which is called a boiler combustion optimization problem, and it is also an important research direction [1]. It is widely known that the boiler combustion system has complex characteristics with non-linearity, strong coupling and large hysteresis, making it difficult to be optimized by traditional mathematical and statistical methods. To address this problem, many researchers commonly used artificial neural networks to build the thermal efficiency model or a NOx emissions model based on the boiler’s historical operation data [2,3,4,5]. And based on the established model, all kinds of optimization algorithms were applied to adjust the running parameters for improving the thermal efficiency and reducing NOx emissions. In addition, solving the optimization problem of the boiler combustion process is beneficial to the ecological modernization of power plants [6,7].
For modeling, Zhou et al. [8,9,10] applied artificial neural networks (ANNs) and support vector machine (SVM) variants to develop the prediction models for thermal efficiency and NOx emission concentration. The experimental results showed that ANNs or SVM variants can obtain good prediction results. These research results had an important impact on the subsequent research on the boiler combustion optimization problem. Si et al. [11] applied an improved SVM to achieve a dynamic adaptive model of NOx emission concentration, which provided a new idea for online modeling. Li et al. applied the extreme learning machine (ELM) and fast learning network (FLN) to model thermal efficiency and NOx emission [12,13,14]. The experimental results showed that the ELM and FLN can obtain higher accuracy and faster computing speed than ANNs and SVMs. Ma et al. [15] proposed a hyper-parameter self-optimized broad learning system to model the thermal efficiency and emission concentrations of Nox and SO2.
For optimization, researchers were keen on applying meta-heuristic optimization algorithms to solve the boiler combustion optimization problem [16,17,18]. Rahat et al. [19] applied a kind of novel multi-objective optimization algorithm to solve the contradiction between thermal efficiency and NOx emission concentration. Wang et al. [20] combined random forest (RF) with an improved flower pollination algorithm to build a NOx emission concentration model and reduce NOx emissions, separately. Hu et al. [21] used a deep neural network and multi-objective optimization algorithm to balance boiler thermal efficiency and NOx emission concentration. Three years ago, Xu et al. proposed several methods to solve the boiler combustion optimization problem, including the ICFAR-IPSO-LSTM-IMOPSO model [22], CIPSODM-LSTM algorithm framework [23], data-driven evolutionary optimization and online retrieval method [24]. In our group, Li and Ma et al. [25,26,27] used an improved artificial bee colony (ABC) and teaching and learning optimization (TLBO) to enhance the thermal efficiency and reduce the NOx emission concentration. In summary, artificial neural networks and meta-heuristic optimization algorithms can effectively solve the boiler combustion optimization problem.

1.2. Motivation

In this paper, a new multi-objective optimization framework is proposed to solve the boiler combustion optimization problem. The proposed framework combines the interpretable CatBoost model by TreeSHAP and a modified slime mould algorithm (MSMA). The interpretable CatBoost model is firstly used to build the synthetic model of boiler thermal efficiency and NOx emissions based on the historical operation data of one 330 MW circulation fluidized bed boiler (CFBB). Based on the established synthetic model, the modified slime mould algorithm is used to optimize the CFBB’s adjustment operation parameters for improved thermal efficiency and reduced NOx emissions. The experimental results show that the interpretable CatBoost model can obtain better model performance than RF [28,29], SVMs [30,31] and MLP [32,33,34], and the MSMA can improve boiler thermal efficiency and reduce NOx emissions.
Now, the new proposed framework needs to be explained. For the model method, CatBoost [35,36,37] is a kind of novel ensemble learning model that was first proposed by Yandex in 2017, and TreeSHAP [38,39] is an interpretable method for black box models that was proposed by Lundberg in 2017. Combining CatBoost with TreeSHAP can enhance its interpretability, which is called the interpretable CatBoost model in this paper. For the optimization algorithm, a slime mould algorithm (SMA) [40] is a kind of novel meta-heuristic algorithm that was first proposed by Lee et al. in 2020, which mainly simulates the changes of behavior and morphology of slime mould in nature during foraging. The SMA has several advantages, including a simple principle, few parameters and good global search capability. In order to further enhance the convergence accuracy and convergence speed of the original SMA, two mechanisms are proposed and introduced in the SMA. Firstly, iterative chaotic mapping with infinite collapse (ICMIC) [41] is used to initialize the population individuals for increasing population diversity and enhancing global search capability. Secondly, a reverse interpolation strategy (RIS) is proposed on the basis of opposition-based learning (OBL) [42,43] to update the positions of the population individuals obtained from each iteration, which helps to improve the quality of the solution and strengthen the local search ability. Compared with five other state-of-the-art meta-heuristic algorithms (including the butterfly optimization algorithm (BOA) [44], whale optimization algorithm (WOA) [45], seagull optimization algorithm (SOA) [46] and sparrow search algorithm (SSA) [47]) on 10 benchmark testing functions and two classical constrained optimization problems, the proposed MSMA outperforms on most testing functions. Moreover, compared to the original SMA, the MSMA is able to obtain better or equivalent convergence accuracy with less convergence time without increasing the calculation burden.

1.3. Contributions

In order to solve the boiler combustion optimization problem, a new multi-objective optimization framework is firstly designed in this paper. The contributions of this paper are summarized as follows:
  • A kind of novel multi-objective optimization framework is proposed, which combines interpretable CatBoost by TreeSHAP and a modified slime mould algorithm. Particularly, this framework has model interpretability and adaptivity.
  • A kind of modified slime mould algorithm is proposed to solve benchmark testing functions and constrained optimization problems.
The structure of this paper is as follows: the study object and basic methods are introduced in Section 2; the modified SMA is presented in Section 3; Section 4 contains the application of the proposed framework to the boiler combustion optimization problem; and the conclusions are presented in Section 5.

2. Related Work

2.1. Study Object

The study object of this paper is a 330 MW circulating fluidized bed boiler (CFBB). The main study focus of this paper is to maximize the thermal efficiency (TE) while minimizing the NOx emissions concentration of the CFBB. To achieve the optimization goal, a kind of new multi-objective optimization framework integrating the interpretable CatBoost model and MSMA is proposed in this paper. The flow chart of the proposed framework is shown in Figure 1, and the core steps are summarized as follows:
Step 1. The interpretable CatBoost model combined with TreeSHAP is used to build the synthetic model of boiler thermal efficiency and NOx emissions based on historical operation data. Note that the detailed descriptions of the interpretable CatBoost and TreeSHAP are given in the literature [35,36,37,38,39], separately.
Step 2. Based on the established model in step 1, the value ranges of adjustable parameters of the CFBB with positive impact on thermal efficiency and negative impact on NOx emissions are obtained by further analyzing the features affecting the thermal efficiency and NOx emission concentration.
Step 3. Based on the established model and selected parameter features, the boiler’s adjustment parameters are optimized by a modified SMA with the fitness function of maximizing the thermal efficiency and minimizing the NOx emission concentration of the CFBB. And the constraints for the value ranges of the adjustable parameters of the CFBB are obtained in step 2.

2.2. Slime Mould Algorithm (SMA)

In this subsection, a detailed descriptions of the slime mould algorithm is given. The specific content of the SMA is explained below. In order to understand the following method, the symbols used in this section are explained in Table 1.
The slime mould algorithm (SMA) is a kind of novel meta-heuristic algorithm that was first proposed by Lee et al. in 2020, which is inspired by foraging behavior and the morphology of slime mould. Slime mould is a eukaryotic organism that lives in a humid and cold environment and takes its nutrients mainly from external organic matter. When slime mould approaches its food source, its bio-oscillator generates a propagation wave through the vein to increase the flow of cytoplasm. The higher the concentration of food, the stronger the propagation wave generated by the bio-oscillator and the faster the cytoplasmic flow. The SMA has a simple principle, few adjustment parameters and strong global search capability. Therefore, the SMA is used to solve the boiler combustion optimization problem. Its basic principle is as follows.
(1) The stage of approaching food
In the approaching food phase, the slime mould approaches the food through the odor in the air, and the expression of its approaching behavior is shown in Equation (1).
X ( t + 1 ) = X b e s t ( t ) + vb W × X A t X B t ,   r < q v c × X ( t ) ,   r q
where X ( t + 1 ) and X ( t ) are the positions of the slime mould individuals at the t + 1 th and t th iterations, respectively; X b e s t ( t ) is the best position of the slime mould individual at the t th iteration; X A t and X B t are the two randomly selected slime mould individuals at the t th iteration; the range of v b is [ a , a ] , a = a r c t a n h ( 1 ( t / P ) ) , t is the current number of iterations and P is the total number of iterations; the range of v c is from 1 linearly decreasing to 0; r is a random value between 0 and 1; q = t a n h ( f ( X i ) S ) , i = 1,2 , , Q , f ( X i ) is the fitness value of the i th slime mould individual; S is the optimal fitness value in all iterations; Q is the population size of the slime mould individuals; and W is the weight of a slime mould individual, and its expression is shown in Equation (2).
W ( SortIndex   ( X i ) ) = 1 + r × log f b e s t f ( X i ) f b e s t f w o r s t + 1 ,   c o n d i t i o n 1 r × log f b e s t f ( X i ) f b e s t f w o r s t + 1 ,   o t h e r
where condition is the top half of the population individual in terms of fitness value; f b e s t and f w o r s t are the best and worst fitness value at the current iteration, respectively; S o r t I n d e x   ( X i ) is the sequence of sorted fitness values; log ( ) is used to slow down the rate of change in the values and stabilize the change in the contraction frequency; c ondition and o t h e r are used to explain the process of slime mould adjusting its position according to the food concentration, where the higher the food concentration, the greater the weight of slime mould near the area; if the food concentration is low, slime mould will turn to search for other areas, and the weight of slime mould in the area will become smaller. The fitness value evaluation graph of slime mould is shown in Figure 2.
The approaching food stage simulates the positive and negative feedback process between the weight of slime mould and the food concentration. Equation (10) updates the individual positions according to the change in the optimal position X b e s t and the fine tuning of v b , v c and W . The role of r is to form a search vector at any angle, i.e., to search the solution space in any direction, thus increasing the possibility of finding the optimal solution and better simulating the circular and fan-shaped structural motion of the slime mould as it approaches the food.
(2) The stage of enveloping food
The enveloping food phase simulates the contraction pattern of the slime mould venous tissue. The higher the concentration of food in contact with the venous tissue, the stronger the propagation wave generated by the bio-oscillator and the faster the cytoplasmic flow; the lower the concentration of food in contact with the venous tissue, the weaker the propagating waves generated by the biological oscillator and the slower the cytoplasmic flow. Based on the above principle, the expression of its enveloping behavior is shown in Equation (3).
X ( t + 1 ) = d × ( X u b X l b ) + X l b ,   r < u   X b e s t ( t ) + v b W × X A t X B t ,   r < q v c × X ( t ) ,   r q
where d is the random value generated within 0 to 1; X u b and X l b are the upper and lower bounds of the search space, respectively; and u is the proportion of randomly distributed slime mould individuals, and it is used to switch between the global search phase and the local search phase.
(3) The stage of grabbing food
During the food grasping phase, the slime mould uses the propagation waves generated by the bio-oscillator to change the velocity of cytoplasmic flow in the venous tissue, simulating the variation of the width of the venous tissue of slime mould with the variation of the oscillation frequency of the bio-oscillator by v b , v c and W . The slime mould approaches the food more slowly when the food concentration is low and more quickly when it finds good quality food. This stage is actually the update of the v b , v c and W .

3. Modified Slime Mould Algorithm (MSMA)

In order to further improve the convergence accuracy and convergence speed of the original SMA, a kind of modified slime mould algorithm (MSMA) is proposed by introducing two mechanisms. The detailed description of the two mechanisms is given in Section 3.1. In addition, ten benchmark testing functions and two classical constraint optimization problems are used to verify the performance of the MSMA in Section 3.2.

3.1. Improved Mechanism

(1) Iterative chaotic mapping with infinite collapse (ICMIC)
In swarm intelligence optimization algorithms, the quality of the initial population individuals has a large impact on the algorithm performance. A good initial population is conducive to obtaining a global optimal solution. The original SMA uses a random initialization method to initialize the population without a priori knowledge, which may easily lead to the population diversity missing. Logistic and tent chaotic mapping functions are commonly utilized to initialize population individuals for enhancing the population diversity. However, compared with these two chaotic mapping methods, the ICMIC chaotic mapping method has better uniform traversal and faster convergence. Therefore, the ICMIC chaos mapping method is introduced to generate population individuals.
The expression of the ICMIC chaotic mapping is shown in Equation (4).
Z i + 1 = sin β π z i ,   β ( 0 , + ) 1 Z i 1 ,   Z i 0
where β is a random value that follows the normal distribution in the interval ( 0 , + ) .
Next, the chaotic sequence individual is converted to the real solution space, and the conversion formula is shown in Equation (5).
X i = X l b + X u b X l b 1 + Z i 2
where x u b and x l b are the upper and lower bounds of the search space, respectively; and Z i is the generated chaotic sequence.
(2) Reverse interpolation strategy (RIS)
In order to further enhance the local exploration capability of the original SMA, a reverse interpolation strategy (RIS) is proposed, which combines elite opposition-based learning and quadratic interpolation. The RIS makes it easy for population individuals to jump out of local optimal solutions.
Opposition-based learning (OBL) is an optimization mechanism that was first proposed by Tizhoosh in 2005 whose main principle is to compute and evaluate the inverse solution simultaneously for the current feasible solution and select the better solution as the next generation individual. Elite opposition-based learning (EOBL) expands the search space based on the former by exploiting the fact that elite individuals include richer effective information than general individuals.
Assuming that the elite individual of the slime mould population at the t th iterations (i.e., the current optimal solution) is X b e s t = x i 1 , x i 2 , , x i j , , x i M , where M is the dimension of the search space, the expression of its elite opposition solution X = x i 1 , x i 2 , , x i j , , x i M is shown in Equation (6).
X = θ X l b + X u b X b e s t
where θ is the random value that follows the normal distribution in the interval [0, 1]; X u b and X l b are the upper and lower bounds of the search space, respectively.
When the generated elite inverse solution exceeds the maximum boundary range of the search space, the out-of-bounds reset is performed using the random generation method, whose expression is shown in Equation (7).
X = r a n d X l b + X u b
In order to further enhance the utility of EOBL and improve the global search capability of the algorithm, the quadratic interpolation method is introduced to further update the elite solutions obtained by EOBL.
Assuming that X A and X B are the two randomly selected slime mould individuals at the t th iteration and different from X A and X B , then the fitness values of X A , X B and elite opposition solution X are f ( X A ) , f ( X B ) and f ( X ) . The expression of the solution after the quadratic interpolation X = x i 1 , x i 2 , , x i j , , x i D is shown in Equation (8).
X = X 2 X B 2 × f ( X A ) + X A 2 X 2 × f ( X B ) + X B 2 X A 2 × f ( X ) 2 [ X X B × f ( X A ) + X A X × f ( X B ) + X B X A × f ( X ) ]
When the generated solution after quadratic interpolation is not as good as the elite inverse solution, it is selected according to the expression shown in Equation (9).
X = X ,   f ( X ) < f ( X ) X .   f ( X ) f ( X )
(3) Overview of MSMA
The flow chart of the MSMA is shown in Figure 3, and the specific steps are summarized as follows.
Step 1: Determine the population size, the maximum number of iterations and the fitness function.
Step 2: Generate chaotic initialized population individuals according to Equations (4) and (5) and calculate the fitness values.
Step 3: Calculate the weight W and the parameter a according to Equations (1) and (2).
Step 4: Generate random number r , and then judge the size of random value r and parameter u ; if r < u , then update the individual position according to the first formula of Equation (3); otherwise, update parameters q , v b , v c . Judge the size of r and parameter q ; if r < q , then update the individual position according to the second formula of Equation (3); otherwise, update the individual position according to the third formula of Equation (3).
Step 5: Calculate the fitness values to obtain the current optimal solution.
Step 6: Update the global optimal solution according to Equations (6)–(9).
Step 7: Determine whether the end condition is satisfied, and if so, output the global optimal solution and the fitness value; otherwise, repeat steps 2–6.

3.2. Experiments

In order to demonstrate the performance of the MSMA, ten benchmark testing functions are chosen from the CEC-2005 functions, and two classical constrained optimization problems are employed in this subsection. All the experiments including those in Section 4 are implemented on 11th Gen Intel(R) Core (TM) i7-11800H @ 2.30 GHz 2.30 GHz Processor and 32 GB RAM. All algorithms are coded and carried out in Python version 3.9 and Windows 11 Professional.

3.2.1. Benchmark Testing Functions

The ten benchmark testing functions are shown in Table 2, where F1–F5 are single-peak testing functions and F6–F10 are multi-peak testing functions. The single-peak testing function has only one global optimal solution, which is mainly used to test the convergence speed and exploitation capability of the optimization algorithm. The multi-peak testing function has not only one global optimal solution, but also several local optimal solutions, which are mainly used to test the ability to jump out of the local optimal solution and the exploration capability of the optimization algorithm. In addition, five meta-heuristic algorithms are used to compare with the MSMA, including the original SMA [40], butterfly optimization algorithm (BOA) [44], whale optimization algorithm (WOA) [45], seagull optimization algorithm (SOA) [46] and sparrow search algorithm (SSA) [47]. According to references [40,44,45,46,47], parameters of the five algorithms are given in Table 3. For each testing function, the optimization algorithm runs independently 30 times. Three metrics are used to test the performance of the MSMA and other algorithms in terms of the mean, standard deviation (S.D.) and running time. The mean is the average of thirty optimum fitness values, and the running time is the average running time of thirty experiments. The population size of every algorithm is 50, and the maximum iteration number is 500. These testing functions are set to 30, 50 and 100 dimensions, separately. The experimental results are denoted from Table 4, Table 5 and Table 6.
The best optimal solutions in terms of the mean, standard deviation and running time are shown in boldface. And it is noted that the smaller the value of these statistics metrics, the better the performance of the optimization algorithm. In addition, the convergence curves of the MSMA and other algorithms are shown in Figure 4.
As can be seen from Table 4, Table 5 and Table 6 and Figure 4, the MSMA has the lowest mean and lowest standard deviation of all the methods on functions F1, F2, F3, F4, F9 and F10. The SSA has the smallest mean and standard deviation for functions F5 and F6. The BOA has the smallest mean and standard deviation for function F7. The SOA has the smallest mean and standard deviation for function F8. And the SOA has the shortest calculation time of all the algorithms for all functions. For the functions F5, F6, F7 and F8, although the MSMA does not obtain the best model performance, it outperforms the original SMA in both mean and standard deviation.

3.2.2. Constrained Optimization Problem

(1) Optimal design of three-bar truss problem
In the optimal design of the three-bar truss problem, the variables x 1 , x 2 and x 3 are the cross-sectional areas of the three bars, respectively, and by symmetry x 1 = x 3 . And the objective of the optimal design of the three-bar truss is to minimize the volume of the three-bar truss by adjusting the cross-sectional area ( x 1 , x 2 ) . This optimal design problem has a nonlinear fitness function, three nonlinear inequality constraints and two continuous decision variables, as shown in Equations (10)–(15).
Fitness function:
m i n f ( x ) = 2 2 x 1 + x 2 100
Constraints:
0.001 x 1 1
0.001 x 2 1
2 x 1 + x 2 2 x 1 2 + 2 x 1 x 2 2 2 0
x 2 2 x 1 2 + 2 x 1 x 2 2 2 0
1 2 x 2 + x 1 2 2 0
The optimal solutions of the MSMA and the other algorithms used for comparison on the optimal design of the three-bar truss problem are shown in Table 7, and the convergence curve of six algorithms on this optimal design problem is shown in Figure 5.
As can be seen from Table 7 and Figure 5, the MSMA has the lowest mean and standard deviation of all the methods, and the WOA has the shortest running time.
(2) Optimal design of tension–compression spring problem
In the optimal design of the tension–compression spring problem, the variables x 1 , x 2 and x 3 are the spring coil diameter, spring coil diameter and number of coils wound, respectively. The objective of the optimal design of the tension–compression spring is to minimize the weight of the tension–compression spring while meeting the minimum deflection, vibration frequency and shear stress. This optimal design problem has a nonlinear fitness function, four nonlinear inequality constraints and three continuous decision variables, as shown in Equations (16)–(23).
Fitness function:
m i n f ( x ) = x 3 + 2 x 2 x 1 2
Constraints:
0.05 x 1 2
0.25 x 2 1.3
2 x 3 15
1 x 2 3 x 3 71785 x 1 4 0
4 x 2 2 x 1 x 2 12566 x 2 x 1 3 x 1 4 + 1 5108 x 1 2 1 0
1 140.45 x 1 x 2 2 x 3 0
x 1 + x 2 1.5 1 0
The optimal solutions of the MSMA and the other algorithms used for comparison on the optimal design of the tension–compression spring problem are shown in Table 8. And the convergence curves of six algorithms on this optimal design problem are shown in Figure 6.
As can be seen from Table 8 and Figure 6, the WOA has the lowest mean, lowest standard deviation and shortest calculation time of all the methods for this optimal design problem. Although the MSMA did not obtain the best model performance for this optimal design problem, it outperforms the original SMA on both the mean and standard deviation.

3.3. Discussion and Analysis

By solving the ten benchmark testing functions and two constrained optimization problems, the experimental results reveal that the MSMA is an effective meta-heuristic algorithm. The MSMA is good at handling the single-peaked function extremum problems, because it can obtain the optimal solution on four single-peaked functions F1, F2, F3 and F4. The MSMA needs to be further enhanced in handling multi-peaked function extremum problems because it can only obtain the optimum solution on two multi-peaked functions, F9 and F10. Although the MSMA did not obtain the optimal solution on functions F5, F6, F7, and F8, it outperformed the original SMA for all four functions. In addition, the MSMA also showed good performance on two constrained optimization problems.
The time complexity of calculating the SMA depends on the following aspects: population initialization of slime mould, evaluation and ranking of fitness values, weight update, and position update of slime mould individuals. Since the slime mould population size is Q , the research problem dimension is M , and the maximum number of iterations is P , the time complexity of the original SMA is O ( M + P × Q × ( 1 + l o g Q + M ) according to the description of the SMA and the operation rules of time complexity. For the MSMA, the time complexity of ICMIC chaotic mapping to initialize the slime mould population is O ( D ) , the time complexity of the fitness evaluation and ranking is O ( Q + Q × l o g Q ) , the time complexity of the weight update is O ( Q × M ) , the time complexity of the position update of slime mould individuals is O ( Q × M ) , and the reverse interpolation strategy to generate new positions is O ( Q × M ) ; then, the overall time complexity of the MSMA is O ( M + P × Q × ( 1 + l o g Q + M ) . The time complexity of the MSMA is the same as that of the original SMA, indicating that the improved algorithm does not increase the computational burden.
Overall, the MSMA is more effective than the original SMA, and it can obtain better or equivalent convergence accuracy with fewer convergence times without adding much calculation time.

4. Boiler Combustion Optimization

In this section, the proposed multi-objective optimization framework is used to solve the boiler combustion optimization problem. Firstly, the description of boiler operation data is given in Section 4.1. Secondly, the interpretable CatBoost is applied to model the boiler thermal efficiency and NOx emissions concentration in Section 4.2. Based on the established model, data correlation analysis is conducted, determining the range of the parameters to be optimized. In Section 4.4, the proposed MSMA is used to optimize these adjustment parameters for improving thermal efficiency and reducing NOx emission concentration at the same time.

4.1. The Source and Description of Boiler Data

There are mainly 26 input data samples that affect the thermal efficiency (TE) and nitrogen oxide (NOx) emission concentration of the CFBB, including the boiler load, coal feed rate, average bed temperature, air flow, air temperature, etc. The symbols and descriptions of each sample are shown in Table 9. A total of 2880 sets of data were collected from the DCS of the CFBB. Some of these data are shown in Table 10.

4.2. Modeled by Interpretable CatBoost

In this subsection, the interpretable CatBoost model is used to build a comprehensive model of the boiler thermal efficiency and NOx emissions concentration. In order to evaluate the model superiority of the interpretable CatBoost model, three state-of-the-art methods are employed to compare the model accuracy and generalization capability, including random forest (RF), support vector machine (SVM), and multi-layer perceptron (MLP). The parameter settings of each method are given in Table 11. Four metrics are used to verify the performance of each method, such as root mean square error (RMSE), mean absolute error (MAE), mean absolute percentage error (MAPE) and goodness of fit (R2) of ten-fold cross-validation. For training data, the smaller the parameters (RMSE, MAE and MAPE) and the closer the R2 value is to one, the higher the model accuracy. For testing data, the smaller the parameters (RMSE, MAE and MAPE) and the closer the R2 value is to one, the better the generalization capability.
The data shown in Table 10 were normalized. Then, the total 2880 sets of data were divided into training data and testing data in a ratio of 7:3 according to the rule of ten-fold cross-validation. The experimental results are shown in Table 12.
As shown in Table 12, compared with RF, SVM and MLP, the interpretable CatBoost model can obtain better model accuracy and adequate generalization capability in the feature models established for the thermal efficiency and NOx emissions concentration of the CFBB both on the training set and testing set.
The fitting diagrams and error diagrams of the interpretable CatBoost model for modeling the two objectives of the CFBB on the partial testing set are shown in Figure 7 and Figure 8, where the red line and the blue line in the fitting diagram indicate the predicted values and real values of the partial testing set, and the light blue line in the error diagram represents the error of the predicted values compared to the real values for the partial testing set.
As shown in Figure 7 and Figure 8, the interpretable CatBoost model can effectively establish the feature models for the two objectives of the CFBB. The two feature models all have a great fitting effect and small fitting error on the testing set.

4.3. Data Feature Analysis

In response to the weak interpretation of CatBoost, an interpretable method combined TreeSHAP is applied to investigate the effect of the combustion condition of the CFBB on the thermal efficiency and NOx emission concentration.
The feature analysis importance diagram and summary diagram of thermal efficiency are shown in Figure 9.
The feature analysis dependency diagram of thermal efficiency is shown in Figure 10.
The feature analysis importance diagram and summary diagram of NOx emission concentration is shown in Figure 11.
The feature analysis dependency diagram of thermal efficiency is shown in Figure 12.
In the feature analysis importance diagrams of Figure 9 and Figure 11, the horizontal coordinate represents the Shapley value. The larger the Shapley value, the greater the feature importance. As can be seen in Figure 9, the top five features that have the greatest impact on the thermal efficiency of the CFBB are 6CEMSO2, 6CEMSTEMP, 08A061, 06T401 and 08A051. As can be seen in Figure 11, the top five features that have the greatest effect on the NOx emission concentration of the CFBB are 6CEMSO2, 17I011, 06F061, 6CEMSTEMP and 17I021.
In the feature analysis summary diagrams of Figure 9 and Figure 11, each feature point represents the corresponding feature sample, the change in color from blue to red indicates the change in the value of the feature sample from small to large, and the positive and negative Shapley values indicate the positive and negative effects of the feature sample with the target. In order to explore more visually the positive and negative effects of different values of each feature on the target, the feature analysis dependency diagrams shown in Figure 10 and Figure 12 are more important.
Taking the first subgraph in Figure 10 and Figure 12 as an example, for the feature AFCOALQ, its original value range is [27.67, 68.12]. When it has a positive impact on the thermal efficiency of the CFBB, its value range is [27.67, 60.40], and when it has a negative impact on the NOx emission concentration of the CFBB, its value range is [27.67, 63.86], so it can be obtained that the value range when its impact on thermal efficiency is positive and its impact on NOx emission concentration is negative is [27.67, 56.90]. And the value ranges of the other adjustable parameters of the combustion system of the CFBB are shown in Table 13.

4.4. Parameters Optimized by MSMA

In this subsection, the MSMA is applied to optimize the boiler thermal efficiency and NOx emissions concentration based on the established models. The objective function is shown in Equation (24), where X is a vector containing 13 adjustable operation parameters, i.e., AFCOALQ ( x 1 ), BFCOALQ ( x 2 ), CFCOALQ ( x 3 ), DFCOALQ ( x 4 ), 05F051 ( x 5 ), 05F061 ( x 6 ), 06F051 ( x 7 ), 06F052 ( x 8 ), 06F061 ( x 9 ), 06F062 ( x 10 ), 6CEMSO2 ( x 11 ), 08A051 ( x 12 ) and 08A061 ( x 13 ). f T E ( X ) and f N O x ( X ) are the function value of boiler thermal efficiency and NOx emission concentration, respectively. The constraints are shown in Equations (26)–(38), i.e., the range of adjustable operation parameters.
Objective function:
m i n   f X = f N O x X f T E ( X )
X = [ x 1 , x 2 , x 3 , x 4 , x 5 , x 6 , x 7 , x 8 , x 9 , x 10 , x 11 , x 12 , x 13 ]
Constraints:
27.67 x 1 56.90
33.03 x 2 51.09
42.49 x 3 61.93
33.92 x 4 58.11
151.97 x 5 456.13
91.78 x 6 448.35
15.05 x 7 762.78
121.21 x 8 56.63
320.80 x 9 1205.40
144.19 x 10 547.852
3.15 x 11 5.14
0.74 x 12 1.07
0.01 x 13 0.20
The population size and the maximum number of iterations of the MSMA are set to 40 and 50, respectively. Overall, 10% of the operation data shown in Table 10 were randomly selected to be optimized. The partial optimization results of the MSMA are recorded in Table 13. And the optimization curves of the MSMA for boiler thermal efficiency and NOx emission concentration are shown in Figure 13.
As can be seen in Table 14 and Figure 13, the MSMA can effectively optimize the combustion system of the CFBB, which is able to reduce the NOx emission concentration while ensuring a high thermal efficiency. Among them, the average optimization ratio for thermal efficiency reaches +0.68%, the average optimization ratio for NOx emission concentration reaches −37.55%, and the average optimization time is 6.40 s, which is much less than the time for each interval sampling, so the actual industrial demand can be met.

5. Conclusions

This paper designs a new multi-objective optimization framework to solve the boiler combustion optimization problem. The proposed framework has several functions, such as modeling, data correlation analysis and optimization. Actually, the framework integrates the interpretable CatBoost model and a modified slime mould algorithm. The framework is used to improve the thermal efficiency and reduce the NOx emissions of one 330 MW CFBB. The experimental results show that the proposed framework can on average improve the boiler thermal efficiency by 0.68% while reducing the NOx emission concentration by 37.55%, and the optimized adjustment operation parameters can meet the actual industrial demand. In addition, a kind of modified slime mould algorithm is proposed to improve the convergence accuracy of the original SMA. Through the benchmark functions test, the MSMA outperforms the original SMA.
In future research, we are planning to use the proposed framework to solve modeling and optimization problems for different types of complex systems. Moreover, we will further enhance the convergence accuracy of the MSMA through several mechanisms and design more efficient multi-objective algorithms based on the MSMA.

Author Contributions

Conceptualization, S.G. and Y.M.; methodology, S.G.; software, S.G.; validation, Y.M.; formal analysis, S.G.; investigation, Y.M.; resources, Y.M.; data curation, Y.M.; writing—original draft preparation, S.G. and Y.M.; writing—review and editing, S.G. and Y.M.; visualization, S.G.; supervision, Y.M.; project administration, Y.M.; funding acquisition, Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (Grant No. 62203332) and the Natural Science Foundation of Tianjin (Grant No. 20JCQNJC00430).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

This manuscript does not report data generation or analysis.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhou, H. Study on Some Key Issues in NOx Control and Combustion Optimization of Large Power Plant Boilers. Doctoral Dissertation, Zhejiang University, Hangzhou, China, 2004. [Google Scholar]
  2. Zhou, H.; Zhao, J.; Zheng, L.; Wang, C.; Cen, K. Modeling NOx emissions from coal-fired utility boilers using support vector regression with ant colony optimization. Eng. Appl. Artif. Intell. 2012, 25, 147–158. [Google Scholar] [CrossRef]
  3. Zhou, H.; Cen, K. Combining Neural Network or Support Vector Machine with Optimization Algorithms to Optimize the Combustion. In Combustion Optimization Based on Computational Intelligence; Advanced Topics in Science and Technology in China; Springer: Singapore, 2018. [Google Scholar] [CrossRef]
  4. Yao, Z.; Romero, C.; Baltrusaitis, J. Combustion optimization of a coal-fired power plant boiler using artificial intelligence neural networks. Fuel 2023, 344, 128145. [Google Scholar] [CrossRef]
  5. Li, G.; Niu, P.; Wang, H.; Liu, Y. Least Square Fast Learning Network for modeling the combustion efficiency of a 300 WM coal-fired boiler. Neural Netw. Off. J. Int. Neural Netw. Soc. 2014, 51, 57–66. [Google Scholar]
  6. Akopov, A.S.; Beklaryan, L.A.; Saghatelyan, A.K. Agent-based modelling for ecological economics: A case study of the Republic of Armenia. Ecol. Model. 2017, 346, 99–118. [Google Scholar] [CrossRef]
  7. Cappelletti, F.; Manuguerra, L.; Rossi, M.; Germani, M.; Sartini, M. Design optimization for environmental sustainability: A framework to select the best eco-design strategy. Proc. Des. Soc. 2023, 3, 1297–1306. [Google Scholar] [CrossRef]
  8. Zheng, L.; Zhou, H.; Cen, K.; Wang, C. A comparative study of optimization algorithms for low NOx combustion modification at a coal-fired utility boiler. Expert Syst. Appl. 2009, 36, 2780–2793. [Google Scholar] [CrossRef]
  9. Wu, F.; Zhou, H.; Ren, T.; Zheng, L.; Cen, K. Combining support vector regression and cellular genetic algorithm for multi-objective optimization of coal-fired utility boilers. Fuel 2009, 88, 1864–1870. [Google Scholar] [CrossRef]
  10. Zhou, H.; Zheng, L.; Cen, K. Computational intelligence approach for NOx emissions minimization in a coal-fired utility boiler. Energy Convers. Manag. 2010, 51, 580–586. [Google Scholar] [CrossRef]
  11. Si, F.; Romero, C.E.; Yao, Z.; Xu, Z.; Morey, R.L.; Liebowitz, B.N. A new approach for function approximation in boiler combustion optimization based on modified structural AOSVR. Expert Syst. Appl. 2009, 36, 8691–8704. [Google Scholar] [CrossRef]
  12. Li, G.; Niu, P.; Duan, X.; Zhang, X. Fast learning network: A kind of novel artificial neural network with a fast learning speed. Neural Comput. Appl. 2014, 24, 1683–1695. [Google Scholar] [CrossRef]
  13. Niu, P.; Ma, Y.; Li, G. Model NOx emission and thermal efficiency of CFBB based on an ameliorated extreme learning machine. Soft Comput. 2018, 22, 4685–4701. [Google Scholar] [CrossRef]
  14. Li, X.; Niu, P.; Li, G.; Liu, J. An Adaptive Extreme Learning Machine for Modeling NOx Emission of a 300 MW Circulating Fluidized Bed Boiler. Neural Process. Lett. 2017, 46, 643–662. [Google Scholar] [CrossRef]
  15. Ma, Y.; Niu, P.; Yan, S.; Li, G. A modified online sequential extreme learning machine for building circulation fluidized bed boiler’s NOx emission model. Appl. Math. Comput. 2018, 334, 214–226. [Google Scholar] [CrossRef]
  16. Ma, Y.; Xu, C.; Wang, H.; Wang, R.; Liu, S.G.; Gu, X. Model NOx, SO2 Emissions Concentration and Thermal Efficiency of CFBB Based on a Hyper-Parameter Self-Optimized Broad Learning System. Energies 2022, 15, 7700. [Google Scholar] [CrossRef]
  17. Fan, Y.; Liu, X.; Zhang, C.; Li, C.; Li, X.; Wang, H. Dynamic prediction of boiler NOx emission with graph convolutional gated recurrent unit model optimized by genetic algorithm. Energy 2024, 294, 130957. [Google Scholar] [CrossRef]
  18. Xi, H.; Liao, P.; Wu, X. Simultaneous parametric optimization for design and operation of solvent-based post-combustion carbon capture using particle swarm optimization. Appl. Therm. Eng. 2020, 184, 116287. [Google Scholar] [CrossRef]
  19. Rahat, A.A.; Wang, C.; Everson, R.M.; Fieldsend, J.E. Data-driven multi-objective optimisation of coal-fired boiler combustion systems. Appl. Energy 2018, 229, 446–458. [Google Scholar] [CrossRef]
  20. Wang, F.; Ma, S.; Wang, H.; Li, Y.H.; Qin, Z.; Zhang, J. A hybrid model integrating improved flower pollination algorithm-based feature selection and improved random forest for NOx emission estimation of coal-fired power plants. Measurement 2018, 125, 303–312. [Google Scholar] [CrossRef]
  21. Hu, X.; Niu, P.; Wang, J.; Zhang, X. Multi-objective prediction of coal-fired boiler with a deep hybrid neural networks. Atmos. Pollut. Res. 2020, 11, 1084–1090. [Google Scholar] [CrossRef]
  22. Xu, W.; Huang, Y.; Song, S.; Chen, B.; Qi, X. A novel online combustion optimization method for boiler combining dynamic modeling, multi-objective optimization and improved case-based reasoning. Fuel 2023, 337, 126854. [Google Scholar] [CrossRef]
  23. Xu, W.; Huang, Y.; Song, S. On-line combustion optimization framework for coal-fired boiler combining improved cultural algorithm, deep learning, multi-objective evolutionary algorithm with improved case-based reasoning technology. Fuel 2024, 358, 130225. [Google Scholar] [CrossRef]
  24. Xu, W.; Poh, K.; Song, S.; Huang, Y. Research on reducing pollutant, improving efficiency and enhancing running safety for 1000 MW coal-fired boiler based on data-driven evolutionary optimization and online retrieval method. Appl. Energy 2025, 377, 123958. [Google Scholar] [CrossRef]
  25. Li, G.; Niu, P.; Ma, Y.; Wang, H.; Zhang, W. Tuning extreme learning machine by an improved artificial bee colony to model and optimize the boiler efficiency. Knowl. Based Syst. 2014, 67, 278–289. [Google Scholar] [CrossRef]
  26. Ma, Y.; Zhang, X.; Song, J.; Chen, L. A modified teaching-learning-based optimization algorithm for solving optimization problem. Knowl. Based Syst. 2020, 212, 106599. [Google Scholar] [CrossRef]
  27. Ma, Y.; Wang, H.; Zhang, X.; Hou, L.; Song, J. Three-objective optimization of boiler combustion process based on multi-objective teaching–learning based optimization algorithm and ameliorated extreme learning machine. Mach. Learn. Appl. 2021, 5, 100082. [Google Scholar] [CrossRef]
  28. Biau, G.; Scornet, E. A random forest guided tour. TEST 2015, 25, 197–227. [Google Scholar] [CrossRef]
  29. Couronné, R.; Probst, P.; Boulesteix, A. Random forest versus logistic regression: A large-scale benchmark experiment. BMC Bioinform. 2018, 19, 270. [Google Scholar] [CrossRef]
  30. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-based Learning Methods; Cambridge University Press: Cambridge, MA, USA, 2000. [Google Scholar]
  31. Steinwart, I.; Christmann, A. Support Vector Machines; Springer: New York, NY, USA, 2008. [Google Scholar]
  32. Huang, G.; Zhu, Q.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541), Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar]
  33. Huang, G.; Zhu, Q.; Siew, C.K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  34. Huang, G.; Zhou, H.; Ding, X.; Zhang, R. Extreme Learning Machine for Regression and Multiclass Classification. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2012, 42, 513–529. [Google Scholar] [CrossRef]
  35. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. arXiv 2017. [Google Scholar] [CrossRef]
  36. Hancock, J.T.; Khoshgoftaar, T.M. CatBoost for big data: An interdisciplinary review. J. Big Data 2020, 7, 94. [Google Scholar] [CrossRef] [PubMed]
  37. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient boosting with categorical features support. arXiv 2018, arXiv:1810.11363. [Google Scholar]
  38. Lundberg, S.M.; Erion, G.G.; Lee, S. Consistent Individualized Feature Attribution for Tree Ensembles. arXiv 2018, arXiv:1802.03888. [Google Scholar]
  39. Lundberg, S.M.; Lee, S. A Unified Approach to Interpreting Model Predictions. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  40. Li, S.; Chen, H.; Wang, M.; Heidari, A.; Mirjalili, S.M. Slime mould algorithm: A new method for stochastic optimization. Future Generarion Comput. Syst. 2020, 111, 300–323. [Google Scholar] [CrossRef]
  41. He, D.; He, C.; Jiang, L.; Zhu, H.; Hu, G. Chaotic characteristics of a one-dimensional iterative map with infinite collapses. IEEE Trans. Circuits Syst. I-Regul. Pap. 2001, 48, 900–906. [Google Scholar]
  42. Tizhoosh, H.R. Opposition-Based Learning: A New Scheme for Machine Intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (CIMCA-IAWTIC’06), Vienna, Austria, 28–30 November 2005; Volume 1, pp. 695–701. [Google Scholar]
  43. Seif, Z.; Ahmadi, M.B. An opposition-based algorithm for function optimization. Eng. Appl. Artif. Intell. 2015, 37, 293–306. [Google Scholar] [CrossRef]
  44. Arora, S.; Singh, S. Butterfly optimization algorithm: A kind of novel approach for global optimization. Soft Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  45. Mirjalili, S.M.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  46. Dhiman, G.; Kumar, V. Seagull optimization algorithm: Theory and its applications for large-scale industrial engineering problems. Knowl. Based Syst. 2019, 165, 169–196. [Google Scholar] [CrossRef]
  47. Xue, J.; Shen, B. A kind of novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
Figure 1. The flow chart of the proposed framework.
Figure 1. The flow chart of the proposed framework.
Biomimetics 09 00717 g001
Figure 2. The fitness value evaluation graph of slime mould.
Figure 2. The fitness value evaluation graph of slime mould.
Biomimetics 09 00717 g002
Figure 3. The flow chart of MSMA.
Figure 3. The flow chart of MSMA.
Biomimetics 09 00717 g003
Figure 4. The convergence curves of six algorithms on 30, 50 and 100 dimensional functions.
Figure 4. The convergence curves of six algorithms on 30, 50 and 100 dimensional functions.
Biomimetics 09 00717 g004aBiomimetics 09 00717 g004b
Figure 5. The convergence curve of six algorithms on the optimal design of three-bar truss problem.
Figure 5. The convergence curve of six algorithms on the optimal design of three-bar truss problem.
Biomimetics 09 00717 g005
Figure 6. The convergence curve of six algorithms on the optimal design of tension–compression spring problem.
Figure 6. The convergence curve of six algorithms on the optimal design of tension–compression spring problem.
Biomimetics 09 00717 g006
Figure 7. The fitting diagram and error diagram of the interpretable CatBoost model for modeling the thermal efficiency.
Figure 7. The fitting diagram and error diagram of the interpretable CatBoost model for modeling the thermal efficiency.
Biomimetics 09 00717 g007
Figure 8. The fitting diagram and error diagram of the interpretable CatBoost model for modeling the NOx emissions concentration.
Figure 8. The fitting diagram and error diagram of the interpretable CatBoost model for modeling the NOx emissions concentration.
Biomimetics 09 00717 g008
Figure 9. The feature analysis importance diagram and summary diagram of thermal efficiency.
Figure 9. The feature analysis importance diagram and summary diagram of thermal efficiency.
Biomimetics 09 00717 g009
Figure 10. The feature analysis dependency diagram of thermal efficiency.
Figure 10. The feature analysis dependency diagram of thermal efficiency.
Biomimetics 09 00717 g010
Figure 11. The feature analysis importance diagram and summary diagram of NOx emission concentration.
Figure 11. The feature analysis importance diagram and summary diagram of NOx emission concentration.
Biomimetics 09 00717 g011
Figure 12. The feature analysis dependency diagram of thermal efficiency.
Figure 12. The feature analysis dependency diagram of thermal efficiency.
Biomimetics 09 00717 g012
Figure 13. The optimization curves of MSMA for thermal efficiency and NOx emission concentration.
Figure 13. The optimization curves of MSMA for thermal efficiency and NOx emission concentration.
Biomimetics 09 00717 g013
Table 1. The symbol explanation of SMA and modified SMA.
Table 1. The symbol explanation of SMA and modified SMA.
SymbolExplanation
X ( t ) The positions of the slime mould individuals at the t th iterations
X b e s t ( t ) The best position of the slime mould individual at the t th iteration
X A t & X B t The randomly selected slime mould individual at the t th iteration
v b A random value
a A changing value
P The total number of iterations
v c A random value
r A random value
q A changing value
f ( · ) The output function of the fitness value
X i The position of the i th slime mould individual
S The optimal fitness value in all iterations
Q The population size of the slime mould individuals
W The weight of the slime mould individual
f b e s t The best fitness value at the current iteration
f w o r s t The worst fitness value at the current iteration
SortIndex ( · ) The output function of sorted fitness values
d A random value
X u b The upper bounds of the search space
X l b The lower bounds of the search space
u The proportion of randomly distributed slime mould individuals
Z i The chaotic sequence
β A random value
X A & X B The randomly selected slime mould individual at the t th iteration
X The elite reverse solution
X The reverse interpolation solution
Table 2. The description of ten benchmark testing functions.
Table 2. The description of ten benchmark testing functions.
FunctionExpressionRange f m i n
F1 f 1 ( x ) = i = 1 n   x i 2 [ 100 ,   100 ] n 0
F2 f 2 ( x ) = i = 1 n   x i + Π i = 1 n x i [ 10 , 10 ] n 0
F3 f 3 ( x ) = i = 1 n   j = 1 i   x j 2 [ 100 , 100 ] n 0
F4 f 4 ( x ) = m a x i   x i , 1 i n [ 10 , 10 ] n 0
F5 f 5 ( x ) = i = 1 n 1   100 x i + 1 x i 2 2 + x i 1 2 [ 30 , 30 ] n 0
F6 f 6 ( x ) = i = 1 n   x i + 0.5 2 [ 100 , 100 ] n 0
F7 f 7 ( x ) = i = 1 n   i x i 4 + r a n d o m [ 0,1 ] [ 1.28 , 1.28 ] n 0
F8 f 8 ( x ) = i = 1 n   x i s i n x i [ 500 , 500 ] n 418.9829 × n
F9 f 9 ( x ) = i = 1 n   x 1 2 10 c o s 2 π x i + 10 [ 5.12 , 5.12 ] n 0
F10 f 10 ( x ) = 20 e x p 0.2 1 n i = 1 n   x i 2 e x p 1 n i = 1 n   c o s 2 π x i + 20 + e [ 32 , 32 ] n 0
Table 3. Parameters setting of six optimization algorithms.
Table 3. Parameters setting of six optimization algorithms.
Algorithm Algorithm Parameters Setting
MSMAthe location update parameter is 0.03, the chaos mapping parameter is 0.7, the proportion of elite individuals is 10%
SMAthe location update parameter is 0.03
BOAthe witching probability is 0.8, the power exponent is 0.1, the sensory modality is 0.1
WOAthe probability parameter of strategy execution is 0.5
SOAthe ratio of flight path to speed is 0.7
SSAthe early warning value is 0.8, the proportion of discoverers is 0.2, the proportion of aware of dangerous sparrows is 0.1
Table 4. The comparison metrics for six algorithms on 30 dimensional functions.
Table 4. The comparison metrics for six algorithms on 30 dimensional functions.
FunctionMetricsMSMASMABOAWOASOASSA
F1mean2.98 × 10−3102.31 × 10−2425.72 × 10−103.40 × 10−221.83 × 10−2001.16 × 10−12
S.D.0.000.002.51 × 10−117.38 × 10−220.006.27 × 10−12
running time4.01 s3.87 s2.37 s2.48 s1.13 s3.07 s
F2mean5.12 × 10−1651.94 × 10−1571.11 × 10−65.13 × 10−156.85 × 10−1223.00 × 10−5
S.D.0.001.041 × 10−1563.92 × 10−77.16 × 10−153.07 × 10−1211.25 × 10−4
running time4.19 s4.05 s2.54 s2.61 s1.22 s3.35 s
F3mean4.00 × 10−2656.74 × 10−1965.35 × 10−107.84 × 10−33.72 × 1042.53 × 10−10
S.D.0.000.004.13 × 10−111.98 × 10−22.90 × 1041.35 × 10−9
running time6.53 s6.16 s8.10 s4.52 s3.05 s8.80 s
F4mean9.69 × 10−1676.06 × 10−1551.55 × 10−71.45 × 10−52.23 × 10−264.29 × 10−8
S.D.0.003.26 × 10−1547.42 × 10−92.53 × 10−51.20 × 10−251.51 × 10−7
running time4.16 s4.01 s2.43 s2.57 s1.17 s3.20 s
F5mean2.82 × 1012.84 × 1012.89 × 1012.71 × 1012.74 × 10−11.59 × 10−5
S.D.2.33 × 10−11.24 × 10−12.73 × 10−21.025.56 × 10−13.31 × 10−5
running time4.45 s4.29 s2.58 s2.69 s1.27 s3.48 s
F6mean8.29 × 10−29.47 × 10−25.791.153.45 × 10−29.50 × 10−8
S.D.3.68 × 10−24.40 × 10−25.29 × 10−15.51 × 10−13.01 × 10−11.26 × 10−7
running time4.64 s4.52 s2.62 s2.80 s1.28 s3.55 s
F7mean1.35 × 10−42.11 × 10−46.73 × 10−54.85 × 10−41.17 × 10−46.18 × 10−4
S.D.1.23 × 10−41.91 × 10−46.17 × 10−54.66 × 10−41.08 × 10−44.38 × 10−4
running time4.30 s4.15 s2.64 s2.70 s1.26 s3.43 s
F8mean−1.00 × 104−8.80 × 103−2.44 × 103−7.22 × 103−1.25 × 104−7.74 × 103
S.D.6.17 × 1025.61 × 1024.90 × 1022.89 × 1026.89 × 1012.05 × 103
running time4.28 s4.15 s2.47 s2.57 s1.12 s3.22 s
F9mean0.000.001.04 × 1021.51 × 10−140.001.24 × 10−10
S.D.0.000.008.03 × 1014.38 × 10−140.006.65 × 10−10
running time4.15 s4.00 s2.41 s2.58 s1.20 s3.30 s
F10mean4.44 × 10−164.44 × 10−161.85 × 10−73.46 × 10−124.44 × 10−162.99 × 10−6
S.D.0.000.008.74 × 10−99.37 × 10−120.009.42 × 10−6
running time4.31 s4.13 s2.91 s2.69 s1.30 s3.62 s
Table 5. The comparison metrics for six algorithms on 50 dimensional functions.
Table 5. The comparison metrics for six algorithms on 50 dimensional functions.
FunctionMetricsMSMASMABOAWOASOASSA
F1mean1.32 × 10−2772.46 × 10−2226.09 × 10−102.05 × 10−174.06 × 10−2043.27 × 10−11
S.D.0.000.002.87 × 10−114.84 × 10−170.001.72 × 10−10
running time6.68 s6.50 s3.64 s4.18 s1.75 s5.01 s
F2mean2.89 × 10−1472.07 × 10−1391.23 × 10201.91 × 10−121.75 × 10−1211.05 × 10−6
S.D.1.53 × 10−1461.11 × 10−1386.67 × 10202.22 × 10−129.02 × 10−1214.18 × 10−6
running time6.60 s6.43 s3.77 s4.16 s1.77 s5.07 s
F3mean7.00 × 10−2103.50 × 10−1485.54 × 10−101.331.45 × 1052.34 × 10−10
S.D.0.001.88 × 10−1473.87 × 10−113.636.87 × 1041.26 × 10−9
running time9.79 s9.26 s11.98 s6.86 s4.51 s13.29 s
F4mean1.25 × 10−1586.62 × 10−1391.63 × 10−71.97 × 10−41.99 × 10−302.04 × 10−7
S.D.6.33 × 10−1583.56 × 10−1388.79 × 10−92.32 × 10−41.06 × 10−298.90 × 10−7
running time6.71 s6.50 s3.59 s4.20 s1.74 s5.01 s
F5mean4.86 × 1014.85 × 1014.89 × 1014.78 × 1019.32 × 10−12.08 × 10−5
S.D.1.27 × 10−11.49 × 10−13.49 × 1027.57 × 10−12.493.03 × 10−5
running time6.94 s6.70 s3.65 s4.23 s1.80 s5.18 s
F6mean1.051.251.07 × 1013.198.93 × 10−21.91 × 10−7
S.D.2.98 × 10−13.30 × 10−16.71 × 10−17.21 × 10−16.86 × 10−23.65 × 10−7
running time6.77 s6.55 s3.51 s4.15 s1.74 s5.02 s
F7mean1.54 × 10−43.29 × 10−46.24 × 10−59.77 × 10−41.42 × 10−45.69 × 10−4
S.D.1.30 × 10−43.57 × 10−47.12 × 10−51.03 × 10−31.10 × 10−43.80 × 10−4
running time6.93 s6.69 s3.87 s4.37 s1.87 s5.32 s
F8mean−1.59 × 104−1.36 × 104−3.22 × 103−1.19 × 104−2.09 × 104−1.64 × 104
S.D.6.45 × 1028.04 × 1026.77 × 1024.55 × 1024.23 × 1014.80 × 103
running time6.89 s6.70 s3.71 s4.21 s1.68 s5.08 s
F9mean0.000.001.59 × 1028.33 × 10−140.008.46 × 10−8
S.D.0.000.001.59 × 1027.58 × 10−140.004.43 × 10−7
running time6.83 s6.55 s3.62 s4.25 s1.80 s5.16 s
F10mean4.44 × 10−164.44 × 10−161.88 × 10−74.82 × 10−104.44 × 10−169.30 × 10−8
S.D.0.000.006.70 × 10−94.60 × 10−100.003.18 × 10−7
running time7.60 s7.33 s4.60 s4.77 s2.10 s6.07 s
Table 6. The comparison metrics for six algorithms on 100 dimensional functions.
Table 6. The comparison metrics for six algorithms on 100 dimensional functions.
FunctionMetricsMSMASMABOAWOASOASSA
F1mean1.25 × 10−2492.78 × 10−1936.31 × 10−101.51 × 10−134.51 × 10−2086.83 × 10−12
S.D.0.000.002.42 × 10−113.30 × 10−130.003.64 × 10−11
running time12.82 s12.49 s6.47 s8.15 s3.18 s9.46 s
F2mean8.63 × 10−1267.83 × 10−1114.82 × 10506.93 × 10−101.86 × 10−1201.29 × 10−6
S.D.4.65 × 10−1254.21 × 10−1102.44 × 10518.53 × 10−101.00 × 10−1196.32 × 10−6
running tim × 1014.025313.56537.42088.81653.45 s10.3113
F3mean4.71 × 10−1648.45 × 10−926.00 × 10−101.79 × 1027.00 × 1051.26 × 10−6
S.D.0.004.55 × 10−913.30 × 10−112.23 × 1023.60 × 1056.82 × 10−6
running time21.50 s20.25 s25.96 s15.12 s9.72 s29.13 s
F4mean4.64 × 10−1531.87 × 10−1421.68 × 10−72.43 × 10−33.92 × 10−326.58 × 10−9
S.D.2.14 × 10−1521.01 × 10−1416.52 × 10−92.21 × 10−32.04 × 10−312.27 × 10−8
running time13.03 s12.64 s6.54 s8.22 s3.20 s9.58 s
F5mean9.87 × 1019.87 × 1019.89 × 1019.78 × 1014.85 × 10−12.31 × 10−5
S.D.1.16 × 10−11.14 × 10−12.46 × 10−26.45 × 10−16.17 × 10−15.55 × 10−5
running time13.26 s12.84 s6.55 s8.20 s3.23 s9.67 s
F6mean9.821.06 × 1012.30 × 1011.03 × 1011.76 × 10−11.60 × 10−7
S.D.1.281.208.00 × 10−11.231.77 × 10−12.08 × 10−7
running time14.65 s14.17 s7.16 s9.14 s3.57 s10.61 s
F7mean2.41 × 10−43.69 × 10−44.88 × 10−51.10 × 10−31.58 × 10−44.35 × 10−4
S.D.2.08 × 10−44.09 × 10−43.56 × 10−51.50 × 10−31.35 × 10−43.43 × 10−4
running time13.58 s13.17 s6.98 s8.61 s3.40 s10.08 s
F8mean−2.92 × 104−2.47 × 104−4.47 × 103−2.40 × 104−4.18 × 104−3.40 × 104
S.D.1.24 × 1031.31 × 1039.75 × 1028.97 × 1021.85 × 1029.05 × 103
running time14.00 s13.64 s7.04 s8.72 s3.15 s10.05 s
F9mean0.000.001.22 × 1022.48 × 10−80.002.94 × 10−9
S.D.0.000.002.74 × 1021.33 × 10−70.001.58 × 10−8
running time13.04 s12.63 s6.45 s8.27 s3.21 s9.58 s
F10mean4.44 × 10−164.44 × 10−161.93 × 10−77.94 × 10−84.44 × 10−162.94 × 10−7
S.D.0.000.006.85 × 10−91.70 × 10−70.001.51 × 10−6
running time13.99 s13.54 s7.52 s8.78 s3.52 s10.57 s
Table 7. The comparison of model performance for six algorithms on the optimal design of three-bar truss problem.
Table 7. The comparison of model performance for six algorithms on the optimal design of three-bar truss problem.
MetricsMSMASMABOAWOASOASSA
Mean2.63 × 1022.63 × 1022.64 × 1022.68 × 1022.67 × 1022.63 × 102
Std2.45 × 10−33.68 × 10−35.38 × 10−14.703.953.75 × 10−2
Min2.63 × 1022.63 × 1022.64 × 1022.64 × 1022.63 × 1022.63 × 102
Max2.63 × 1022.63 × 1022.66 × 1022.80 × 1022.76 × 1022.64 × 102
Time0.69 s0.64 s0.74 s0.33 s0.40 s0.70 s
Table 8. The comparison of model performance for six algorithms on the optimal design of tension–compression spring problem.
Table 8. The comparison of model performance for six algorithms on the optimal design of tension–compression spring problem.
MetricsMSMASMABOAWOASOASSA
Mean1.35 × 10−21.40 × 10−28.66 × 10321.31 × 10−26.66 × 10312.33 × 1032
Std8.79 × 10−42.95 × 10−33.39 × 10324.99 × 10−42.49 × 10324.22 × 1032
Min1.26 × 10−21.26 × 10−21.31 × 10−21.26 × 10−21.29 × 10−21.27 × 10−2
Max1.62 × 10−22.93 × 10−21.00 × 10331.48 × 10−21.00 × 10331.00 × 1033
Time0.77 s0.72 s0.68 s0.35 s0.37 s0.61 s
Table 9. Description of feature symbols.
Table 9. Description of feature symbols.
SymbolDescription
17ANO037Boiler load
AFCOALQCoal feeder feed rate 1
BFCOALQCoal feeder feed rate 2
CFCOALQCoal feeder feed rate 3
DFCOALQCoal feeder feed rate 4
18ANO073Average bed temperature in the upper part of the dense phase zone of the left side furnace
18ANO074Average bed temperature in the upper part of the dense phase zone of the furnace chamber
18ANO075Average bed temperature in the upper part of the dense phase zone of the right side furnace
05F051Primary air flow rate at the inlet of the left duct burner
05F061Right side duct burner inlet primary air flow rate
06F051Total secondary air flow rate on the left side
06F052Left inner secondary air distribution flow rate
06F061Total secondary air flow rate on the right side
06F062Secondary air distribution flow rate in the right side
05T457Left duct burner inlet primary air temperature
05T467Right side duct burner inlet primary air temperature
17I011Limestone powder conveying motor current 1
17I021Limestone powder conveying motor current 2
6CEMSO2Flue gas O2 concentration
6CEMSTEMPFlue gas temperature
08A051Carbon content of fly ash at the inlet of the left side dust collector
08A061Carbon content of fly ash at the inlet of the right side dust collector
05T402Primary fan inlet temperature 1
05T403Primary fan inlet temperature 2
06T401Secondary fan inlet temperature 1
06T402Secondary fan inlet temperature 2
NOXNOX emission concentration
TEBoiler thermal efficiency
Table 10. The operation conditions of circulation fluidized bed boiler.
Table 10. The operation conditions of circulation fluidized bed boiler.
No.17ANO037AFCOALQBFCOALQCFCOALQDFCOALQ18ANO07318ANO074
173.5237.6039.1739.2737.87865.03863.21
273.4039.1740.4140.4039.01863.65862.11
373.5239.2840.4140.4439.04862.89861.65
287866.3031.4847.2132.7435.99855.49858.57
287966.0531.1047.1132.3935.46855.84857.58
288065.0732.1748.2933.7136.94856.58856.53
No.18ANO07505F05105F06106F05106F05206F06106F062
1861.27236.65228.6423.89196.35412.66188.05
2860.54237.11238.25107.43148.38384.34216.19
3860.44219.26371.45188.84168.03397.97211.99
2878861.63181.49252.21359.57120.06353.54146.48
2879859.29226.35297.07190.77121.21352.30170.98
2880856.41202.55304.39233.81125.69347.88153.15
No.05T45705T46717I01117I0216CEMSO26CEMSTEMP08A051
1269.34267.38112.87101.665.55152.660.85
2269.34267.38109.93101.545.55152.660.85
3269.34267.38109.3297.965.55152.660.85
2878264.82261.96101.7784.577.27149.110.63
2879264.82261.96104.0284.807.27149.110.63
2880264.82261.96100.2884.617.27149.110.63
No.08A06105T40205T40306T40106T4026CEMSNOXTE
10.3225.6524.9023.8925.73131.5290.55
20.2725.6524.9023.8925.73134.3590.59
30.0625.6524.9023.8925.73134.8890.75
28780.3924.5823.2823.1923.75174.2589.96
28790.2624.5823.2823.1923.75175.3990.06
28800.1324.7123.2123.2223.84174.6390.13
Table 11. Parameter settings of four methods.
Table 11. Parameter settings of four methods.
Method Parameters Setting
CatBoostiterations = 1000, learning_rate = 0.03, max_depth = 6, l2_leaf_reg = 3.
RFn_estimators = 100, max_depth = None, min_samples_split = 2, min_samples_leaf = 1, max_features = Auto.
SVMkernel = rbf, gamma = scale, C = 1.0, epsilon = 0.1.
MLPhidden_size = 100, activation_function = relu, learning_rate = 0.001, iteration = 1000.
Table 12. The experimental results of four models on two objectives.
Table 12. The experimental results of four models on two objectives.
ObjectiveDatasetMetricsInterpretable CatBoostRFSVMMLP
TETraining dataRMSE1.85 × 10−35.75 × 10−34.66 × 10−24.96 × 10−2
MAE1.45 × 10−33.90 × 10−33.81 × 10−23.81 × 10−2
MAPE2.95 × 10−38.36 × 10−36.79 × 10−26.78 × 10−2
R299.98%99.98%86.53%85.62%
Testing dataRMSE5.27 × 10−31.53 × 10−24.72 × 10−25.08 × 10−2
MAE3.51 × 10−31.05 × 10−23.84 × 10−23.91 × 10−2
MAPE7.61 × 10−32.10 × 10−26.86 × 10−26.95 × 10−2
R299.85%98.68%85.78%84.47%
NOxTraining dataRMSE1.19 × 10−21.22 × 10−25.76 × 10−27.32 × 10−2
MAE9.27 × 10−38.20 × 10−34.79 × 10−25.64 × 10−2
MAPE2.76 × 10−21.78 × 10−28.94 × 10−21.06 × 10−1
R299.48%99.45%84.62%74.51%
Testing dataRMSE2.92 × 10−23.28 × 10−26.10 × 10−27.44 × 10−2
MAE2.13 × 10−22.21 × 10−24.96 × 10−25.71 × 10−2
MAPE4.25 × 10−24.48 × 10−29.27 × 10−21.07 × 10−1
R296.79%95.84%82.46%73.47%
Table 13. The value range of adjustable parameters of the combustion system of the CFBB.
Table 13. The value range of adjustable parameters of the combustion system of the CFBB.
Decision VariableOriginal RangeRange of Positive Effect on TERange of Negative Effect on NOxReduced Range
AFCOALQ[27.67, 68.12][27.67, 60.40][27.67, 63.86][27.67, 56.90]
BFCOALQ[25.17, 61.44][33.03, 61.44][25.17, 56.31][33.03, 51.09]
CFCOALQ[25.54, 61.93][35.00, 61.93][39.59, 61.93][42.49, 61.93]
DFCOALQ[27.23, 67.89][33.92, 58.11][27.23, 62.84][33.92, 58.11]
05F051[123.59, 487.95][123.59, 487.95][151.97, 456.13][151.97, 456.13]
05F061[91.78, 513.12][91.78, 458.88][91.78, 448.35][91.78, 448.35]
06F051[15.05, 762.78][15.05, 762.78][15.05, 762.78][15.05, 762.78]
06F052[120.06, 869.41][121.21, 856.65][121.21, 869.41][121.21, 856.63]
06F061[318.67, 1217.79][320.80, 1205.40][318.67, 1217.79][320.80, 1205.40]
06F062[144.19, 888.01][144.19, 728.56][144.19, 888.01][144.19, 547.852]
6CEMSO2[3.15, 8.13][3.15, 5.49][3.15, 5.14][3.15, 5.14]
08A051[0.61, 1.47][0.61, 1.07][0.74, 1.47][0.74, 1.07]
08A061[−0.01, 0.40][−0.01, 0.20][0.01, 0.40][0.01, 0.20]
Table 14. The optimization results of MSMA for thermal efficiency and NOx emission concentration.
Table 14. The optimization results of MSMA for thermal efficiency and NOx emission concentration.
No.TE
(Pre-Opt)
TE
(Optimized)
TE
(Opt-Ratio)
NOx
(Pre-Opt)
NOx
(Optimized))
NOx
(Opt-Ratio)
Time
990.3291.12+0.89%139.9296.59−30.97%6.39 s
1090.4091.13+0.81%138.6999.60−28.19%6.17 s
1590.7091.15+0.49%134.4289.80−33.19%6.33 s
2290.6291.21+0.65%149.76102.44−31.60%6.16 s
2390.7291.21+0.54%150.44101.99−32.21%6.50 s
285490.6291.06+0.49%173.41102.81−40.71%6.36 s
285790.7191.11+0.43%173.10102.95−40.53%6.12 s
286190.7991.04+0.28%168.9099.99−40.80%6.09 s
287190.2491.04+0.89%168.5294.66−43.83%6.07 s
287390.2391.06+0.93%170.9697.38−43.04%6.43 s
Average90.4791.08+0.68%148.9589.03−37.55%6.40 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, S.; Ma, Y. A Multi-Objective Optimization Framework That Incorporates Interpretable CatBoost and Modified Slime Mould Algorithm to Resolve Boiler Combustion Optimization Problem. Biomimetics 2024, 9, 717. https://doi.org/10.3390/biomimetics9110717

AMA Style

Gao S, Ma Y. A Multi-Objective Optimization Framework That Incorporates Interpretable CatBoost and Modified Slime Mould Algorithm to Resolve Boiler Combustion Optimization Problem. Biomimetics. 2024; 9(11):717. https://doi.org/10.3390/biomimetics9110717

Chicago/Turabian Style

Gao, Shan, and Yunpeng Ma. 2024. "A Multi-Objective Optimization Framework That Incorporates Interpretable CatBoost and Modified Slime Mould Algorithm to Resolve Boiler Combustion Optimization Problem" Biomimetics 9, no. 11: 717. https://doi.org/10.3390/biomimetics9110717

APA Style

Gao, S., & Ma, Y. (2024). A Multi-Objective Optimization Framework That Incorporates Interpretable CatBoost and Modified Slime Mould Algorithm to Resolve Boiler Combustion Optimization Problem. Biomimetics, 9(11), 717. https://doi.org/10.3390/biomimetics9110717

Article Metrics

Back to TopTop