You are currently viewing a new version of our website. To view the old version click .
Energies
  • Article
  • Open Access

27 November 2023

Optimizing the Controlling Parameters of a Biomass Boiler Based on Big Data

,
,
,
,
,
,
and
1
School of Energy, Power and Mechanical Engineering, North China Electric Power University, Beijing 102206, China
2
School of New Energy, North China Electric Power University, Beijing 102206, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section A4: Bio-Energy

Abstract

This paper presents a comprehensive method for optimizing the controlling parameters of a biomass boiler. The historical data are preprocessed and classified into different conditions with the k-means clustering algorithm. The first-order derivative (FOD) method is used to compensate for the lag of controlling parameters, the backpropagation (BP) neural network is used to map the controlling parameters with the boiler efficiency and unit load, and the ant colony optimization (ACO) algorithm is used to search the opening of air dampers. The results of the FOD-BP-ACO model show an improvement in the boiler efficiency compared to the predicted values of FOD-BP and the data compared to the historical true values were observed. The results suggest that this FOD-BP-ACO method can also be used to search and optimize other controlling parameters.

1. Introduction

As a renewable energy, biomass energy is drawing a lot of attention around the world. The installed biomass power energy capacity has exceeded 120 GW (as shown in Figure 1) [1].
Figure 1. The total installed capacity of global biomass power generation from 2012 to 2021 [1].
To improve the efficiency of the plant, the big data analysis method was attempting to optimize the operation of the boiler. Li et al. [2] proposed an artificial neural network (ANN) and showed better regression accuracy and generalization ability with faster learning speed. Krzywanski et al. [3,4] built a feedforward ANN to predict SO2 emissions. Gao et al. [5] proposed a method based on improved particle swarm (HPSO) and support vector machine (SVM) algorithms for the prediction of the NOx concentration in flue gas. Li et al. [6] used the Lasso algorithm of machine learning to conduct a correlation analysis of boiler control parameters and operating state parameters, and extracted highly correlated control parameters and operating state parameters. A nonlinear combined deep confidence network (NCDBN) was used to predict the exhaust gas’s boiler efficiency and NOx content.
L. D. Blackburn [7] proposed a constrained dynamic optimization using a recurrent neural network model combined with a meta-heuristic optimizer. An improvement was achieved for the simulated coal-fired boiler compared to the non-optimal situation. Yao et al. [8] proposed an ANN model to improve the unit heat rate by optimizing the boiler load, boiler excess oxygen (O2), fly ash unburned carbon, and descaling flow rate for coal-fired boilers.
Wei Tian and Yu Cao proposed an innovative study where they optimized the backpropagation (BP) neural network based on the firefly algorithm (FA). They also introduced the sparrow search algorithm (SSA), altering the search mechanism of BP to address the issue of local optima affecting traditional BP [9].
However, the above studies do not consider the large inertia of boiler parameters. Delays in predicting parameters may hinder the ability to take timely corrective actions and lead to suboptimal performance.
Hence, this study considers the effect of delay parameter compensation, and an FOD-BP-ACO model is proposed. The individual steps for the optimization of control parameters are summarized as follows:
Step I: Data cleaning with the PauTa criterion and classification with the k-means method.
Step II: Parameter selection with the average impact value (AIV) method.
Step III: Compensating the lag of controlling parameters with the first-order derivative (FOD) method.
Step IV: Modeling with backpropagation (BP) neural networks and random forest.
Step V: Optimizing the opening of the air damper through the ant colony algorithm.
The biomass direct-fired 130 t/h (9.2 MPa and 540 °C) boiler structure diagram is showed in Figure 2 and all the used data are collected from this power plant.
Figure 2. Simplified diagram of the biomass boiler.

2. Materials and Methods

2.1. Data Processing and Analysis

Boiler operational data frequently exhibit anomalies due to various sources of interference, such as noise, sensor malfunctions, abnormal operational conditions, system instability, and other factors. It is necessary to assess the reliability of the measurement data [10].
(1) Data cleaning
The threshold of 3 σ -based PauTa [11] criterion is used as the outlier detection method. The standard deviation is calculated with Equation (1):
σ = k = 1 n X k X ¯ 2 n
where  n  is the number of observations,  X ¯  is the mean of the observations, and  σ  is the standard deviation.
The residual error is calculated with Equation (2):
R n k = X k X =
where  R n  is the residual, and  X =  is the estimated value of the observation.
When the residual error surpasses 3 σ , it is categorized as a gross error and the data are screened out; otherwise, it is deemed a random error and the data are kept. Thus, the dataset is refined, reducing its size from 22,900 to 22,477 groups.
(2) Data classification
The operational status of the biomass boiler tends to vary with unit load. It is difficult for a single model to accurately predict parameters for all operational scenarios. Therefore, the k-means algorithm is employed to classify the pre-cleaning data and carried out as follows [12,13,14]:
(1) Normalized raw data;
(2) Initialize k center points;
(3) Data partitioning.
Calculate the distance between each sample  x i  in the dataset  D = x 1 ,   x 2 ,   ,   x m  and k center points  μ j :
d i j = x i μ j 2 2
Divide each data sample  x i  into clusters  u j  corresponding to the shortest distance  d i j  from the center point.
(4) Update center point
Recalculate the center points of samples  C j j = 1 , 2 , , k  within each cluster:
u i = 1 C j x C j x
Repeat Steps (3) and (4) until the iteration ends or the partitioning results remain unchanged for each iteration.
The choice of the k value in k-means significantly impacts condition classification accuracy. The elbow method helps determine the optimal k value as follows:
(a) Calculate the sum of squared errors (SSE) value:
Assume there are  n 1 ,   n 2 , ,   n k  data samples in k clusters. The SSE is the sum of squared distances from each data point to its cluster center, calculated using Equation (5):
S S E = j = 1 k i = 1 n x i j u j 2 2
(b) Plot SSE values against different k values.
(c) The SSE vs. k graph typically resembles an elbow, and the k value at the “elbow point” indicates the optimal cluster count for classification.
Three parameters (steam drum pressure, feedwater flow rate, and unit load) which are strongly correlated with the boiler operational state are used as sample co-ordinates for k-means classification. As shown in Figure 3, the k value corresponding to the elbow point is 2. Then, the experimental dataset is partitioned into two categories (condition 1 and condition 2), delineated by the boundaries of steam drum pressure, feedwater flow rate, and unit load. The parameter ranges for each type of operating condition, along with the quantity of data encompassed within each operational group, are itemized in Table 1, respectively.
Figure 3. SEE values at different k values.
Table 1. Parameter ranges of various operating conditions after classification.

2.2. Parameter Extraction

To improve model accuracy and reduce modeling time, tens of thousands of measurement points in the distributed control system (DCS) are screened. The average influence value method based on the BP neural network is used to screen the characteristic parameters of two output parameters, respectively:
(1)
Construct samples
Convert the initialization dataset into an independent variable matrix  X = X 1 ,   X 2 ,   ,   X m T  with l rows and m columns and a dependent variable matrix  Y = Y 1 ,   Y 2 ,   ,   Y m T  with m rows and 1 column, where l represents the number of data samples and m is the number of feature parameters.
(2)
Normalization processing
Normalized input parameters:
x i , j = X i , j X m i n i , j X m a x i , j X m i n i , j
Normalized output parameters:
y i = Y i Y m i n Y m a x Y m i n
(3)
Establish a model
Using the normalized data from the previous step as input and output parameters, combined with the BP neural network algorithm, establish a prediction model for the target parameters.
(4)
Average impact value (AIV) analysis
Enlarge and shrink each column of the input parameter matrix to 1.1 and 0.9 times the original data to obtain the scaling matrices, as shown in Equations (8) and (9):
x j 1 = x 11 , x 12 , , 1.1 x 1 j , , x 1 n x 21 , x 22 , , 1.1 x 2 j , , x 2 n x m 1 , x m 2 , , 1.1 x mj , , x mn
x j 2 = x 11 , x 12 , , 0.9 x 1 j , , x 1 n x 21 , x 22 , , 0.9 x 2 j , , x 2 n x m 1 , x m 2 , , 0.9 x mj , , x mn
Place the scaled matrices  x j 1  and  x j 2  into the BP neural network model established in step (3) for prediction, and obtain the prediction results  y j 1  and  y j 2 . The difference between two squares of  y j 1  and  y j 2  is calculated to obtain the influence value IV of the corresponding characteristic variable on the output parameter.
I V j = y j 1 2 y j 2 2
(5)
Calculate the average of each IV value to obtain the AIV:
A I V j = 1 m i = 1 m I V i j
(6)
Calculate contribution rate:
η j = A I V j i = 1 m A I V i
If the cumulative contribution rate of the set of feature variables reaches 80%, it is considered that there is a significant connection between the set of feature parameters and the output parameters, which are used as input parameters for the optimizing model.

2.3. Algorithm Optimization

In this study, the backpropagation neural network [9,15,16] with an input layer, a single hidden layer, and an output layer is selected as the core algorithm. The activation function for the hidden layer is set to sigmoid. At the same time, the random forest algorithm [17,18] is employed for comparison.
In practical applications, using only the BP algorithm poses challenges due to the asynchronous changes between input parameters, influenced by the control system and the inertia of combustion within the furnace, and the corresponding output parameters. The sampling process, indexed at the same time point, may compromise the model’s accuracy, especially when dealing with terminal output parameters like unit load in DCS systems, directly impacting the predictive model’s precision. Therefore, this study innovatively combines first-order derivative (FOD) with the BP algorithm. Additionally, a search mechanism based on ant colony optimization (ACO) is incorporated [19,20,21], enabling the FOD-BP-ACO model to effectively enhance boiler thermal efficiency.
Assume that the probability of ant k (k = 1, 2, …, m) transitioning from location i to location j at a certain moment is  p i j k . The calculation formula is as follows:
p i j k = τ i j α   ·   η i j β j J k i τ i j α   ·   η i j β , j ϵ   J k i 0 ,   o t h e r w i s e
In the equation,  τ i j α  represents the amount of pheromone between location i and location j on the path at a certain moment,  η i j β  represents the expected level of ant transitioning from i to j J k i  is the set of accessible locations for the ant in the next step,  α  is the importance factor of pheromone, and  β  is the importance factor of heuristics. The operation of ant colony optimization is shown in Figure 4.
Figure 4. The operation of ant colony optimization.
When m ants complete one iteration of path traversal, typically, the ACO model is used to update the pheromone. The following formula can be used to update the pheromone concentration:
τ i j t + 1 = 1 ρ τ i j t + Δ τ i j t
In this formula,  ρ  is the evaporation rate (a value between 0 and 1), and  Δ τ i j  represents the amount of pheromone deposited by the ants on the path from i to j during the iteration.
This study does not aim to solve a path problem. Therefore, the two equations mentioned above will be rewritten to solve the maximization of a specified function:
p i = m a x s 1 , m τ s t τ i t m a x s 1 , m τ s t
In this formula,  τ i 0 = F X , τ 0  represents the initialization of pheromone values for the data.  F X  represents the specified objective function and the optimization problem is transformed into a maximization problem; the formula for updating the pheromone concentration can be modified as follows:
τ i t + 1 = 1 ρ τ i t + Q × F X
Q  represents the total pheromone concentration released by the ants after completing one iteration or cycle. As  F X  increases, it indicates a higher value of the objective function, which, in turn, leads to a larger overall pheromone concentration at the location where the ants are located.
The introduction of the ACO algorithm significantly enhances the local search capability of the BP algorithm. Additionally, to improve the precision of the predictive model, this study proposes the optimization of BP through the introduction of FOD. The FOD is used to estimate the delay time of a particular parameter relative to another output parameter within a given system or process, and the specific process is as follows:
Firstly, calculate the Pearson correlation coefficient of the unit load characteristic parameters with Equation (17):
r = i = 1 n X I X ¯ Y i Y ¯ i = 1 n X I X ¯ 2 i = 1 n Y i Y ¯ 2
Next, conduct a lag analysis by selecting representative parameters with a correlation coefficient greater than 0.5 relative to the unit load from the input parameters. Subsequently, the interpolation function is fitted to the scattered data points over a complete period, the FOD of the fitted function is performed, and the estimation of the parameter relative to the output parameter is obtained by comparing the time difference between adjacent extreme points delay.
Assuming there are n hysteresis-compensated operating state parameters, for  i = 1 , 2 n , define the latency  t d i  as shown in the following Equation (18):
P t p = 0 P t p t s P t p + t s < 0 B i t b = 0 B i t b i t s B i t b i + t s < 0 0 < t b i t p < t d m a x t d i = t b t p
where  P  is for unit load,  B i  is the ith hysteresis-compensated operating state parameter,  t p  is for the unit load interpolation function corresponding to the extreme value point,  t b  represents the ith running parameter interpolating the extremum point corresponding to the time,  t s  represents DCS sampling interval,  t d i  represents the latency for running parameter  B i , and    t d m a x  is for maximum lag time of biomass boiler unit load.
The FOD-BP algorithm corrects the weights and thresholds of hidden layer nodes by propagating the error values generated after forward propagation back to the hidden layer, distributing them across the nodes. This process is iteratively repeated until the iteration reaches the minimum error value and concludes.
The FOD-BP model is a hybrid neural network based on the traditional BP algorithm. To enhance the local search capability of the BP algorithm, the ACO algorithm is introduced into the network’s weight-updating mechanism. The FOD algorithm adjusts the delay time difference to provide accurate input parameters for the BP algorithm, improving the model accuracy. By combining the FOD and ACO algorithms, it effectively obtains more accurate input parameters, thereby enhancing the precision of the BP model and maximizing thermal efficiency through control parameter adjustments.

3. Results and Discussion

3.1. FOD Compensation Result

As shown in Figure 5, there exists a certain linear correlation between the unit load and furnace temperature. Moreover, it is observable that, when the time changes, the load changes exhibit a noticeable lag compared to the furnace temperature curve. Therefore, we introduce the first derivative method to improve the model accuracy.
Figure 5. Unit load–furnace temperature variation.
Thus, parameters whose Pearson correlation coefficients are greater than 0.50 with the unit load are selected as compensation parameters in Table 2.
Table 2. Hysteresis analysis parameters.
The average lag time for each compensation parameter is shown in Table 3. The compensation effect is verified using data from unit load operating condition 2, 70% of the data are randomly selected as the training set, 30% of the data are used as the test set, and the model prediction results before and after compensation are shown in Figure 6. The mean relative error (MRE) is reduced by 15.64% after hysteresis compensation compared to before hysteresis (as shown in Table 4). This indicates that FOD compensation can improve forecasting accuracy to a certain extent.
Table 3. Lag time for compensation parameter.
Figure 6. Compensation effect: (a) compensation before unit load forecast results; and (b) unit load forecast results after compensation.
Table 4. Performance evaluation parameters of boiler thermal efficiency prediction model before and after compensation.

3.2. Unit Load and Boiler Efficiency Prediction

Predictive models for the unit load and boiler thermal efficiency under two different operating conditions were established using the random forest algorithm and the BP neural network algorithm, and a comparison of the predictive results between the two algorithms was conducted.

3.2.1. Random Forest Model

We used grid-search–cross-validation-optimized hyperparameters for unit load forecasting using a dataset of operating condition 2, with R2 as the evaluation index. Similar tuning was applied to other datasets, yielding optimal points as detailed in Table 5.
Table 5. Optimization results of random forest hyperparameters.
It can be observed from Table 6 that, apart from the unit load condition 1, the relative errors of the random forest predictive models for all operating conditions are below 1%, with R2 values exceeding 0.95. Among them, the single model with the highest accuracy is the boiler thermal efficiency predictive model 1, which has an average MRE of only 0.02%.
Table 6. Evaluation metrics for random forest prediction models.

3.2.2. BP Neural Network

BP neural network models were developed using the PyCharm Community edition 2023.2. The model parameters settings are presented in Table 7.
Table 7. Parameter settings for the BP neural network prediction model.
Taking the unit load as an example, the number of input layer nodes was 38, and the range of hidden layer nodes given according to different empirical formulae is l < 60, with comprehensive consideration, and 11 numbers in the range of 5–55 were taken at intervals of 5. Similarly, the hidden layer node selection was carried out for the thermal efficiency of the boiler, and the BP neural network prediction model was established by using the data after the division of operating conditions, and the root mean square error (RMSE) corresponding to the number of nodes is shown in Table 8.
Table 8. The number of hidden layer nodes of operating condition 1 predictive model corresponds to the RMSE.
The number of nodes corresponding to the lowest RMSE of the above two types of operating conditions is shown in Table 9.
Table 9. The optimal number of hidden layer nodes in a BP neural network.
The results of BP neural network predictions are shown in Figure 7.
Figure 7. Results of BP neural network: (a) prediction results of thermal efficiency of boilers under condition 1; (b) prediction results of thermal efficiency of boilers under condition 2; (c) prediction results of unit load under condition 1; and (d) prediction results of unit load under condition 2.
From Table 10, it can be seen that the mean relative error of the BP neural network predictive model under all operating conditions is within 1.5%, and the value of R2 is above 0.94, among which the single model with the highest accuracy is the boiler thermal efficiency prediction model under operating conditions 1, and its mean relative error is only 0.04%. In summary, the BP neural network prediction model can accurately predict the boiler thermal efficiency and unit load under all operating conditions.
Table 10. Evaluation metrics for BP neural network prediction models.
A detailed comparison is as follows:
(1)
Unit load: For condition 1, the MRE of the random forest predictive model is 1.10%, while the MRE of the BP neural network predictive model is 0.88%. This represents a decrease of approximately 25% in the MRE compared to the random forest predictive results. For condition 2, the random forest predictive model has an MRE of 0.72%, while the BP neural network predictive model has an MRE of 0.58%. In this case, the MRE of the BP neural network predictive model is approximately 24.1% lower than that of the random forest.
(2)
Thermal efficiency: For condition 1, the MRE value of the BP neural network predictive model decreased by approximately 100% compared to the random forest predictive results. For condition 2, the decrease was approximately 300%.
The BP neural network algorithm maintains an MRE of within 1.5% in all predictive models, indicating an overall superior predictive performance compared to random forest. Taking into account both modeling accuracy and stability, the next section will proceed with the optimization of control parameters using the model established by the BP neural network algorithm.

3.3. Optimization of Boiler Efficiency Using Ant Colony Algorithm

One-hour datasets were randomly selected for each type of operating condition and the ant colony algorithm was used to search for the corresponding damper positions that maximize boiler efficiency. The hyperparameters of the ant colony algorithm are shown in Table 11, and the variation range of the independent variable (high-side, center, and low-side boiler grate air damper positions, as well as front and rear furnace wall secondary air damper positions) of the objective function was specified to be 0.9–1.1 times of the original value and the opening of the baffle was not higher than 100%. The FOD-BP neural network models obtained were selected for the forward prediction model. The changes of the high-side, center-side, and low-side exhaust air damper opening, the secondary air damper opening of the front wall and the rear wall of the furnace, and boiler efficiency, before and after optimization, are shown in Table 11. The FOD-BP forecast value is the value obtained from the initial input parameters, and the FOD-BP-ACO forecast value is the value obtained from the input parameters after optimization.
Table 11. Hyperparameter configuration for BP ant colony algorithm.
The results show that the thermal efficiency of the boiler increases by 0.002–0.04% compared with the pre-optimization value. However, it is worth noting that, in certain instances, such as at 2160 s, even with optimization, the thermal efficiency remains below the actual values, albeit with a 0.02% improvement. This is primarily due to the initial predictive model having some degree of error, resulting in significant deviations between the predicted and actual values, even after optimization. Out of the 902 datasets analyzed, 95% of them demonstrated an improvement in boiler thermal efficiency by 0.01–0.06% relative to the actual values when the unit load variation did not exceed 0.5 MW. The performance data for different parameter types is summarized in Table 12.
Table 12. Performance data for different parameter types.

4. Conclusions

In this paper, a prediction model based on FOD-BP-ACO is proposed to optimize the controlling parameter of the biomass boiler. The main conclusions are summarized below:
The k-means clustering method is introduced to categorize data into two distinct operational states based on boundary parameters (steam drum pressure, feedwater flow rate, and unit load). When the delay parameter compensation is used, the average relative error of the model is reduced by 25.78%, which indicates that the delay compensation can effectively improve the accuracy of the prediction. The ant colony algorithm is proposed to optimize the air damper opening and show that the thermal efficiency of the boiler increases by 0.002–0.04% (relative to the predicted value) and 95% of the data demonstrated an improvement in boiler thermal efficiency by 0.01–0.06% relative to the actual values. This improvement is limited, because the model was based on the data of optimized operation condition and fixed biomass fuel. When the type of biomass fuel varies in a large range, FOD-BP-ACO might show a good potential for boiler efficiency improvement and providing a new reference for establishing the intelligent control system of biomass boilers.

Author Contributions

Conceptualization, J.H. and J.Z.; methodology, J.Z.; software, L.W.; validation, J.H., J.Z. and L.W.; formal analysis, J.H.; investigation, J.H.; resources, X.H.; data curation, J.X.; writing—original draft preparation, Y.Z.; writing—review and editing, X.W.; visualization, C.D.; supervision, X.W.; project administration, C.D.; funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, J.J. Zhang, upon reasonable request.

Acknowledgments

The authors would like to express their gratitude to all those who contributed to this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. International Renewable Energy Agency. Renewable Capacity Statistics 2023. Available online: https://www.irena.org/Publications/2023/Mar/Renewable-capacity-statistics-2023 (accessed on 26 March 2023).
  2. Li, G.; Niu, P.; Wang, H.; Liu, Y. Least Square Fast Learning Network for modeling the combustion efficiency of a 300WM coal-fired boiler. Neural Netw. 2014, 51, 57–66. [Google Scholar] [CrossRef] [PubMed]
  3. Krzywanski, J.; Czakiert, T.; Blaszczuk, A.; Rajczyk, R.; Muskala, W.; Nowak, W. A generalized model of SO2 emissions from large- and small-scale CFB boilers by artificial neural network approach: Part 1. The mathematical model of SO2 emissions in air-firing, oxygen-enriched and oxycombustion CFB conditions. Fuel Process. Technol. 2015, 137, 66–74. [Google Scholar] [CrossRef]
  4. Krzywanski, J.; Czakiert, T.; Blaszczuk, A.; Rajczyk, R.; Muskala, W.; Nowak, W. A generalized model of SO2 emissions from large- and small-scale CFB boilers by artificial neural network approach Part 2. SO2 emissions from large- and pilot-scale CFB boilers in O2/N2, O2/CO2 and O2/RFG combustion atmospheres. Fuel Process. Technol. 2015, 139, 73–85. [Google Scholar] [CrossRef]
  5. Gao, X.; Fu, Z.; Zhang, L.; Liu, B. Research on the Application of Big Data Technology in the Development of Coal-Fired Power Plants. J. Shenyang Inst. Eng. 2018, 14, 16–22. [Google Scholar]
  6. Li, Y.Y. Research on the Prediction and Control System of Nitrogen Oxides Emissions from Power Plant Boilers. Ph.D. Dissertation, Northeast Electric Power University, Jilin, China, 2020. [Google Scholar]
  7. Blackburn, L.D.; Tuttle, J.F.; Andersson, K.; Hedengren, J.D.; Powell, K.M. Dynamic machine learning-based optimization algorithm to improve boiler efficiency. J. Process Control. 2022, 120, 129–149. [Google Scholar] [CrossRef]
  8. Yao, Z.; Romero, C.; Baltrusaitis, J. Combustion optimization of a coal-fired power plant boiler using artificial intelligence neural networks. Fuel 2023, 344, 128145. [Google Scholar] [CrossRef]
  9. Tian, W.; Cao, Y. Evaluation model and algorithm optimization of intelligent manufacturing system on the basis of BP neural network. Intell. Syst. Appl. 2023, 20, 200293. [Google Scholar] [CrossRef]
  10. Maharana, K.; Mondal, S.; Nemade, B. A review: Data pre-processing and data augmentation techniques. Glob. Transit. Proc. 2022, 3, 91–99. [Google Scholar] [CrossRef]
  11. Xia, J.; Zhang, J.; Wang, Y.; Han, L.; Yan, H. WC-KNNG-PC: Watershed clustering based on k-nearest-neighbor graph and Pauta Criterion. Pattern Recognit. 2022, 121, 108177. [Google Scholar] [CrossRef]
  12. Ay, M.; Özbakır, L.; Kulluk, S.; Gülmez, B.; Öztürk, G.; Özer, S. FC-Kmeans: Fixed-centered K-means algorithm. Expert Syst. Appl. 2023, 211, 118656. [Google Scholar] [CrossRef]
  13. Selim, S.Z.; Ismail, M.A. K-means-type algorithms: A generalized convergence theorem and characterization of local optimality. IEEE Trans. Pattern Anal. Mach. Intell. 1984, 6, 81–87. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, B.; Fu, Z.; Wang, P.; Liu, L.; Gao, M.; Liu, J. Big-Data-Mining-Based Improved K-Means Algorithm for Energy Use Analysis of Coal-Fired Power Plant Units: A Case Study. Entropy 2018, 20, 702. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, J.; Zou, B.; Liu, M.; Li, Y.; Ding, H.; Xue, K. Milling force prediction model based on transfer learning neural network. J. Intell. Manuf. 2020, 32, 947–956. [Google Scholar] [CrossRef]
  16. Huang, X.; Li, Q.; Tai, Y.; Chen, Z.; Zhang, J.; Shi, J.; Cao, B.; Liu, W. Hybrid deep neural model for hourly solar irradiance forecasting. Renew. Energy 2021, 171, 1041–1060. [Google Scholar] [CrossRef]
  17. Gall, J.; Lempitsky, V. Class-specific Hough forests for object detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  18. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  19. Cui, J.; Guo, X.; Zhan, Y.; Pang, R. An inverse analysis method to identify maximum overfire temperature based on an improved ant colony algorithm. J. Build. Eng. 2022, 59, 105104. [Google Scholar] [CrossRef]
  20. Dorigo, M.; Gambardella, L.M. Ant Colony System: A cooperative learning approach to the traveling salesman problem. IEEE Trans. Evol. Comput. 1997, 1, 53–66. [Google Scholar] [CrossRef]
  21. Mirjalili, S.; Mirjalili, S.S.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.