Next Article in Journal
Development and Application of Film-Forming Nano Sealing Agent for Deep Coal Seam Drilling
Previous Article in Journal
Detection of Short-Circuit Faults in Induction Motor Winding Turns Using a Neural Network and Its Implementation in FPGA
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Dynamic Modeling of Bed Temperature in Utility Circulating Fluidized Bed Boilers Using a Hybrid CEEMDAN-NMI–iTransformer Framework

1
School of Energy and Environment, Southeast University, Nanjing 210096, China
2
Inner Mongolia Jingtai Power Generation Co., Ltd., Ordos 010399, China
3
Beijing Jingneng Power Co., Ltd., Beijing 100123, China
4
School of Intelligent Manufacturing, Jianghan University, Wuhan 430056, China
*
Author to whom correspondence should be addressed.
Processes 2025, 13(3), 816; https://doi.org/10.3390/pr13030816
Submission received: 9 February 2025 / Revised: 3 March 2025 / Accepted: 4 March 2025 / Published: 11 March 2025
(This article belongs to the Section Energy Systems)

Abstract

:
Circulating fluidized bed (CFB) boilers excel in low emissions and high efficiency, with bed temperature serving as a critical indicator of combustion stability, heat transfer efficiency, and pollutant reduction. This study proposes a novel framework for predicting bed temperature in CFB boilers under complex operating conditions. The framework begins by collecting historical operational data from a power plant Distributed Control System (DCS) database. Next, the Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) algorithm is employed to decompose the raw signals into distinct modes. By analyzing the trade-offs of combining modes with different energy levels, data denoising and outlier reconstruction are achieved. Key features are then selected using Normalized Mutual Information (NMI), and the inflection point of NMI values is used to determine the number of variables included. Finally, an iTransformer-based model is developed to capture long-term dependencies in bed temperature dynamics. Results show that the CEEMDAN-NMI–iTransformer framework effectively adapts to diverse datasets and performs better in capturing spatiotemporal relationships and delivering superior single-step prediction accuracy, compared to Gated Recurrent Unit (GRU), Long Short-Term Memory (LSTM), and Transformer models. For multi-step predictions, the model achieves accurate forecasts within 6 min and maintains an R2 above 0.95 for 24 min predictions, demonstrating robust predictive performance and generalization.

1. Introduction

In pursuit of carbon neutrality, low-carbon operation and environmental protection have become key focus areas in thermal power generation [1]. Among coal-fired utility boilers, circulating fluidized bed (CFB) boilers stand out for their low pollutant emissions and high combustion efficiency, achieved through lower peak temperatures and the recirculation of unburned particles [2]. An important indicator for evaluating the performance of CFB boilers is bed temperature [3], which directly influences combustion stability, heat transfer efficiency, and pollutant reduction. Therefore, accurately predicting and regulating bed temperature is essential for improving the overall performance of CFB boilers.
Performance prediction methods for utility boilers are categorized into mechanism-based and data-driven models. Mechanism-based models [4] rely on physical and chemical principles to simulate combustion, fluidization, and heat transfer processes. Gu et al. [5] used the Multi-Phase Particle-In-Cell (MP-PIC) method to simulate pressurized oxy-fuel coal combustion, demonstrating reduced CO and NOx emissions with increased pressure. Gürel et al. [6] employed computational particle fluid dynamics (CPFD) to study lignite combustion in CFB boilers, showing improved combustion efficiency but higher NOx emissions with increased bed material sphericity. Wu et al. [7] compared the Two-Fluid Model (TFM) and Dense Discrete Phase Model (DDPM) for gas–solid hydrodynamics, finding the DDPM more effective in capturing solid-phase distributions. Huang et al. [8] studied the tri-combustion of coal, biomass, and oil sludge using CPFD, revealing improved combustion but increased NOx emissions under certain blends. Cam et al. [9] optimized air nozzle designs for Turkish lignite boilers with computational fluid dynamics (CFD). Liu et al. [10] used CPFD to analyze a 440 t/h CFB boiler, showing moderate secondary air rates enhanced combustion and reduced NO emissions, while excessive rates caused instability. In summary, mechanism-based models provide valuable insights into fluidized bed operations. However, this kind of model faces limitations such as simplifying assumptions, high computational demands, and challenges in adapting to real-time variations or system changes.
Data-driven modeling offers several advantages over mechanism-based approaches. It can rapidly identify nonlinear relationships and adapt to changing operational conditions [11]. This makes data-driven methods particularly suitable for predicting complex, time-varying parameters such as bed temperature. In the recent years, data-driven modeling with machine learning methods has been widely used in CFB modeling. Kartal et al. [12] developed a deep learning-based ANN model to accurately predict the lower heating value (LHV) of syngas in a CFB gasifier. Ma et al. [13] proposed a modified sequential extreme learning machine (SIOS-ELM) for real-time NOx emission modeling in a 330 MW CFB boiler, improving accuracy and generalization by dynamically adjusting weights and thresholds. Li et al. [14] introduced an adaptive extreme learning machine (A-ELM) with teaching–learning-based optimization (TLBO) to model NOx emissions in a 300 MW CFB boiler, outperforming six other methods in approximation and generalization. Wang et al. [15] used a two-step K-means clustering algorithm to analyze particle clusters in CFB risers, enhancing understanding of gas–solid interactions and cluster evolution. Cui et al. [16] developed a combustion prediction model for S-CO2 CFB boilers using an adaptive gray wolf optimizer and support vector machine (AGWO-SVM), enabling efficient design and operation.
Compared to traditional machine learning methods, deep learning offers superior capability in extracting deep features, resulting in enhanced modeling performance. Adams et al. [17] used deep neural networks and least squares support vector machines to predict SOx and NOx emissions in coal-fired CFB boilers, achieving up to 40% improved accuracy by incorporating dynamic coal and limestone properties. Hong et al. [18] combined Long Short-Term Memory (LSTM) neural networks and a dynamic time warping (DTW) algorithm for real-time risk prediction of bed inventory overturn in pant-leg CFB boilers, enhancing operational safety. Despite progress in dynamic bed temperature modeling for CFB boilers, several challenges remain:
(1)
Data quality issues: As thermal power plants shoulder greater responsibilities for deep load adjustments to achieve carbon neutrality, utility boilers operate under increasingly flexible conditions [19]. This results in dirtier and noisier operating data, with more frequent measurement errors and outliers. These issues compromise data reliability and weaken the robustness of data-driven predictive models.
(2)
Challenges in dynamic modeling: The complexity of factors affecting CFB temperatures requires dynamic models to accurately capture spatial and temporal correlations between variables.
  • Spatial Correlation: Bed temperature is shaped by interactions among the flow field, temperature field, and chemical reaction field, which depend on variables such as primary air (PA) and secondary air (SA) flow distributions, recycled particles from different return feeders, and flue gas recirculation. Capturing these high-dimensional relationships is a great challenge.
  • Temporal Correlation: Bed temperature exhibits thermal inertia and time delay, necessitating models capable of learning long-term dependencies. Current approaches [20,21] often use correlation analysis to identify the time-step-specific variables that are most closely related to the target and then build dynamic models. However, changes in operating load can alter delay times and key time series variables, highlighting the need for deeper exploration of temporal relationships to enhance model accuracy and adaptability.
This study introduces a novel framework for predicting bed temperature in CFB boilers. The framework first collects historical operational data from a Distributed Control System (DCS) database. Subsequently, the Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) algorithm is employed to denoise the data, thereby enhancing their reliability. Key features are then identified using Normalized Mutual Information (NMI) to reduce model complexity. Finally, an iTransformer-based model is developed to effectively capture long-term dependencies in bed temperature dynamics. Compared to Gated Recurrent Unit (GRU), LSTM, and Transformer models, the iTransformer demonstrates superior accuracy and robustness. Comprehensive evaluations validate its effectiveness in both single-step and multi-step predictions, offering a reliable and efficient solution for optimizing boiler operation.

2. System Description and Variable Primary Selection

2.1. System Description

This study takes a 300 MW coal-fired circulating fluidized bed (CFB) boiler from Inner Mongolia Jingtai Power Generation Co., Ltd. in Ordos, China as the subject, with its three-dimensional (3-D) structure shown in Figure 1a. Figure 1b provides a top view of the furnace. There are twelve platen superheater (SH) panels, six reheater (RH) panels, and two water-cooled evaporator panels in the upper furnace. The hot flue gas and particles generated from combustion transfer heat to these heating surfaces and then flows into three steam-cooled cyclones. Each cyclone is equipped with a split-structure return feeder, which recirculates unburnt particles to the fluidized bed for reburning.
The combustion system of the boiler is located in the lower furnace, part of the primary air (PA) is used to fluidize the bed material (i.e., slag particles) through the bottom distributor plate, while the rest transports coal particles into the furnace through eight coal-feeding ports at the front (see Figure 1c). Secondary air (SA) is introduced via two layers of nozzles: eight on both the upper and lower front walls (see Figure 1c), ten on the upper rear wall (see Figure 1d), and six on the lower rear wall (see Figure 1d).

2.2. Variable Analysis and Primary Selection

Furnace bed temperature is directly influenced by multiple factors, including the combustion process, air distribution, fluidization state, and dynamic heat transfer balance within the furnace. In a CFB boiler, parameters such as unit load and oxygen content reflect the balance among coal, working fluid, and air supply, shaping the overall combustion environment. The distribution of coal and air affects the staged combustion and the local air-fuel ratio, which are critical for stable temperature regulation. Additionally, pressures at various points in the furnace represent the fluidization state of the bed, while water-side parameters such as steam pressure and temperature influence the dynamic heat transfer balance within the furnace. The temperature and pressure within the flue gas passage impact the bed temperature through enthalpy–temperature and pressure–flow pathways, highlighting their role in heat and mass transfer dynamics. These variables collectively characterize the complex interactions within the combustion, fluidization, heat transfer, and flue gas flow processes.
Based on the mechanism analysis, a total of 130 variables in 28 categories potentially related to bed temperature in a CFB boiler are first selected, as shown in Table 1, with their ranges determined according to real operational data. Subsequent correlation analysis will be conducted to identify the key factors influencing bed temperature.

3. Methodology

3.1. Algorithm Framework

In this study, we propose a novel algorithm framework to predict the bed temperature of CFB boiler, as shown in Figure 2. The calculation steps are described below.
Step 1: Data collection. Historical operational data, including variables such as unit load, coal feeding rate, air flow rate, and bed pressure, are collected from the DCS database, forming the data basis for developing an accurate predictive model.
Step 2: Data denoising. The data are denoised using the CEEMDAN algorithm, which is a robust method to handle nonlinear and non-stationary signals. This process eliminates noise and outliers from sensor measurements caused by interference or error, ensuring more reliable data and enhancing the robustness of the predictive model.
Step 3: Feature selection. The prediction target (i.e., bed temperature) is influenced by a large number of factors and variables, as shown in Table 1. Identifying the most relevant variables in such a vast dataset is challenging, but critical for improving model prediction accuracy and interpretability. To address this issue, Normalized Mutual Information (NMI) is applied to identify the variables that are most highly correlated with bed temperature. This step reduces model complexity and improves model validity by focusing on the major features.
Step 4: Model Development and performance analysis.
Bed temperature is influenced by factors such as chemical reactions, flow fields, and temperature fields, all of which exhibit strong inertia and time delay. Accurately predicting bed temperature dynamics is challenging, and the advantage of iTransformer model to capture long-term dependencies provides a valuable solution to this problem. Therefore, an iTransformer-based model was developed by individually embedding each feature and modeling the dynamic behavior of bed temperature.
The proposed iTransformer-based model was comprehensively evaluated on its generalization ability and single-step and multi-step prediction accuracies under different training, validation, and testing datasets. It was also compared with other models such as GRU, LSTM, and Transformer. These evaluations highlight the effectiveness of the iTransformer model in predicting bed temperature dynamics with improved accuracy and robustness.

3.2. Data Denoising

The data denoising process eliminates noise and outliers from sensor measurements caused by interference or error, ensuring more reliable data and enhancing the robustness of modeling. In this study, CEEMDAN is employed to perform data denoising.

3.2.1. Empirical Mode Decomposition

The empirical mode decomposition (EMD) algorithm is a method used to decompose a signal into a set of intrinsic mode functions (IMFs) and a residual [22]. The steps of EMD are given below.
Step 1: Identify local extrema. Given an original signal s(t), we determine its set of local maxima {tmax,i, s(tmax,i)} and local minima {tmin,i, s(tmin,i)}.
Step 2: Construct envelopes. Based on the local maxima and minima sets, we construct the upper envelope emax(t) and lower envelope emin(t) using cubic spline interpolation.
Step 3: Calculate the mean envelope. We calculate the mean of the upper and lower envelopes m(t) = (emax(t) + emin(t))/2.
Step 4: Extract the component. We subtract the mean envelope m(t) from the original signal s(t) to obtain the detail component h(t) = s(t) − m(t).
Step 5: Check IMF criteria. If h(t) satisfies criterion 1 (the number of zero crossings and extrema differ by at most one) and criterion 2 (the upper and lower envelopes have zero means), we designate h(t) as the first IMF, denoted as c1(t) (i.e., Max_IMF in Figure 2). Otherwise, we repeat steps 1–4 with h(t) as the new signal until an IMF is obtained.
Step 6: Subtract IMF and iterate calculation. We subtract the extracted IMF from the original signal to obtain a residual signal r1(t) = s(t) − c1(t) and then replace s(t) with r1(t) and repeat steps 1~5 to extract the next IMF c2(t). We repeat this process until the residual signal rn(t) can be treated as noise or becomes a monotonic function.
Step 7: Complete decomposition. The original signal can be decomposed into several IMFs and one residual s t = i = 1 n c i t + r n t .

3.2.2. CEEMDAN

CEEMDAN is an improved EMD algorithm that incorporates adaptive noise into the decomposition process [23], as shown in Figure 2. The key steps are summarized below.
Step 1: We add white noise series wi(t) with standard normal distribution N(0,1) to the original signal s(t): si(t) = s(t) + wi(t), i = 1,2,…, N.
Step 2: The signal with white noise si(t) is decomposed using the EMD algorithm to obtain the first IMF (i.e., Max_IMF in Figure 2).
Step 3: We calculate the average of Max_IMF I M F 1 ¯ t = 1 N i = 1 N M a x _ I M F 1 , i t and the first stage residual r 1 t = s t I M F 1 ¯ t .
Step 4: We continue to add white noise wi(t) to the residual r1(t) and perform step 2 and step 3 to obtain I M F 2 ¯ t = 1 N i = 1 N M a x _ I M F 2 , i t and the second stage residual r 2 t = r 1 t I M F 2 ¯ t .
Step 5: The isolating process is repeated k times until the residual becomes a monotony function and cannot be decomposed by EMD. The original and denoised signals can be written as s t = i = 1 k I M F i ¯ t + r k t and d e n o i s e _ s i g n a l t = j S I M F j ¯ t + r k t , respectively. S is the set of the selected modes and is discussed in the subsequent chapter.

3.2.3. Denoising Strategy

Our denoising strategy preserves the main data trend while effectively eliminating gross errors, with each variable analyzed individually for optimal performance. Using CEEMDAN, the original signal is decomposed into intrinsic mode functions (IMFs), where low-energy IMFs typically represent noise, and high-energy IMFs capture essential data patterns. To maintain data integrity, we systematically analyze the energy distribution of IMFs, removing only those that contribute to noise while ensuring the reconstructed signal remains consistent with the original trend. The components are progressively merged from high to low energy, with the fusion process halted when further removal starts to distort the dominant trend. This threshold, defined as the point just before the trend begins to deviate significantly, represents the minimum number of components to retain. The selected components are then used to reconstruct the denoised signal, ensuring noise reduction without compromising critical data patterns across different operating conditions.

3.3. Feature Selection

For the prediction target (i.e., bed temperature, Y in Table 1), the NMI algorithm is used to identify the most relevant features from X1~X130. The mutual information value [24] between feature X and target Y can be calculated by Equation (1):
M I X , Y = x X y Y P X Y x , y log P X Y x , y P X x P Y y
where P X Y x , y is the joint probability density function of X and Y. P X x and P Y y are the marginal probability distribution of X and Y, respectively.
The NMI value [25] between feature X and target Y can be calculated by Equation (2):
N M I X , Y = 2 M I X , Y H X H Y
where H X and H Y are the entropy of X and Y and can be calculated by Equation (3) and Equation (4), respectively.
H X = x X P x log P x
H Y = y Y P y log P y

3.4. Dynamic Modeling Based on iTransformer

iTransformer is an enhanced Transformer [26,27] variant and excels in capturing long-range dependencies and hierarchical patterns. The model framework of iTransformer [28] is shown in Figure 3. iTransformer achieves greater efficiency and scalability through an innovative attention mechanism and optimized architecture. The calculation steps are summarized as follows.
Step 1: Input embedding based on variables. The input of the embedding process is defined as X T × M , T is the time step length of the input, and M is the number of features selected via NMI algorithm. Compared with Transformer, iTransformer adopts an alternative approach by embedding each series individually into a univariate token:
H 0 = M L P X . transpose
where H 0 M × D represents variate token set. Using Multi-Layer Perceptron (MLP), the input length for each feature changes from T (length of time step) to D (token dimension). X.transpose means the matrix transposition operation for X.
Step 2: Run Trm blocks n times. Each time, the steps given below are performed.
Multi-head self-attention (Self_Attn) with residual connections and layer normalization (LayerNorm) is first applied on variate tokens:
H l 1 = L a y e r N o r m H l 1 + S e l f _ A t t n H l 1
Secondly, feed-forward network (Feed_Forward) with residual connections and layer normalization is conducted:
H l = L a y e r N o r m H l 1 + F e e d _ F o r w a r d H l 1
where l = 1 , 2 , , n . H l M × D is the variate token set of the l-th iteration.
Step 3: Project the variate token set to the predicted series using MLP:
Y ^ = M L P H n . t r a n s p o s e

4. Results and Discussion

4.1. Impact of Data Denoising

When performing deep and flexible load adjustments for the power grid, the operating conditions of CFB boilers are highly complex and variable. This leads to the presence of significant noise in the collected operational data. Noise refers to irrelevant, erroneous, or inaccurate information within the data which can negatively impact the training and prediction performance of models, ultimately leading to reduced model accuracy. To ensure the high quality of the dataset, denoising is essential. Denoising involves decomposing the data signal into multiple modal components, such as low-frequency and high-frequency components, and removing the components with the lowest energy. The remaining components are then reconstructed into a new dataset with reduced noise and improved quality. Typically, low-energy components represent noise in the signal, and removing these components can effectively reduce noise.
During the denoising process, the algorithm decomposes the original signal into multiple modal components, as shown in Figure 4. The energy of each component represents the signal strength and amplitude information contained in that component. The greater the energy of a component, the more it contributes to the overall signal, and it often represents the primary component or key feature. Conversely, components with lower energy contribute less and often represent noise or irrelevant variables. Figure 5 shows the energy values of these components. Determining the components to be removed is a critical aspect of data denoising. Removing too many modal components may discard useful information from the signal, causing the denoised data signal curve to deviate significantly from the original signal curve.
Figure 6, Figure 7, Figure 8 and Figure 9 show the comparison between the reconstructed and original signals for four different features. When denoising the “coal feeding rate” feature, reconstructing the data after removing more than six components resulted in a trend that diverges notably from the original signal, indicating that useful information is discarded. To effectively remove noise without altering the signal trend, six of the lowest-energy components are removed. Similarly, for the “SA flow rate” feature, the reconstructed curve deviates from the original signal curve when six components are removed, so five components are chosen for removal instead. Conversely, removing too few components leaves significant noise in the signal, failing to improve data quality. For the “outlet oxygen content” feature, the reconstructed data with four components removed still contains noticeable noise, while removing five components effectivity mitigates the noise and retains the signal trend.
As can be seen from Figure 6, the effect of denoising on the coal feeding rate signal is the best when the six components with the lowest energy are removed. When 3–5 components are removed, some noise is still not completely eliminated in the reconstructed signal. After more than six components are removed, the reconstructed signal has obvious signal trend loss, and the reconstructed signal has a deviation from the original signal trend.
As can be seen from Figure 7, the effect of denoising on the dilution water flow rate signal is the best when the four components with the lowest energy are removed. After removing 2–3 components, there is still noise in the reconstructed signal, which will affect subsequent processing and prediction. After removing more than four components, the reconstructed signal will obviously lose the signal trend, and the reconstructed signal will lose the original signal.
As shown in Figure 8, the effect of denoising on the outlet oxygen content signal is the best when the five components with the lowest energy are removed. The noise in the reconstructed signal is still incomplete when 3–4 components are removed, which will affect subsequent work. After the removal of six components, the reconstructed signal has a tendency to be lost, and when more components are removed, the reconstructed signal obviously loses more information contained in the original signal.
It can be seen from Figure 9 that the effect of denoising is the best when the five components with the lowest energy are removed from the SA flow rate signal. There is still incomplete noise in the reconstructed signal when the 3–4 components are removed, and the effect of denoising is not the best. When six components are removed, the reconstructed signal trend is lost, and when more components are removed, the reconstructed signal curve significantly shifts more than the original signal.
The reconstructed signal curves for the different features after precise denoising are depicted in Figure 10. For the “dilution water flow rate” feature, data denoising is much more challenging since data variation is more complex, with significant fluctuations and mixed noise. To ensure there is no difference between the reconstructed and original data, four components are removed for this feature.

4.2. Analysis of Feature Selection

CFB boilers involve numerous parameters, resulting in high-dimensional field data. This high-dimensional dataset contains redundant features and features with weak correlations to the target. To prevent these irrelevant or weakly relevant features from causing model overfitting, and to avoid high computational costs and dimensionality issues associated with excessive data dimensions, it is necessary to perform feature selection. This process identifies features with strong correlations to the target and uses them as inputs for the model. In this way, the model can extract meaningful patterns from the input features, enabling accurate and efficient predictions.
The correlation between each boiler feature and the target variable was determined by the Normalized Mutual Information (NMI) values. In Figure 11, the features with NMI values greater than 0.8 have a strong correlation with the target variable (i.e., the boiler bed temperature). The results indicate that the magnitude and distribution of the Coal Seeding Air Volume play a crucial role in influencing bed temperature. The Coal Seeding Air Volume refers to the airflow responsible for transporting coal particles into the boiler. In the CFB boiler examined in this study, there are four coal feeders, each connected to two coal conveying pipelines, resulting in a total of eight pipelines labeled 11 to 18. Each pipeline is equipped with four air injection ports. For instance, pipeline 11 contains four ports, 11/11, 11/12, 11/13, and 11/14, which correspond to the 11/11 Coal Seeding Air Volume, 11/12 Coal Seeding Air Volume, 11/13 Coal Seeding Air Volume, and 11/14 Coal Seeding Air Volume, as shown in Figure 11.
Figure 12 shows the features with NMI values less than 0.8, with a red threshold line of 0.665. There is a distinct separation between features above the threshold (orange bars) and those below it (blue bars). Features above the threshold (0.665) have NMI values closer to 0.8, indicating that they are more strongly correlated with the target variable and are suitable as input features to the model.
Ultimately, the weakly correlated features corresponding to the blue bars in Figure 12 are discarded, while the strongly correlated features represented by the orange bars are selected as input features for the predictive model.

4.3. Impact of Dataset Splits

In Figure 13, the unit load refers to the power supply load of the CFB thermal power plant, which determines the boiler’s operating conditions. As shown in the figure, the dataset used in this study includes various operating scenarios, such as increasing load, decreasing load, and steady load conditions. Therefore, the data in this study can be used to validate the bed temperature prediction performance under different operating conditions. To validate the performance of the proposed prediction model, the experimental dataset is divided into training, validation, and testing sets in the ratios. If the size of the training set decreases and the model’s performance significantly deteriorates, this indicates that the model achieves high prediction accuracy only for the training data, while its generalization ability remains weak. Therefore, to investigate the impact of training data volume on the model’s generalization ability, we studied two dataset split ratios: 7:1:2 and 6:2:2 for training, validation, and testing, respectively. The 7:1:2 split, or even higher training data ratios, is a commonly adopted data partitioning method. Based on this (dataset split ratio 7:1:2), we further explored the model’s generalization ability under a lower training data ratio of 6:2:2.
The impact of different training dataset lengths on model prediction performance is compared. Figure 14 shows the distribution of prediction errors under these two dataset splits. The results indicate that the different dataset splits have a minimal impact on prediction errors. Regardless of the dataset splits, the model’s performance across the training, validation, and testing sets remains consistent, which demonstrates the model has a strong stability and an excellent generalization ability. It is evident in Figure 14 that the proposed iTransformer model performs exceptionally well under both dataset splits, achieving an R2 value of 0.99. This confirms that the model has a strong ability to capture the trends and dependencies of the input features independent of the changes in the dataset splits.
For these two data splits, the MAE changes by 0.02 mg/m3, 0.04 mg/m3, and 0.05 mg/m3 in the training, validation, and testing datasets, respectively. The RMSE changes by 0.05 mg/m3, 0.13 mg/m3, and 0.09 mg/m3 in the training, validation, and testing datasets, respectively. For metrics such as MAE and RMSE, slight fluctuations are observed due to the changes in the dataset, but they stay low and have no significant impact. The prediction model maintains consistent error levels across different datasets, proving its robust predictive performance and outstanding generalization capability.

4.4. Comparison with Other Models

Figure 15 shows the prediction effect results of the iTransformer model and other comparison models on the test set. It can be seen from Figure 15 that, on the whole, the iTransformer model performs well in terms of the trend of prediction and the change trend of measured values, while other models show obvious deviations. This shows that the forecast result of the iTransformer model is closer to the real value, and the change trend is more in line with the actual situation. Although all the models achieved a better prediction effect in the training set, the prediction effects of the comparison models were significantly worse for the new test set, which also indicated that the established iTransformer model had a strong generalization ability.
The performance of the models was evaluated based on three metrics: R2, MAE, and RMSE. The results are shown in Figure 16. The iTransformer model demonstrates excellent predictive performance, achieving R2 values of 0.99 across the training, validation, and testing sets. iTransformer significantly outperforms the comparative models LSTM (0.97, 0.88, 0.87) and GRU (0.96, 0.88, 0.88) and surpasses the Transformer model with the same architecture (0.98, 0.96, 0.96).
In CFB boiler operation, the bed temperature exhibits inertia, with significant long-term dependencies in the data. LSTM and GRU models fail to accurately capture these dependencies because they gradually forget previously learned information during the training process. Specifically speaking, LSTM and GRU utilize gating mechanisms to mitigate the vanishing gradient problem, but they still suffer from memory loss over long sequences. As a result, they struggle to capture long-term dependencies in CFB boiler operation data, leading to less accurate predictions. In contrast, the iTransformer and Transformer models incorporate multi-head attention mechanisms, enabling them to focus on the global information within the input data. In terms of MAE, iTransformer consistently achieves lower MAE values across all datasets compared to other models and thereby has smaller absolute errors between predicted and actual values. For the testing set, the MAE of iTransformer is 0.41 mg/m3, significantly lower than those of LSTM (20.19 mg/m3), GRU (18.8 mg/m3), and Transformer (11.77 mg/m3). Similarly, with respect to RMSE, iTransformer exhibits substantially lower values than the other models, further confirming its superior predictive accuracy and performance.
iTransformer’s advantage over Transformer lies in its treatment of variable relationships. Transformer treats multiple variables at the same time point as a single temporal token and applies attention mechanisms to derive correlations, while significant differences between variables can negatively impact its predictive performance. In contrast, the internal mechanism of iTransformer plays a crucial role in its predictive performance. By treating each variable as an independent token with a full sequence, the model effectively captures both intra-variable dependencies and inter-variable correlations. This structural design reduces interference between variables and ensures a more refined feature extraction process. The attention mechanism in iTransformer is thereby better utilized to focus on relevant long-term dependencies, leading to more stable and accurate predictions. Comparisons on the three metrics confirm that the iTransformer model exhibits outstanding predictive performance and a robust generalization ability.

4.5. Performance in Multi-Step Prediction

Figure 17 compares the multi-step prediction accuracy between the Transformer and iTransformer models. In the experiment, each time step corresponds to 30 s. It is necessary to explain the errors in Figure 17. The error at each point in the figure represents the error of the output sequence with that length. For example, the errors at x = 48 and x = 96 do not correspond to the errors at the 48th and 96th time steps, but rather to the errors of the output sequences predicted for the next 48 and 96 time steps, respectively.
For the testing dataset, as the number of time steps increases, the prediction errors of both models grow rapidly, resulting in reduced R2. This can be attributed to two primary factors: First, Transformer-based models have already effectively handled long-term dependencies. With sufficient data and model parameters, they can achieve excellent long-range forecasting performance, similar to GPT and DeepSeek. However, in industrial applications, due to limitations in training data volume and hardware constraints such as GPUs, the model’s parameter size is restricted, leading to increased long-term prediction errors. Secondly, despite the self-attention mechanism enabling Transformer-based models to capture long-range dependencies, in time series forecasting, distant historical data often has a weaker influence on far-future predictions compared to more recent data. As a result, Transformer-based models tend to prioritize learning short-term patterns, while their ability to accurately capture long-term trends is relatively weaker, leading to a decline in predictive accuracy over longer horizons. However, iTransformer consistently achieves higher accuracy than Transformer. Notably, when predicting the next 48 time steps (24 min), iTransformer maintains an R2 value above 0.95, while the R2 of Transformer drops below 0.90. For iTransformer, the prediction error becomes evident after predicting the next 12 steps (6 min), so iTransformer is reliable in predicting bed temperature changes within the next 6 min with high accuracy in practical applications.

5. Conclusions

This study proposed a CEEMDAN-NMI–iTransformer framework for predicting bed temperature in utility CFB boilers, addressing challenges in data quality, feature selection, and long-term dependency modeling. The model was validated using real operational data from a 300 MW CFB boiler and outperformed traditional deep learning models in single-step and multi-step predictions. The key conclusions are as follows:
(1)
The CEEMDAN algorithm significantly enhances data quality for predictive modeling. For each feature, the appropriate components are analyzed and selected to preserve the overall signal trend while effectively removing substantial noise points.
(2)
The NMI-based feature selection reduces complexity while preserving accuracy. From 130 operational variables, the most relevant features with NMI values greater than 0.665 were chosen, enhancing computational efficiency. This process also emphasizes key factors, such as the magnitude and distribution of the coal seeding air, that play a crucial role in the system’s performance.
(3)
The iTransformer model outperforms LSTM, GRU, and Transformer, achieving an R2 of 0.99 for single-step predictions across different training dataset sizes, and maintaining an R2 above 0.95 for 24 min forecasts, demonstrating superior capability in capturing long-term dependencies.
Our future work will enhance the iTransformer model with adaptive attention mechanisms to dynamically adjust feature importance in real time, improving its ability to handle nonlinear fluctuations and sudden operational changes in CFB boilers, further strengthening prediction accuracy, robustness, and generalization.

Author Contributions

Conceptualization, Q.L. and H.M.; methodology, Q.L., C.Y. and H.M.; investigation, Q.L., G.W. and Q.B.; software, G.W., X.L. and L.W.; data curation, G.W., Q.B. and W.L.; validation, X.L., L.W. and F.S.; formal analysis, Q.L., X.L. and W.L.; writing—original draft preparation, Q.L.; writing—review and editing, Q.L. and H.M.; visualization, Q.B. and C.Y.; supervision, H.M. and F.S.; resources, G.W. and F.S.; funding acquisition, H.M.; project administration, H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Beijing Jingneng Power Co., Ltd.

Data Availability Statement

The dataset is available from the authors upon request.

Acknowledgments

Special thanks are also extended to the editors and reviewers who made valuable comments to improve this paper.

Conflicts of Interest

Authors Qianyu Li, Guanglong Wang, Xian Li, Qing Bao and Wei Li were employed by the company Inner Mongolia Jingtai Power Generation Co., Ltd. Author Qianyu Li was employed by the company Beijing Jingneng Power Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhu, Y.; Yu, C.; Jin, W.; Shi, L.; Chen, B.; Xu, P. Mechanism-enhanced data-driven method for the joint optimization of boiler combustion and selective catalytic reduction systems considering gas temperature deviations. Energy 2024, 291, 130432. [Google Scholar] [CrossRef]
  2. Huang, X.; Bai, Z.; Zhu, X.; Wang, S.; Mu, L.; Gong, L. Gas-Solid Flow, Combustion Characteristics, and Gas Emissions in a 75 t/h CFB Boiler Based on the CPFD Method. J. Therm. Sci. 2025, 34, 323–336. [Google Scholar] [CrossRef]
  3. Hong, F.; Long, D.; Chen, J.; Gao, M. Modeling for the bed temperature 2D-interval prediction of CFB boilers based on long-short term memory network. Energy 2020, 194, 116733. [Google Scholar] [CrossRef]
  4. Yu, C.; Xiong, W.; Ma, H.; Zhou, J.; Si, F.; Jiang, X.; Fang, X. Numerical investigation of combustion optimization in a tangential firing boiler considering steam tube overheating. Appl. Therm. Eng. 2019, 154, 87–101. [Google Scholar] [CrossRef]
  5. Gu, J.; Shao, Y.; Zhong, W. 3D simulation on pressurized oxy-fuel combustion of coal in fluidized bed. Adv. Powder Technol. 2020, 31, 2792–2805. [Google Scholar] [CrossRef]
  6. Gürel, B.; Dolgun, G.K.; İpek, O.; Keçebaş, A. Combustion characteristics of low-quality lignite for different bed material sphericities in a circulating fluidized bed boiler: A numerical study. Particuology 2024, 90, 364–382. [Google Scholar] [CrossRef]
  7. Wu, Y.; Liu, D.; Hu, J.; Ma, J.; Chen, X. Comparative study of two fluid model and dense discrete phase model for simulations of gas–solid hydrodynamics in circulating fluidized beds. Particuology 2021, 55, 108–117. [Google Scholar] [CrossRef]
  8. Huang, X.; Jin, X.; Dong, L.; Li, R.; Yang, K.; Li, Y.; Deng, L.; Che, D. CPFD numerical study on tri-combustion characteristics of coal, biomass and oil sludge in a circulating fluidized bed boiler. J. Energy Inst. 2024, 113, 101550. [Google Scholar] [CrossRef]
  9. Çam, M.M.; Soyhan, H.S.; Al Qubeissi, M.; Çelik, C. Designing a new bell-type primary air nozzle for large-scale circulating fluidized bed boilers. Fuel 2023, 335, 127065. [Google Scholar] [CrossRef]
  10. Liu, H.; Sun, H.; Bi, Y.; Jia, C.; Zhang, L.; Li, Y.; Qin, H.; Wang, Q. Effect of secondary air on NO emission in a 440 t/h circulating fluidized bed boiler based on CPFD method. Particuology 2023, 83, 18–31. [Google Scholar] [CrossRef]
  11. Zhu, Y.; Yu, C.; Fan, W.; Yu, H.; Jin, W.; Chen, S.; Liu, X. A novel NOx emission prediction model for multimodal operational utility boilers considering local features and prior knowledge. Energy 2023, 280, 128128. [Google Scholar] [CrossRef]
  12. Kartal, F.; Özveren, U. A deep learning approach for prediction of syngas lower heating value from CFB gasifier in Aspen plus®. Energy 2020, 209, 118457. [Google Scholar] [CrossRef]
  13. Ma, Y.; Niu, P.; Yan, S.; Li, G. A modified online sequential extreme learning machine for building circulation fluidized bed boiler’s NOx emission model. Appl. Math. Comput. 2018, 334, 214–226. [Google Scholar] [CrossRef]
  14. Li, X.; Niu, P.; Li, G.; Liu, J. An adaptive extreme learning machine for modeling NOx emission of a 300 MW circulating fluidized bed boiler. Neural Process. Lett. 2017, 46, 643–662. [Google Scholar] [CrossRef]
  15. Wang, T.; Deng, A.; He, Y.; Wu, B.; Gao, R.; Tang, T. Artificial intelligence-based approach for cluster identification in a CFB riser. Chem. Eng. Sci. 2023, 268, 118379. [Google Scholar] [CrossRef]
  16. Cui, Y.; Zou, Y.; Jiang, S.; Zhong, W. Combustion Characteristic Prediction of a Supercritical CO2 Circulating Fluidized Bed Boiler Based on Adaptive GWO-SVM. ACS Omega 2023, 8, 10160–10175. [Google Scholar] [CrossRef]
  17. Adams, D.; Oh, D.H.; Kim, D.W.; Lee, C.H.; Oh, M. Prediction of SOx–NOx emission from a coal-fired CFB power plant with machine learning: Plant data learned by deep neural network and least square support vector machine. J. Clean. Prod. 2020, 270, 122310. [Google Scholar] [CrossRef]
  18. Hong, F.; Chen, J.; Zhang, Z.; Wang, R.; Gao, M. Time series risk prediction based on LSTM and a variant DTW algorithm: Application of bed inventory overturn prevention in a pant-leg CFB boiler. IEEE Access 2020, 8, 156634–156644. [Google Scholar] [CrossRef]
  19. Chen, S.; Yu, C.; Zhu, Y.; Fan, W.; Yu, H.; Zhang, T. NOx formation model for utility boilers using robust two-step steady-state detection and multimodal residual convolutional auto-encoder. J. Taiwan Inst. Chem. Eng. 2024, 155, 105252. [Google Scholar] [CrossRef]
  20. Tang, Z.; Wang, S.; Chai, X.; Cao, S.; Ouyang, T.; Li, Y. Auto-encoder-extreme learning machine model for boiler NOx emission concentration prediction. Energy 2022, 256, 124552. [Google Scholar] [CrossRef]
  21. Seol, Y.; Lee, S.; Lee, J.; Kim, C.W.; Bak, H.S.; Byun, Y.; Yoon, J. An Interpretable Time Series Forecasting Model for Predicting NOx Emission Concentration in Ferroalloy Electric Arc Furnace Plants. Mathematics 2024, 12, 878. [Google Scholar] [CrossRef]
  22. Lin, Y.; Yan, Y.; Xu, J.; Liao, Y.; Ma, F. Forecasting stock index price using the CEEMDAN-LSTM model. N. Am. J. Econ. Financ. 2021, 57, 101421. [Google Scholar] [CrossRef]
  23. Zhang, C.; Zhao, Y.; Zhao, H. A novel hybrid price prediction model for multimodal carbon emission trading market based on CEEMDAN algorithm and window-based XGBoost approach. Mathematics 2022, 10, 4072. [Google Scholar] [CrossRef]
  24. Estévez, P.A.; Tesmer, M.; Perez, C.A.; Zurada, J.M. Normalized mutual information feature selection. IEEE Trans. Neural Netw. 2009, 20, 189–201. [Google Scholar] [CrossRef]
  25. Wen, Q.; Tu, X.; Zhou, L.; Singh, V.P.; Chen, X.; Lin, K. Mutual-information of meteorological-soil and spatial propagation: Agricultural drought assessment based on network science. Ecol. Indic. 2025, 170, 113004. [Google Scholar] [CrossRef]
  26. Kim, D.K.; Kim, K. A convolutional transformer model for multivariate time series prediction. IEEE Access 2022, 10, 101319–101329. [Google Scholar] [CrossRef]
  27. Wu, S.; Xiao, X.; Ding, Q.; Zhao, P.; Wei, Y.; Huang, J. Adversarial sparse transformer for time series forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17105–17115. [Google Scholar]
  28. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. itransformer: Inverted transformers are effective for time series forecasting. arXiv 2023, arXiv:2310.06625. [Google Scholar]
Figure 1. Structural diagram of the studied CFB boiler.
Figure 1. Structural diagram of the studied CFB boiler.
Processes 13 00816 g001
Figure 2. Model framework of the present study.
Figure 2. Model framework of the present study.
Processes 13 00816 g002
Figure 3. Model framework of iTransformer.
Figure 3. Model framework of iTransformer.
Processes 13 00816 g003
Figure 4. Decomposition of original signals using CEEMDAN.
Figure 4. Decomposition of original signals using CEEMDAN.
Processes 13 00816 g004aProcesses 13 00816 g004b
Figure 5. Energies of decomposed IMFs for different signals.
Figure 5. Energies of decomposed IMFs for different signals.
Processes 13 00816 g005
Figure 6. Construction of coal feeding rate by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Figure 6. Construction of coal feeding rate by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Processes 13 00816 g006
Figure 7. Construction of dilution water flow rate by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Figure 7. Construction of dilution water flow rate by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Processes 13 00816 g007
Figure 8. Construction of oxygen content by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Figure 8. Construction of oxygen content by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Processes 13 00816 g008
Figure 9. Construction of SA flow rate at B side by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Figure 9. Construction of SA flow rate at B side by IMFs of different energies. The blue curve represents the original curve, the orange curve represents the denoised curve, and the red box highlights the issues that need attention during local denoising.
Processes 13 00816 g009
Figure 10. Effects of denoising on different signals. The blue curve represents the original curve and the orange curve represents the denoised curve.
Figure 10. Effects of denoising on different signals. The blue curve represents the original curve and the orange curve represents the denoised curve.
Processes 13 00816 g010
Figure 11. Features with an NMI value greater than 0.8.
Figure 11. Features with an NMI value greater than 0.8.
Processes 13 00816 g011
Figure 12. Features with an NMI value less than 0.8. The red dashed line represents the threshold of the NMI value, and indicators with NMI values above the threshold are used as features for modeling.
Figure 12. Features with an NMI value less than 0.8. The red dashed line represents the threshold of the NMI value, and indicators with NMI values above the threshold are used as features for modeling.
Processes 13 00816 g012
Figure 13. Dataset for different splitting ratios. The red dashed lines represent the boundaries between the training set, validation set, and test set data.
Figure 13. Dataset for different splitting ratios. The red dashed lines represent the boundaries between the training set, validation set, and test set data.
Processes 13 00816 g013aProcesses 13 00816 g013b
Figure 14. Prediction accuracies for different dataset splitting ratios.
Figure 14. Prediction accuracies for different dataset splitting ratios.
Processes 13 00816 g014aProcesses 13 00816 g014b
Figure 15. Overall prediction of the test set.
Figure 15. Overall prediction of the test set.
Processes 13 00816 g015
Figure 16. Comparison of different models, (a) the R2 values of the four algorithms on three datasets, (b) the MAE values of the four algorithms on three datasets, (c) the RMSE values of the four algorithms on three datasets.
Figure 16. Comparison of different models, (a) the R2 values of the four algorithms on three datasets, (b) the MAE values of the four algorithms on three datasets, (c) the RMSE values of the four algorithms on three datasets.
Processes 13 00816 g016
Figure 17. Performance in long-term prediction, (a) the R2 values of multi-step predictions for the iTransformer and Transformer models, (b) the MAE values of multi-step predictions for the iTransformer and Transformer models, (c) the MRSE values of multi-step predictions for the iTransformer and Transformer models.
Figure 17. Performance in long-term prediction, (a) the R2 values of multi-step predictions for the iTransformer and Transformer models, (b) the MAE values of multi-step predictions for the iTransformer and Transformer models, (c) the MRSE values of multi-step predictions for the iTransformer and Transformer models.
Processes 13 00816 g017
Table 1. Variables related to bed temperature.
Table 1. Variables related to bed temperature.
Feature NameSymbolUnitVariable NumberValues Range
Unit LoadX1MW1[73.31, 311.20]
Fresh Steam PressureX2MPa1[8.54, 17.20]
Fresh Steam TemperatureX3°C1[501.08, 542.65]
Back PressureX4kPa1[5.26, 26.83]
Steam Pressure before the Primary RHX5MPa1[0.80, 3.72]
Steam Temperature in the RH Hot SectionX6°C1[507.00, 543.30]
Coal Feeding RateX7t/h1[48.07, 245.35]
Coal Feeding Rate at Left and Right SidesX8~X9t/h2[20.40, 130.82]
Total Air Flow RateX10km3/h1[233.69, 806.51]
Ratio of Air to CoalX11 1[2.67, 11.79]
Total PA Flow RateX12km3/h1[223.56, 348.87]
Total SA Flow RateX13km3/h1[0.00, 500.96]
PA and SA Flow Rate at Left and Right SidesX14~X17km3/h4[0.00, 278.54]
PA and SA Temperature at Air Preheater OutletX18~X21°C4[142.18, 232.98]
PA and SA Pressure at Air Preheater OutletX22~X25kPa4[0.65, 13.36]
Pressure at Upper FurnaceX26~X27Pa2[−443.02, 462.12]
Temperature at Furnace OutletX28~X29°C2[493.48, 914.22]
Furnace Differential PressureX30~X31kPa2[0.11, 1.12]
SA Damper OpeningX32~X71%40[2.21, 100.69]
Boiler EfficiencyX72%1[0.00, 93.41]
Bed PressureX73~X75kPa3[1.26, 10.23]
Pressure at Air ChamberX76kPa1[8.28, 13.06]
Temperature at Cyclone InletX77~X79°C3[581.53, 1046.54]
Air Flow Rate at Coal Distributor OutletX80~X111m3/h32[852.43, 3272.22]
Coal Feeding Rate at Feeder OutletX112~X119t/h8[−0.04, 51.18]
Temperature of Coal Feeder ChuteX120~X127°C8[30.83, 140.55]
Temperature at Secondary Fan OutletX128~X129°C2[7.57, 39.33]
Oxygen content at SCR outletX130%1[0.01, 9.39]
Bed Temperature (Object)Y°C1[694.84, 957.91]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Q.; Wang, G.; Li, X.; Yu, C.; Bao, Q.; Wei, L.; Li, W.; Ma, H.; Si, F. Robust Dynamic Modeling of Bed Temperature in Utility Circulating Fluidized Bed Boilers Using a Hybrid CEEMDAN-NMI–iTransformer Framework. Processes 2025, 13, 816. https://doi.org/10.3390/pr13030816

AMA Style

Li Q, Wang G, Li X, Yu C, Bao Q, Wei L, Li W, Ma H, Si F. Robust Dynamic Modeling of Bed Temperature in Utility Circulating Fluidized Bed Boilers Using a Hybrid CEEMDAN-NMI–iTransformer Framework. Processes. 2025; 13(3):816. https://doi.org/10.3390/pr13030816

Chicago/Turabian Style

Li, Qianyu, Guanglong Wang, Xian Li, Cong Yu, Qing Bao, Lai Wei, Wei Li, Huan Ma, and Fengqi Si. 2025. "Robust Dynamic Modeling of Bed Temperature in Utility Circulating Fluidized Bed Boilers Using a Hybrid CEEMDAN-NMI–iTransformer Framework" Processes 13, no. 3: 816. https://doi.org/10.3390/pr13030816

APA Style

Li, Q., Wang, G., Li, X., Yu, C., Bao, Q., Wei, L., Li, W., Ma, H., & Si, F. (2025). Robust Dynamic Modeling of Bed Temperature in Utility Circulating Fluidized Bed Boilers Using a Hybrid CEEMDAN-NMI–iTransformer Framework. Processes, 13(3), 816. https://doi.org/10.3390/pr13030816

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop