Next Article in Journal
Optimization of Artificial Intelligence Algorithm Selection: PIPRECIA-S Model and Multi-Criteria Analysis
Next Article in Special Issue
Generator-Level Transient Stability Assessment in Power System Based on Graph Deep Learning with Sparse Hybrid Pooling
Previous Article in Journal
Exploiting Personalized Observation Frequency for Proportional Integral Derivative-Based Diabetes Management
Previous Article in Special Issue
Research on Multi-Microgrid Electricity–Carbon Collaborative Sharing and Benefit Allocation Based on Emergy Value and Carbon Trading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Surrogate-Assisted Intelligent Adaptive Generation Framework for Cost-Effective Coal Blending Strategy in Thermal Power Units

1
Artificial Intelligence Research Center, Information Center of China Building Materials Industry, Beijing 100037, China
2
School of Economics and Management, North China Electric Power University, Beijing 102206, China
3
School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China
4
Economic and Technological Research Institute, State Grid Fujian Electric Power Co., Ltd., Fuzhou 350013, China
5
Institute of Artificial Intelligence, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(3), 561; https://doi.org/10.3390/electronics14030561
Submission received: 4 November 2024 / Revised: 23 January 2025 / Accepted: 28 January 2025 / Published: 30 January 2025

Abstract

:
The coal cost of coal-fired units accounts for more than 70% of the total power generation cost. In addition to determining coal costs, coal blending strategies (CBS) significantly impact various types of costs, such as pollutant removal and emissions. To address these issues, we propose a framework for generating cost-effective CBS. The framework includes a unit output condition recognition module (UOCR) that enables the adaptive classification of output conditions based on historical operation datasets, performing intelligent condition recognition with Imitator and pre-trained image classification models using blending strategies and unit parameters as inputs. The cost-effective strategy generation module (CESG) employs a surrogate model to evaluate the economic viability of strategies in terms of coal and environmental costs, among other factors. It also employs UOCR as another surrogate model to validate strategy feasibility. Cost-effective strategies are generated via a population-based metaheuristic algorithm. In the case study, the UOCR achieved an average training accuracy of 96.64%, and the generated cost-effective strategies reduced costs by an average of 3.37% compared to currently implemented strategies.

1. Introduction

Thermal power generation, predominantly coal-fired, plays a crucial role as the “ballast” in China’s electricity supply structure, accounting for 66.3% of the nation’s total power generation as of 2023 [1]. Amid increasing grid dispatch volatility and pressures from electricity market bidding, cost control in coal-fired power units faces new challenges. Within the cost composition of coal-fired units, coal expenses alone account for over 70% of total production costs. Coal blending is a technique widely applied in coal-fired units to optimize combustion efficiency, reduce power generation costs, and meet environmental standards by mixing coals with varying prices and qualities in specific ratios. This technique directly determines coal production costs and significantly impacts pollutant removal and emissions-related processes. Therefore, improving the coal blending strategy remains a core priority in controlling production costs for coal-fired units.
Research on cost control in coal-fired units is primarily divided into internal cost control and external cost control [2]. Internal costs include fuel costs, equipment and maintenance costs, labor costs, and financial investment costs, among others, among which fuel costs are the most significant component, typically accounting for over 70% of power generation costs. External costs encompass environmental costs and resource consumption costs. With increasingly stringent environmental regulations on thermal power in countries and regions such as China, the optimization of environmental costs for coal-fired units has become a key issue of focus [3,4,5,6].
Coal blending decisions not only determine coal costs but also define coal quality indicators such as sulfur and ash content. Using numerical simulations and combustion experiments, the basic chemical characteristics of blended coal have been relatively well studied [7,8]. Sulfur, ash, and other coal quality factors in blended coal are key contributors to pollutants generated during combustion, which constitute a major source of environmental costs [9,10,11]. Therefore, ensuring the adoption of rational coal blending strategies (CBS) during operation is of great practical significance for the overall cost control of coal-fired units [12].
A multi-constraint optimization model was proven to be an effective approach for optimizing CBS. Yan et al. proposed a hybrid dynamic coal blending method, employing a multi-stage dynamic decision model to balance economic benefits with environmental protection, optimizing coal procurement, blending, and distribution plans to achieve the dual goals of reducing carbon and PM10 emissions [13]. Lv et al. introduced a two-layer coal blending optimization method based on an equilibrium strategy to achieve coordinated reductions in carbon and PM10 emissions under uncertain conditions [14]. Amini et al. developed a coal blending optimization strategy based on a robust optimization model, aiming to maximize economic profit and reduce blending risk while addressing uncertainties in coal quality, sampling, and measurement [15]. This model leverages fuzzy expectations and possibility measures to optimize the blending scheme, balancing the conflict between environmental protection and economic benefits. Yuan et al. proposed a coal blending optimization method based on petrographic characteristics, using Xgboost and support vector regression to construct a coke quality prediction model and employing a multi-constraint optimization model to minimize coal costs while meeting quality requirements [16]. Nawaz et al. examined coal blending strategies from a supply chain perspective, aiming to reduce emissions while maintaining generation efficiency. They simulated co-combustion patterns of various coal types and biomass to assess the technical and economic feasibility of coal blending strategies across different scenarios [17].
However, in the study of CBS optimization models, the approach to ensuring that coal blending strategies meet power generation requirements is often overly simplified. The use of thermoelectric conversion parameters [13,14] or simple calorific value estimates of coal [15,16,17] clearly falls short of accurately reflecting the complex and dynamic thermoelectric conversion processes of power units. During the operation of coal-fired units, ensuring that power generation meets dispatch requirements remains the most critical technical metric. These factors necessitate adjusting coal blending strategies in practical operations based on human experience rather than optimization models, resulting in inefficiencies in the decision-making process and economic losses.
On the other hand, coal-fired power units, as a mature type of generation technology, benefit from advanced digital control systems such as distributed control systems and safety instrumented systems, enabling the effective collection and storage of extensive operational data. Moreover, with the rapid advancement of parameter control research [18], output condition recognition with high accuracy was already achieved using unit operation data as an input [19,20]. Consequently, it becomes feasible to leverage data-driven models to more accurately reflect the power generation corresponding to CBS.
To address the aforementioned issues, we developed a data-driven intelligent adaptive generation framework for cost-effective CBS, comprising two main components: a unit output recognition module (UOCR) and a cost-effective strategy generation module (CESG). The primary contributions are as follows:
  • The framework can intelligently and adaptively recognize the output conditions of units with reasonable accuracy based on CBS and supplementary feature parameters.
  • Building on the first component, we designed a surrogate-assisted optimization model to generate cost-effective CBS.
  • The framework provides extensibility in terms of algorithm selection, such as recognition and optimization algorithms.
The following sections first introduce the current experience-based coal blending decision-making process and its limitations. Then, the composition and technical details of the proposed framework are described in detail, including a novel dedicated neural network called ’Imitator.’ Subsequently, a case study and analysis of the framework were conducted using real operational data from a coal-fired unit. Next, the main contributions of this work are discussed, and directions for future research are suggested. Finally, the paper concludes with a summary and key findings.

2. Coal Blending Decision-Making for Thermal Power Units

As shown in Figure 1, the CBS has substantial influence on various types of generation costs. Among these, coal cost is directly determined by the CBS and typically accounts for over 70% of the total operating costs in thermal power units, making it a central concern in generation cost management.
Other cost types related to CBS can be collectively referred to as auxiliary costs, which are significantly impacted by coal quality indicators such as sulfur and ash content in the blended coal. These costs are highly beneficial for a comprehensive evaluation of the economic efficiency of CBS and specifically include desulfurization and denitrification costs, emission tax costs, and equipment electricity costs. Desulfurization and denitrification costs refer to the expenses incurred from treating raw flue gas to meet environmental regulatory requirements. Emission tax costs are taxes paid by thermal power plants on pollutants ultimately emitted by the units. Equipment electricity costs refer to the expenses associated with electricity purchased from the external grid for operating equipment such as coal crushers, feeders, and coal conveyor belts.
Currently, the formulation of CBS in most coal-fired power plants is primarily based on human experience. Figure 2 illustrates the typical decision-making process for developing CBS. Upon receiving an output dispatch requirement, fuel or operations managers refer to the CBS menu to establish the blending ratio for each coal type and set the total coal feed rate, forming an initial blending strategy. This strategy is then continuously adjusted based on feedback from the unit’s actual output until the output reaches the required range, resulting in the final CBS. The CBS menu is structured with fixed unit conditions, segmented by output ranges based on theoretical upper and lower output limits. It provides reference values for blending strategies, which, based on past experience, are feasible in terms of coal quality (e.g., calorific value) and coal cost efficiency.
This experience-based decision-making process for generating CBS focuses on meeting dispatch conditions while considering some economic factors. However, as grid dispatch becomes increasingly complex and pressures from electricity market bidding intensify, this approach has begun to reveal two key limitations:
  • The output condition classification that serves as a reference for initial CBS may not adapt well to actual circumstances. This decision-making process provides fixed condition classifications based on theoretical output limits and derives an initial CBS from past experience. However, during operation, the grid’s dispatch of unit output adjusts according to the actual power load, with outputs often concentrated within a specific range of the theoretical limits. This range may be covered by only a minimal number of reference condition classifications, resulting in an initial CBS with insufficient granularity that lacks the adaptability to adjust dynamically to real production needs, thus differing significantly from the final feasible CBS. Additionally, the initial CBS primarily focuses on coal cost, to some extent overlooking other cost types.
  • The method of adjusting the initial CBS to form the final CBS is inefficient and fails to ensure economic viability. The process of fine-tuning the initial CBS to a final CBS relies entirely on continuous adjustments based on actual output feedback and manual judgment. This process may lead to fuel wastage and has a high probability of compromising the economic efficiency of the initial blending strategy.

3. Framework for Generating Cost-Effective Coal Blending Strategy

To address the limitations of the experience-based CBS decision-making process, this paper proposes a cost-effective CBS generation framework. This framework allows fuel and operations managers to quickly generate a current cost-effective CBS by specifying output conditions along with a few past blending strategies and unit parameters. Figure 3 illustrates the framework’s components and the connection between the UOCR and the cost-effective strategy generation module.
The UOCR enables the adaptive classification of unit output conditions and can intelligently identify these conditions using CBS and other unit parameters as inputs. This module first performs adaptive classification based on the actual operational dataset. Then, by using the imitator, it transforms CBS and other inputs into dimensions compatible with image classification models, allowing fine-tuned pre-trained models to perform output condition recognition.
The cost-effective strategy generation module is designed with a surrogate-assisted CBS model to generate strategies that account for costs related to coal, environmental factors, and equipment power consumption under specified output conditions. This module first applies the UOCR as a surrogate model to impose output condition constraints in the strategy generation process. Then, it uses a regression surrogate model to evaluate additional costs within the objective function. Finally, a population-based optimization algorithm is employed to solve the strategy generation model, producing a cost-effective CBS.
Figure 4 further summarizes the workflows of the framework. The workflows are divided into two stages. Stage 1 is based on offline historical datasets. Firstly, the labeled output condition dataset was obtained through the adaptive unit output classification of UOCR. Secondly, the historical unit output condition dataset, CBS dataset, and supplementary features dataset were used to fine-tune the Imitator and pre-trained image classification models in UOCR. Finally, based on the historical CBS dataset and auxiliary costs dataset, Auto-ML was used to construct the surrogate model of auxiliary cost in CESG.
Stage 2 is the process of achieving cost-effective CBS generation using the trained-out framework. After inputting real-time CBS data, supplemental features data, and dispatch requirements, CESG generates CBS, updates CBS with the goal of reducing the total cost (including coal cost and auxiliary costs), and finally generates a cost-effective CBS. During this process, UOCR continuously provides CESG with information about the output conditions corresponding to the generated CBS.

3.1. Unit Output Condition Recognition Module

3.1.1. Adaptive Unit Output Condition Classification

Unit output condition classification involves dividing output into different ranges based on magnitude while simultaneously categorizing CBS and relevant unit parameters. This paper proposes an adaptive condition classification method that determines output probability distributions and completes interval division based on past operational data. The set of output data O in the operational dataset is expressed as:
O = { o i i = 1 , 2 , , L } ,
where o i represents any output data point and L is the dataset length. We employed Kernel Density Estimation (KDE) to model the probability density of output data. KDE is a non-parametric method that does not require any assumptions about the data’s distribution, making it particularly suitable for handling complex, irregular, or unknown distributions. This characteristic aligns well with the practical scenario where the output data distribution is often multimodal. Consequently, the process of modeling the probability density of output data can be expressed as
p ( o j ) = 1 L h i = 1 L K e r n e l o j o i h ,
where o j O , h is the smoothing parameter, and K e r n e l ( ) typically uses a Gaussian kernel function. Based on the size of the output dataset and ensuring sufficient training samples for each condition classification, the total number of unit output conditions is set to n . The set of conditions is then defined as
C o n d i t i o n s = { C o n d i t i o n   k | k = 1 , 2 , , n } .
When the output dataset is sufficiently large, the number of conditions will significantly increase compared to the fixed value used in experience-based decision-making models. The upper limit o k for each condition satisfies:
o k 1 o k p ( t ) d t = 1 n , k = 0 , 1 , , n ,
where o 0 represents the minimum output value in the actual dataset and o n represents the maximum output value in the actual dataset. Any unit output condition classification can be expressed as:
C o n d i t i o n   k = { o i | o k 1 o i < o k } , i f   1 k < n { o i | o k 1 o i o k } , i f   k = n , k = 1 , 2 , , n ,
where k is the label for unit condition classification.
Figure 5 illustrates the process of unit output condition classification based on probability distribution for past operational data. By setting the total number of conditions as n , adaptive classification results can be obtained.

3.1.2. Imitator

According to state-space theory, the operating condition of a unit can be described by a state vector composed of several unit parameters, where the features of the state vector are determined by the magnitude of the unit parameters. This bears a strong resemblance to the features formed by pixels of varying values in an image. Therefore, we designed the imitator to expand the unit condition recognition input matrix—comprising CBS and supplementary unit parameters, resembling a grayscale image—into a three-channel input matrix commonly used in image classification models, enabling initial feature extraction and transformation of the unit parameter matrix. Firstly, for inputs, to ensure the comprehensiveness of unit condition recognition, additional unit parameters—such as boiler feedwater temperature, pressure, and flow rate—can be selected as supplementary features alongside CBS. The CBS x C B S t c and supplementary features x S F t s at time t can be expressed as:
x C B S t = [ x 1 t , x 2 t , , x c t ] ,
x S F t = [ x 1 t , x 2 t , , x s t ] ,
where x 1 t represents the total coal flow rate, c denotes the total number of coal types involved in blending, x 2 t , , x c represents the blending ratio of the c 1 type coal, the omitted blending ratios can be calculated using 1 i = 2 c x i t , s indicates the total number of supplementary features, and x S F t composes of scalar values for each supplementary feature at time t . After unifying the subscripts, the combined unit parameter features at time t can be defined as:
x t = [ x C B S t , x S F t ] T = [ x 1 t , x 2 t , , x m t ] T ,
where m = c + s represents the total number of input features. Secondly, since adjustments to certain parameters in thermal power units have significant delays in affecting output, the Imitator’s input can incorporate unit parameter combinations by backtracking τ 1 time steps from time t . Thus, the Imitator input X t τ × m at time t can be expressed as:
X t = [ x t τ + 1 , , x t ] .
To simplify calculations and facilitate subsequent feature dimension transformations, τ = m is typically set such that X t forms a square matrix. Additionally, x i t represents the normalized values of the features:
x i t = x ^ i t min ( { x ^ i j | j = 1 , 2 , , L } ) max ( { x ^ i j | j = 1 , 2 , , L } ) min ( { x ^ i j | j = 1 , 2 , , L } ) , i = 1 , 2 , , m ,
where x ^ i t is the original value of the feature, and max ( { x ^ i j | j = 1 , 2 , , L } ) and min ( { x ^ i j | j = 1 , 2 , , L } ) are the maximum and minimum values of the corresponding feature in the dataset, respectively. Figure 6 illustrates the structure of the input X t . Horizontally, it consists of x t arranged in chronological order, while vertically, each x t is composed of CBS and supplementary unit parameters. After normalization, X t can be viewed as a grayscale image with an equal width and height of m .
In the design of Imitator, as shown in Figure 7, a simple combination of attention mechanisms and transposed convolution matrices is used for feature extraction and transformation of the input. The final output matches the common input dimensions required by pre-trained image classification models.
Firstly, broadcast x t to form X 1 , Q t m × m as the input for the query, using X t as the input for the key and value, and perform the forward propagation calculation for multi-head attention, with the number of heads set as r :
Q 1 , i = X 1 , Q t W 1 , i Q ,
K 1 , i = X t W 1 , i K ,
V 1 , i = X t W 1 , i V ,
h e a d i t = softmax Q 1 , i t ( K 1 , i t ) T d k V 1 , i ,
h ^ 1 t = [ h e a d 1 t , , h e a d r t ] W 1 O ,
where i = 1 , 2 , , r , W 1 , i Q m × d 1 q , W 1 , i V m × d 1 v , and W 1 , i K m × d 1 k are the weight matrices for the head i and Q 1 , i , K 1 , i , and V 1 , i are the query, key, and value for the head i , respectively; d 1 q = d 1 k ; and h e a d i t m × d 1 v is the attention score for the head i , calculated using the dot-product attention mechanism. W O ( d 1 v r ) × m is the output weight matrix of the multi-head attention mechanism, and h ^ 1 m × m is the output of the multi-head attention mechanism. The multi-head attention extracts the similarity information between x t and each unit parameter combination across supplementary time steps. The sum of X Q 1 t and h ^ 1 t yields the first hidden state, h 1 t m × m :
h 1 t = X Q 1 t + h ^ 1 t .
Secondly, three consecutive transposed convolution layers are used to progressively expand the feature dimensions and the number of channels of h 1 :
U 1 t i , j , l = a = 0 w 1 1 b = 0 w 1 1 h 1 t i a , j b K 1 a , b , l ,
U 2 t i , j , g = a = 0 w 2 1 b = 0 w 2 1 l = 0 2 U 2 t i a , j b , l K 2 a , b , l , g ,
h 2 t i , j , g = a = 0 w 3 1 b = 0 w 3 1 l = 0 2 U 2 t i a , j b , l K 3 a , b , l , g
where K 1 w 1 × w 1 × 3 , K 2 w 2 × w 2 × 3 × 3 , K 3 w 3 × w 3 × 3 × 3 are the weight matrices of the three transposed convolutions, represents the matrix indexing operation, U 1 t u 1 × u 1 × 3 , U 2 t u 2 × u 2 × 3 represent the outputs of the first and second convolutional kernels, and h 2 t u × u × 3 is the output of the third convolutional kernel, which serves as the second hidden state. The stride for all three convolutional kernels is set to 1, with zero padding. The width and height h 2 t for each channel can be calculated as follows:
u = m + w 1 + w 2 + w 3 3 .
where u is the required width and height for fine-tuning inputs in pre-trained image classification models, typically set to 224 or 384, and the specific values of w 1 , w 2 , w 3 can be adjusted based on m . Using multiple consecutive transposed convolution layers instead of a single transposed convolution layer is beneficial because, in most cases, m u and multiple layers help control the size of the convolutional kernel.
Finally, X 1 , Q t is scaled and broadcast along the channel dimension to form X 2 , Q t u × u × 3 , with scaling performed using bilinear interpolation:
X 2 , Q t i , j , g = i = 0 1 j = 0 1 ϖ p i ϖ q j X 1 , Q t p i , q j ,
where ϖ p i , ϖ q j represent the interpolation weights in the row and column directions, which depend on the distances between p i and p as well as q i and q . The closer the distance is, the higher the weight is, and p = i m u , q = j m u . Using X 2 , Q t as the query input, attention is computed for each channel:
Q 2 , g = X 2 , Q t : , : , g W 2 , g Q ,
K 2 , g = h 2 t : , : , g W 2 , g K ,
V 2 , g = h 2 t : , : , g W 2 , g V ,
c h a n n e l g t = softmax Q 2 , g t ( K 2 , g t ) T d k V 2 , g ,
h ^ 3 t = [ c h a n n e l 1 t W 2 , 1 O , , c h a n n e l g t W 2 , g O ] ,
where g = 1 , 2 , 3 , W 2 , i Q u × d 2 q , W 2 , i K u × d 2 k , and W 1 , i V u × d 2 v are the weight matrices for the channel g , and Q 2 , g , K 2 , g , and V 2 , g are the query, key, and value for the channel g , respectively. The c h a n n e l g t u × d 2 v is the attention score for the channel g , calculated using the dot-product attention mechanism. W 2 , g O ( d 2 v r ) × u is the output weight matrix of the attention mechanism, and h ^ 3 t u × u × 3 is the output of the attention mechanism. The attention mechanism further explores the similarity information between x t and h 2 t after dimension transformation. By summing X 2 , Q t and h ^ 3 t , the Imitator’s output h t u × u × 3 is obtained:
h t = X 2 , Q t + h ^ 3 t .
In both attention mechanisms used by the Imitator, the results obtained by broadcasting x t are used as input for the query, essentially extracting the similarity information between x t and each unit parameter combination across supplementary time steps. The three consecutive transposed convolutions expand the dimensions and channels of X t to facilitate transforming X t into data input formats commonly used by pre-trained image classification models, such as 224 × 224@3 or 384 × 384@3. Through two addition operations, the feature representation of x t is gradually enhanced in the dimension transformation process, simplifying the complexity of the unit output condition recognition task. In summary, the design of the Imitator’s working mechanism enables the unit output condition recognition task—with the unit parameter matrix as input and classification labels as output—to leverage mature and robust image classification models, such as ResNet [21], ResNeXt [22], and Vision Transformers [23], which are pre-trained on large-scale image classification datasets and widely validated in both academia and industry. This setup allows for selecting different model architectures based on the data characteristics and variations in different unit operation datasets, ensuring the foundational performance of the UOCR.

3.1.3. Intelligent Condition Recognition Based on Pre-Trained Image Classification Models

The Imitator output ht transforms the unit parameter input matrix into the commonly used input data dimensions for fine-tuning pre-trained image classification models. The forward process of the pre-trained image classification model can be expressed as:
Z t = P T I C ( h t ) ,
where Z t d Z denotes the feature vector before entering the classifier and P T I C ( ) represents the forward propagation of the pre-trained image classification model. Next, H t is fed into the classifier:
z ˜ t = W c l f Z t + β ,
where z ˜ t n is the output of the classifier, W c l f d Z × n is the classifier’s weight, and β is the classifier’s bias. The recognition result k ˜ of the UOCR can be obtained by:
k ˜ = Softmax ( z ˜ t ) .
The loss function of the UOCR is defined by the cross-entropy function:
L = i = 1 C z i log ( p i ) ,
where z i n is the one-hot encoding of the true class, with the k -th component as 1 and the others as 0. The k is the classification label for the unit condition. The p i is calculated as follows:
p i = e z ˜ i j = 1 n e z ˜ j .
Through fine-tuning the pre-trained image classification model, the Imitator, and the classifier, a trained UOCR is obtained. This module achieves intelligent recognition of output conditions by using a unit parameter matrix composed of multi-step CBS and supplementary features as input.

3.2. Cost-Effective Strategy Generation Module

3.2.1. Calculation of Power Generation Costs and Coal Quality Indicators

The evaluation of a CBS can be determined through its corresponding power generation cost. To provide a comprehensive measure of CBS generation costs, auxiliary costs such as desulfurization and denitrification costs, emission tax costs, and equipment electricity consumption are included in addition to coal costs. Furthermore, this paper adopts a cost-per-unit-generation approach for cost accounting, with the conversion between unit output and generation given by:
E = t Δ t t o ( t ) d t .
where E represents the power generation of Δ t , measured in k W · h , and o ( t ) is the function of output over time, measured in k W . Based on actual production practices in thermal power plants, when Δ t is small, o ( t ) remains relatively constant, and o t within Δ t can be considered as an average value. Thus,
E = o t Δ t .
In the cost calculation of CBS, we use the lower limit obtained from adaptive classification as the unified o t , and the power generated in Condition k within Δ t is:
E k = o k 1 Δ t ,
enabling a uniform power unit cost within the same condition. This simplified calculation method, which considers a single-point value over a short period as an average value of Δ t , is extended to the entire cost accounting process.
The coal cost C c o a l can be calculated from the CBS as follows:
C c o a l = x 1 t ( i = 2 c p i x i t + p 1 ( 1 i = 2 c x i t ) ) o k 1 Δ t ,
where p i ,   i = 2 , 3 , , c correspond to the prices of each type of coal in the blending strategy and p 1 is the price of the last type of coal that was omitted.
Auxiliary costs C a u x i l i a r y include desulfurization and denitrification costs, emission tax costs, and equipment electricity costs. Desulfurization and denitrification costs are mainly calculated based on the consumption of desulfurization materials such as limestone and urea within Δ t . Emission tax costs are calculated based on the net flue gas flow and the concentrations of nitrogen oxides, sulfur oxides, and dust within Δ t . Equipment electricity costs are calculated based on the electricity consumed by equipment involved in the CBS process within Δ t . The full mathematical formulas for calculating auxiliary costs C a u x i l i a r y are detailed in Figure 3. Here, we use nitrogen oxide emission tax costs as an example:
C N O x = ς N O x V c l e a n p N O x τ N O x o k 1 Δ t ,
where ς N O x represents the concentration of nitrogen oxides in the net flue gas, measured in kg / Nm 3 , p N O x represents the unit price of nitrogen oxides in the emission tax, measured in CNY / kg , τ N O x is the nitrogen oxide equivalent factor, and V c l e a n represents the total volume of net flue gas in standard atmospheric cubic meters, measured in Nm 3 . This calculation depends on the original flue gas flow rate, pressure, temperature, humidity, and environmental atmospheric pressure. It is evident that the CBS cannot directly calculate C a u x i l i a r y through explicit computation. Therefore, we express the power generation cost of CBS as:
C = x 1 t ( i = 2 c p i x i t + p 1 ( 1 i = 2 c x i t ) ) o k 1 Δ t + C a u x i l i a r y ( x C B F t ) .
The CBS requires calculating coal quality indicators such as ash content, sulfur content, calorific value, and volatile matter. These coal qualities of CBS can be calculated using a weighted average as follows:
φ q u a l i t y = i = 2 c x i t φ i , q u a l i t y + ( 1 i = 2 c x i t ) φ 1 , q u a l i t y ,
where φ q u a l i t y represents the coal quality to be calculated, φ i , q u a l i t y represents the coal quality of type i coal, and φ 1 , q u a l i t y represents the coal quality of the omitted type.

3.2.2. Surrogate-Assisted Generation Model

Based on CBS generation cost accounting, coal quality calculations, and the UOCR, this paper demonstrates a surrogate-assisted cost-effective CBS generation model:
min ( C ) = min x 1 t ( i = 2 c p i x i t + p 1 ( 1 i = 2 c x i t ) ) o k 1 Δ t + C a u x i l i a r y ( x C B S t )   s . t .   U O C R ( x C B S t ) = k d i s p a t c h φ s φ s max φ a s h φ a s h max φ v φ v max φ c a l φ c a l min i = 2 c x i t 1 .
Firstly, the constraints in generating the CBS need to account for coal quality and output conditions. Here, U O C R ( ) refers to the surrogate model using the trained UOCR, with real-time values used for the previous m 1 steps of CBS and all supplementary features, combined with the generated CBS as input, to output the recognized unit condition label. The k d i s p a t c h denotes the required unit condition label, and φ s max , φ a s h max , and φ v max are the upper limits for sulfur content, ash content, and volatile matter, respectively. The φ c a l min represents the minimum calorific value required to ensure normal unit operation.
Secondly, the objective function for strategy evaluation uses a surrogate model, specifically C a u x i l i a r y ( ) , which is a regression surrogate model for CBS and auxiliary costs. Auto-ML technology can be employed to identify the most suitable surrogate model from classic regression models such as LightGBM-XT, LightGBM, KNN, RandomForest, and other similar models, along with their weighted ensemble models. By adopting a data-driven surrogated model for auxiliary costs, the strategy generation model ensures scalability. Beyond the auxiliary costs covered in this study, costs associated with carbon footprints and carbon tax could also be incorporated into the objective function under certain scenarios, thereby enabling different types of CBS generation control. The weighted ensemble model of multiple regression models fitting C a u x i l i a r y ( ) can be expressed as:
C a u x i l i a r y ( x C B S t ) = i = 1 M α i C a u x i l i a r y i ( x C B S t ) ,
where M is the total number of models in the weighted ensemble, C a u x i l i a r y i ( ) represents the surrogate model for calculating the i type of auxiliary cost, and α i is the weight of the i surrogate model in the final auxiliary cost calculation.

3.2.3. Cost-Effective CBS Generation

In designing the generation model, we employed data-driven surrogate models for both the evaluation function and constraints, resulting in a mathematically implicit expression. Therefore, the generation of CBS can be achieved using population-based metaheuristic algorithms such as genetic algorithms, differential evolution algorithms, and particle swarm optimization. These algorithms generate a certain number of individuals as candidate CBS, using a fitness function composed of the inverse of the objective function and penalty terms as the evaluation standard for the population. The population is iteratively refined based on fitness scores. After all iterations, the individual with the highest fitness in the population is the generated cost-effective CBS. The mathematical expression of the fitness function is as follows:
F i t t n e s s ( x ˜ C B S , i t ) = 1 C , i f   c o n s t r a i n t ( x ˜ C B S , i t ) P e n a l t y ,   i f   n o t   c o n s t r a i n t ( x ˜ C B S , i t ) ,
where x ˜ C B S , i t represents an individual in the population, i.e., a generated candidate CBS; P e n a l t y is a penalty term, typically much smaller than the inverse of the objective function; and c o n s t r a i n t ( ) represents the constraint check calculation. The iteration strategy of swarm intelligence algorithms generally focuses on individuals with higher fitness in the population, gradually eliminating those that do not meet the constraints. This ensures the effectiveness of the cost-reducing CBS generated under constraint conditions. The pseudocode for the CBS generation process of CESG is shown in Algorithm 1. In Algorithm 1, p o p represents the set population size, denotes the total number of constraints in the generation model, and I t e r a t i o n ( ) refers to the population update process of the selected population-based metaheuristic algorithm, such as selection, crossover, and mutation in genetic algorithms.
Algorithm 1. The Pseudocode for the CBS Generation Process of CESG
Input  p o p , c , X t
Output  b e s t   f i t n e s s , b e s t   x C B S t
Initialize  X ˜ C B F t = r a n d ( p o p , c ) , b e s t   f i t n e s s = , b e s t   x C B S t = 0 c
Electronics 14 00561 i001
return  b e s t   f i t n e s s , b e s t   x C B S t

4. Case Study and Analysis

4.1. Overview of the Sample Coal-Fired Unit and Experimental Environment

This work used a 350 MW coal-fired power unit produced by BABCOCK & WILCOX BEIJING COMPANY LTD, located in a coastal province in southeastern China, as the subject of the case study. During the significant coal price fluctuations caused by international conditions in 2021–2022, the performance of this unit was notably impacted, with a marked increase in power generation costs. As shown in Figure 8, this unit initially classified operating conditions into four types based on theoretical output limits and used reference values from manually derived initial CBS strategies, which were then adjusted according to feedback from actual output to obtain the final CBS. With increasing grid dispatch volatility and pressure from electricity market bidding, this unit urgently required an accurate and cost-effective CBS decision-making process to control power generation costs.
This unit primarily stored raw coal in six coal storage silos: silos 1 and 2 stored Coal Type A, silos 3 and 4 stored Coal Type B, and silos 5 and 6 stored Coal Type C. The three types of coal varied sequentially in price and calorific value and exhibited significant differences in coal quality indicators, such as sulfur content, ash content, and volatile matter. The CBS for this unit can be expressed as:
x C B S t = [ x 1 t , x 2 t , x 3 t ] ,
where x 1 t represents the total coal flow, x 2 t denotes the blending ratio of Coal Type A, and x 3 t denotes the blending ratio of Coal Type B. The blending ratio of Coal Type C can be calculated from the equation.
The experimental environment was equipped with an Intel(R) Xeon(R) Platinum 8369B CPU (2.90 GHz), an Nvidia A10 GPU (24 GB VRAM), and 32 GB RAM, running on the Ubuntu 18.04 operating system. The framework in the case study was implemented primarily using Python 3.9.18, PyTorch 1.12.0, Transformers 4.44.1, and Autogluon 1.0.0.

4.2. Sample Dataset

Let Δ t be set to 1 minute; we selected a continuous dataset of length 5008 from the normal operating data of the sample unit. Table 1 describes the data used as input for intelligent condition recognition, including the original CBS and the six supplementary unit parameters we selected, with a dataset length of 5008.
The total number of adaptive condition classifications n in the unit output recognition module was 6, with a backtracking step length m of 9. The dataset could form 5000 condition recognition samples. Figure 9 illustrates the unit output within the dataset, along with the division of the training and test sets for intelligent condition recognition in a 4:1 ratio. The training and test set division for the auxiliary cost surrogate model was the same as that for intelligent condition recognition.
The coal quality and price data for the three types of coal required by the CBS generation module within the dataset are shown in Table 2.

4.3. Hyperparameter Settings for Proposed Framework

The hyperparameter settings for the UOCR are shown in Table 3. During training, to ensure stable updates, a hyperparameter reduction strategy was applied, reducing the learning rate to 70% of the initial rate at the 15th epoch.
The input X t 9 × 9 of the unit condition recognition module, after normalizing the parameters, yielded the grayscale image shown in Figure 10.
The number of heads in the multi-head attention mechanism within the Imitator was r = 3 , and the transposed convolution kernel weight matrices were K 1 56 × 56 × 3 , K 2 65 × 65 × 3 × 3 , K 3 97 × 97 × 3 × 3 , with a training dropout set to 30%. The output of the Imitator was h t 224 × 224 × 3 . To evaluate the effectiveness of the UOCR, we conducted comparative experiments by pairing the Imitator with various pre-trained image classification models of differing architectures and parameter sizes. As shown in Table 4, these models included ResNet18, ResNet34, ResNet50, ResNet101, ResNeXt50-32X4D, ResNeXt101-32X4D, and Vision Transformer Base (ViT-Base). All image classification models, except ViT-Base, which was pre-trained on Image-21K, were pre-trained on Image-1K, with fine-tuning inputs formatted as 224 × 224@3 images, using the latest pre-trained weights from PyTorch or Transformers. Due to GPU memory limitations in the experimental environment, batch sizes varied depending on the pre-trained model applied within the unit output recognition module. For models with larger parameter sets, batch size was maximized within the available GPU memory.
In the CESG example, constraint thresholds for sulfur content, ash content, volatile matter, and calorific value were set based on the coal quality management requirements and actual conditions of the reference unit, as shown in Table 5.
The models used for the auxiliary cost surrogate experiments included LightGBM [26], LightGBMLarge [26], LightGBMXT [26], CatBoost [27], ExtraTrees [28], KneighborsDist [29], KneighborsUnif [29], NeuralNetTorch [30], RandomForest [31], XGBoost [32], NeuralNetFastAI [33], and their weighted ensemble model, WeightedEnsemble [34]. We let the unit output condition label obtained through adaptive classification in the dataset be equal to the dispatch label k = k d i s p a t c h . The CBS generation model in the case framework can be expressed as:
min ( C ) = min x 1 t ( i = 2 3 p i x i t + p 1 ( 1 i = 2 3 x i t ) ) o k 1 Δ t + C a u x i l i a r y ( x C B S t )   s . t .   U O C R ( x C B S t ) = k x 2 t + x 3 t 1 φ s 1.5 % φ a s h 30 % φ v 20 % φ c a l 4200 .
We employed a roulette-based genetic algorithm as the population-based optimization algorithm [35,36,37], with candidate solutions initialized randomly and iteration strategies including crossover and mutation. The individuals in the genetic algorithm were composed of x C B S t . The penalty term was 1 × 10−2, and the fitness function can be expressed as:
F i t t n e s s ( x ˜ C B s , i t ) = 1 C , i f   c o n s t r a i n t ( x ˜ C B S , i t ) 1 × 10 2 ,   i f   n o t   c o n s t r a i n t ( x ˜ C B S , i t )
The additional hyperparameter settings for the genetic algorithm are shown in Table 6.

4.4. Experimental Results and Analysis of UOCR

The experimental results for the adaptive output condition classification are shown in Figure 11. In the actual dataset, the output was concentrated between 180.0 and 307.1 MW. Figure 11a illustrates the distribution of output points across each classification, where adaptive output condition classification ensured that the total number of output points in each category was approximately the same. This resulted in similar label counts for each type in subsequent intelligent condition recognition training. Figure 11b compares unit output condition classification based on human experience with adaptive condition classification. The adaptive method significantly enhanced the granularity of conditions, supporting a higher level of refinement in CBS.
Figure 12 and Table 7 present the training error, training accuracy, and test accuracy of UOCRs using different pre-trained models. Without adjusting hyperparameters for different pre-trained models and with only 25 epochs of training, all models except ViT-Base achieved over 95% training accuracy and over 80% test accuracy. Among them, the UOCRs using ResNet50 and ResNet101 achieved the highest test accuracy.
As shown in Figure 13, we selected data points from six different unit conditions and input them into the unit condition recognition module using ResNet50 to observe the effect of the Imitator. The first column shows the input in the form of a 9 × 9 grayscale image, where higher pixel values appear closer to white. Columns 2–4 display the 224 × 224 three-channel images processed by the Imitator, where higher pixel values appear darker. In the first column, it is evident that the original inputs from the six different conditions already showed significant differences. In columns 2–4, the outputs processed by the Imitator clearly retained these original input features as the 9 × 9 grayscale image was expanded to a 224 × 224 three-channel image.
Specifically, along the height of the image, areas close to white in the original input resulted in very dark or very light horizontal stripes in the output, reflecting differences in the magnitude of different unit parameters in the input. Along the width of the image, the output showed vertical stripes with varying intensities, corresponding to temporal changes in the same unit parameter in the input. This effect was particularly pronounced in unit conditions 3 to 6, as illustrated in Figure 14. The structural design of the Imitator expanded the input dimensions to 224 × 224@3 and, through its learnable parameters, preserved distinct feature differences among different conditions in the output.
UOCR first divided the output conditions into more granular categories compared to human experience, based on the total set classifications and the actual production dataset. Next, the Imitator, pre-trained image classification model, and classifier were used to achieve intelligent condition recognition based on CBS. The Imitator’s output features, which exhibited distinct differences across various conditions, enabled the pre-trained image classification model—trained on three-channel image datasets and fine-tuned with prior knowledge from pre-trained weights—to achieve higher accuracy in the unit condition recognition task.

4.5. Experimental Results and Analysis of CESG

Using the ResNet50 UOCR, which achieved the highest test accuracy, we selected data points from six conditions to conduct an experimental analysis on CESG. Figure 15 illustrates the selected data points and their preceding eight steps of output conditions used as input for the optimization strategy generation module.
Table 8 presents the experimental results for the auxiliary cost surrogate model. In CESG, we used the weighted ensemble model with the lowest training error as the surrogate model for auxiliary costs.
Figure 16 shows the iteration process of the genetic algorithm in the experimental phase of the CESG. After the fourth generation, all condition experiments yielded over 70 constraint-compliant individuals. In the experiments for Conditions 1, 3, and 5, the total number of compliant individuals stabilized close to the total population size until the end of iterations, while in Conditions 2, 4, and 6, the number fluctuated but generally was maintained above 70% of the total population. Even though the generation model included implicit surrogate models in both the objective function and constraints, using a population-based optimization algorithm proved feasible and efficient in generating candidate solutions for CBS optimization. The abundance of candidate solutions ensured a continuous reduction in total costs, with the final CBS generated in all six conditions showing a significant decrease in total costs compared to the initially generated strategies.
In designing the generation model, we imposed constraints on the generated strategy for output condition, calorific value, ash content, sulfur content, and volatile matter. Figure 17 compares the generated cost-effective CBS, the original strategy, and the constraint thresholds. Across all conditions, the generated cost-effective CBS met the output condition and four coal quality constraints and showed minimal differences from the original strategy in most indicators, ensuring practical applicability.
Figure 18 shows the rate of change in total coal flow; proportions of A coal, B coal, and C coal; total cost; coal cost; and other costs between the generated cost-effective CBS and the original strategy across six unit conditions. Table 9 provides the specific values for these indicators. In the six unit conditions, the generated cost-effective CBS achieved an average total cost reduction of 3.37% and an average coal cost reduction of 3.62% compared to the original strategy. The generated strategies exhibited significant variation across different unit conditions. For instance, in Conditions 1 and 5, the proportion of C coal and total coal flow were notably increased, while in Conditions 2 and 3, the proportion of A coal rose and total coal flow was reduced. Conditions 4 and 6 were obtained by fine-tuning the original strategy. Additionally, the auxiliary costs included in the strategy evaluation often differed substantially from changes in coal costs, which enhanced the global cost-effectiveness of the generated strategy in unit operation.
Overall, CESG can produce differentiated cost-effective CBS tailored to various conditions. These strategies comply with practical production constraints on output conditions and coal quality, significantly reducing total costs compared to the original strategy, and demonstrate strong robustness and applicability.

5. Discussion

In this study, we improved coal blending decision-making for thermal power units through the design and implementation of a cost-effective CBS strategy generation framework. This framework includes a UOCR and a cost-effective strategy generation module. Once the UOCR is trained, the CESG only requires inputting the specified unit conditions and a few unit parameters to generate the cost-effective CBS strategy. Our research shows that the cost-effective CBS strategies generated by this framework comply with practical production constraints and have a significant economic advantage in power generation costs compared to conventional experience-based blending strategies.
The UOCR first achieves adaptive classification of unit output conditions by forming output condition categories that are practical for production based on a set number of classifications and actual operational data. It then uses real blending strategies and selected supplementary unit parameters as inputs; with our Imitator architecture, the pre-trained image classification model, when fine-tuned, can intelligently recognize unit output conditions. Case study results indicate that, with a set classification count of six, the UOCRs using ResNet, ResNeXt, and Vision Transformer architectures achieved an average training accuracy of 96.64%, with a maximum test accuracy of 85.20%.
The CESG employs a surrogate-assisted model to evaluate not only coal costs but also auxiliary costs, such as emission taxes and desulfurization/denitrification costs. For strategy generation constraints, in addition to coal quality limitations, the module generates strategies that produce the desired unit output conditions use UOCR as a constraint. Case study results demonstrate broad applicability across six conditions; the generated strategies meet practical production constraints and, compared to the original strategy, reduce the total costs by an average of 3.37% and the coal cost by 3.62%. In some output conditions, the reduction in total cost and coal cost reached 8.7% and 6.3%, respectively.
Given that coal costs and auxiliary costs, which serve as evaluation targets in this framework, account for at least 70% of actual power generation costs in coal-fired power units—and coal-fired power continues to supply over 60% of electricity in China—the cost-effective CBS decision-making process enabled by this framework holds significant value for reducing costs and enhancing efficiency in power generation.
The proposed surrogate-assisted intelligent adaptive cost-effective CBS strategy generation framework opens avenues for future research and development. The following three points are recommended as references:
  • The design concept of combining the Imitator with pre-trained image classification models in the UOCR module offers promising opportunities for applying larger-scale deep learning models to the thermal power sector. This approach provides a potential pathway for leveraging advanced model architectures with greater parameter capacity in industrial applications.
  • Conducting more extensive case studies is necessary to comprehensively evaluate the framework’s design. Different types of thermal power units, such as ultra-supercritical units or those combined with renewable energy generation, operate under distinct mechanisms, which may pose new challenges to the framework’s stability. Additionally, thermal power units in different regions face varying cost structures, requiring adjustments to auxiliary cost design based on case studies. For instance, European thermal power plants may need to incorporate carbon footprint considerations into cost calculations.
  • The framework’s performance in engineering environments requires further experimentation. In the current case study, the trained framework took 50–70 seconds from initialization to generating a cost-effective CBS (detailed timing information is provided in the Figure 3), leaving room for improvement. However, the experimental environment in this study may not represent the operational conditions available in most engineering applications. The framework needs to be tested in a wider range of operating environments to comprehensively validate its performance.

6. Conclusions

This study proposes a surrogate-assisted intelligent adaptive cost-effective CBS strategy generation framework that reduces reliance on human experience in CBS decision-making and significantly enhances the economic efficiency of coal blending strategies. The framework consists of a UOCR and a cost-effective strategy generation module. The UOCR achieves adaptive classification of unit output conditions based on real data, utilizing CBS and supplementary unit parameters as inputs to intelligently recognize unit output conditions using the Imitator and pre-trained image classification models. Across architectures such as ResNet, ResNeXt, and Vision Transformer, the UOCRs achieved an average training accuracy of 96.64%. The CESG is designed with a surrogate-assisted model to comprehensively evaluate strategy cost-effectiveness and to impose constraints on the generated output conditions and coal quality requirements based on UOCR, producing cost-effective CBS using population-based metaheuristic algorithms that, on average, reduce the total costs by an average of 3.37% and the coal cost by 3.62% compared to the original strategy. In some output conditions, the reduction in total cost and coal cost reached 8.7% and 6.3%, respectively.

Author Contributions

Conceptualization, X.W. and S.W.; data curation, T.W.; formal analysis, X.W.; funding acquisition, T.W.; investigation, X.W.; methodology, X.W.; project administration, X.W.; resources, T.W.; software, X.W.; supervision, T.W.; validation, X.W., S.W., and J.D.; visualization, X.W.; writing—original draft, X.W.; writing—review and editing, S.W. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

Author Wang Teng was employed by the State Grid Corporation. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. National Bureau of Statistics of China. Statistical Communiqué of the People’s Republic of China on the 2023 National Economic and Social Development. Available online: https://www.stats.gov.cn/english/PressRelease/202402/t20240228_1947918.html (accessed on 2 November 2024).
  2. Srinivasan, P.; Shekhar, A. Internalizing the External Cost of Gaseous and Particulate Matter Emissions from the Coal-Based Thermal Power Plants in India. Part. Sci. Technol. 2021, 39, 632–640. [Google Scholar] [CrossRef]
  3. Xie, S.; Qin, P.; Zhang, M.; Xu, J.; Ouyang, T. A High-Efficiency and Eco-Friendly Design for Coal-Fired Power Plants: Combined Waste Heat Recovery and Electron Beam Irradiation. Energy 2022, 258, 124884. [Google Scholar] [CrossRef]
  4. Yuan, X.; Chen, L.; Sheng, X.; Liu, M.; Xu, Y.; Tang, Y.; Wang, Q.; Ma, Q.; Zuo, J. Life Cycle Cost of Electricity Production: A Comparative Study of Coal-Fired, Biomass, and Wind Power in China. Energies 2021, 14, 3463. [Google Scholar] [CrossRef]
  5. Lin, L.; Xu, B.; Xia, S. Multi-Angle Economic Analysis of Coal-Fired Units with Plasma Ignition and Oil Injection during Deep Peak Shaving in China. Appl. Sci. 2019, 9, 5399. [Google Scholar] [CrossRef]
  6. Ali, H.; Phoumin, H.; Weller, S.R.; Suryadi, B. Cost–Benefit Analysis of HELE and Subcritical Coal-Fired Electricity Generation Technologies in Southeast Asia. Sustainability 2021, 13, 1591. [Google Scholar] [CrossRef]
  7. Zhao, B.; Chen, G.; Qin, L.; Han, Y.; Zhang, Q.; Chen, W.; Han, J. Effect of Coal Blending on Arsenic and Fine Particles Emission during Coal Combustion. J. Clean. Prod. 2021, 311, 127645. [Google Scholar] [CrossRef]
  8. Baek, S.H.; Park, H.Y.; Ko, S.H. The Effect of the Coal Blending Method in a Coal Fired Boiler on Carbon in Ash and NOx Emission. Fuel 2014, 128, 62–70. [Google Scholar] [CrossRef]
  9. Khabarova, M.A.; Novikova, O.V.; Khabarov, A.A. State and Perspectives of Power and Industry Applications of Coal. In Proceedings of the 2019 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus), St. Petersburg, Russia, 28–31 January 2019; pp. 985–987. [Google Scholar]
  10. Wang, Y.; Liu, Z.; Huang, H.; Xiong, X. Reducing Measurement Costs of Thermal Power: An Advanced MISM (Mamba with Improved SSM Embedding in MLP) Regression Model for Accurate CO2 Emission Accounting. Sensors 2024, 24, 6256. [Google Scholar] [CrossRef] [PubMed]
  11. Wu, Y.; Ke, Y.; Xu, C.; Xiao, X.; Hu, Y. Eco-Efficiency Measurement of Coal-Fired Power Plants in China Using Super Efficiency Data Envelopment Analysis. Sustain. Cities Soc. 2018, 36, 157–168. [Google Scholar] [CrossRef]
  12. Zaid, M.Z.S.M.; Wahid, M.A.; Mailah, M.; Mazlan, M.A.; Saat, A. Coal Fired Power Plant: A Review on Coal Blending and Emission Issues. AIP Conf. Proc. 2019, 2062, 020022. [Google Scholar] [CrossRef]
  13. Yan, S.; Lv, C.; Yao, L.; Hu, Z.; Wang, F. Hybrid Dynamic Coal Blending Method to Address Multiple Environmental Objectives under a Carbon Emissions Allocation Mechanism. Energy 2022, 254, 124297. [Google Scholar] [CrossRef]
  14. Lv, C.; Xu, J.; Xie, H.; Zeng, Z.; Wu, Y. Equilibrium Strategy Based Coal Blending Method for Combined Carbon and PM10 Emissions Reductions. Appl. Energy 2016, 183, 1035–1052. [Google Scholar] [CrossRef]
  15. Amini, S.H.; Vass, C.; Shahabi, M.; Noble, A. Optimization of Coal Blending Operations under Uncertainty—Robust Optimization Approach. Int. J. Coal Prep. Util. 2022, 42, 30–50. [Google Scholar] [CrossRef]
  16. Yuan, Y.; Qu, Q.; Chen, L.; Wu, M. Modeling and Optimization of Coal Blending and Coking Costs Using Coal Petrography. Inf. Sci. 2020, 522, 49–68. [Google Scholar] [CrossRef]
  17. Nawaz, Z.; Ali, U. Techno-Economic Evaluation of Different Operating Scenarios for Indigenous and Imported Coal Blends and Biomass Co-Firing on Supercritical Coal Fired Power Plant Performance. Energy 2020, 212, 118721. [Google Scholar] [CrossRef]
  18. Huang, S.; Xiong, L.; Zhou, Y.; Gao, F.; Jia, Q.; Li, X.; Li, X.; Wang, Z.; Khan, M.W. Robust Distributed Fixed-Time Fault-Tolerant Control for Shipboard Microgrids With Actuator Fault. IEEE Trans. Transp. Electrif. 2024. [Google Scholar] [CrossRef]
  19. Huang, C.; Li, Z. Data-Driven Modeling of Ultra-Supercritical Unit Coordinated Control System by Improved Transformer Network. Energy 2023, 266, 126473. [Google Scholar] [CrossRef]
  20. Huang, C.; Sheng, X. Data-Driven Model Identification of Boiler-Turbine Coupled Process in 1000 MW Ultra-Supercritical Unit by Improved Bird Swarm Algorithm. Energy 2020, 205, 118009. [Google Scholar] [CrossRef]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  23. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  24. Models and Pre-Trained Weights—Torchvision 0.20 Documentation. Available online: https://pytorch.org/vision/stable/models.html (accessed on 27 October 2024).
  25. Google/Vit-Base-Patch16-384·Hugging Face. Available online: https://huggingface.co/google/vit-base-patch16-384 (accessed on 27 October 2024).
  26. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  27. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased Boosting with Categorical Features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 6639–6649. [Google Scholar]
  28. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely Randomized Trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef]
  29. Cover, T.; Hart, P. Nearest Neighbor Pattern Classification. IEEE Trans. Inf. Theor. 2006, 13, 21–27. [Google Scholar] [CrossRef]
  30. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in PyTorch. 28 October 2017.
  31. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  32. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar]
  33. Howard, J.; Gugger, S. Fastai: A Layered API for Deep Learning. Information 2020, 11, 108. [Google Scholar] [CrossRef]
  34. Zhou, Z.-H. Ensemble Methods: Foundations and Algorithms, 1st ed.; Chapman & Hall/CRC: Boca Raton, FL, USA, 2012; ISBN 978-1-4398-3003-1. [Google Scholar]
  35. Improved Roulette Wheel Selection-Based Genetic Algorithm for TSP|IEEE Conference Publication|IEEE Xplore. Available online: https://ieeexplore.ieee.org/document/7945968 (accessed on 27 October 2024).
  36. Wu, S.; Wang, H.; Yu, W.; Yang, K.; Cao, D.; Wang, F. A New SOTIF Scenario Hierarchy and Its Critical Test Case Generation Based on Potential Risk Assessment. In Proceedings of the 2021 IEEE 1st International Conference on Digital Twins and Parallel Intelligence (DTPI), Beijing, China, 15 July–15 August 2021; pp. 399–409. [Google Scholar] [CrossRef]
  37. Li, Y.; Wu, S.; Wang, H. Adaptive Mining of Failure Scenarios for Autonomous Driving Systems Based on Multi-Population Genetic Algorithm. In Proceedings of the 2024 IEEE Intelligent Vehicles Symposium (IV), Jeju Island, Republic of Korea, 2–5 June 2024; pp. 2458–2464. [Google Scholar]
Figure 1. Coal blending process and its impacted costs.
Figure 1. Coal blending process and its impacted costs.
Electronics 14 00561 g001
Figure 2. Decision-making for CBS based on human experience.
Figure 2. Decision-making for CBS based on human experience.
Electronics 14 00561 g002
Figure 3. Framework for generating cost-effective coal blending strategy.
Figure 3. Framework for generating cost-effective coal blending strategy.
Electronics 14 00561 g003
Figure 4. Framework workflows.
Figure 4. Framework workflows.
Electronics 14 00561 g004
Figure 5. Adaptive unit output condition classification based on probability distribution: (a) probability distribution of output in the actual dataset; (b) output sample points in the actual dataset; (c) unit condition classification results of the actual dataset and the theoretical condition range of the unit.
Figure 5. Adaptive unit output condition classification based on probability distribution: (a) probability distribution of output in the actual dataset; (b) output sample points in the actual dataset; (c) unit condition classification results of the actual dataset and the theoretical condition range of the unit.
Electronics 14 00561 g005
Figure 6. The Imitator’s input matrix of unit parameters composed of multi-step CBS and supplementary features.
Figure 6. The Imitator’s input matrix of unit parameters composed of multi-step CBS and supplementary features.
Electronics 14 00561 g006
Figure 7. The design of Imitator.
Figure 7. The design of Imitator.
Electronics 14 00561 g007
Figure 8. The four unit output conditions and corresponding CBS reference values proposed based on human experience for this unit.
Figure 8. The four unit output conditions and corresponding CBS reference values proposed based on human experience for this unit.
Electronics 14 00561 g008
Figure 9. Unit output in the dataset and division of training and test sets.
Figure 9. Unit output in the dataset and division of training and test sets.
Electronics 14 00561 g009
Figure 10. Sample input for the unit condition recognition module in the case study.
Figure 10. Sample input for the unit condition recognition module in the case study.
Electronics 14 00561 g010
Figure 11. Experimental results of adaptive output condition classification and comparison with original condition classification: (a) distribution of output points across each classification; (b) comparison between adaptive classification and original classification.
Figure 11. Experimental results of adaptive output condition classification and comparison with original condition classification: (a) distribution of output points across each classification; (b) comparison between adaptive classification and original classification.
Electronics 14 00561 g011
Figure 12. Comparison of training error, training accuracy, and test accuracy of UOCRs using different pre-trained models: (a) training error, training accuracy, and test accuracy over generations for UOCRs with different pre-trained models; (b) final training and test accuracy comparison for UOCRs with different pre-trained models.
Figure 12. Comparison of training error, training accuracy, and test accuracy of UOCRs using different pre-trained models: (a) training error, training accuracy, and test accuracy over generations for UOCRs with different pre-trained models; (b) final training and test accuracy comparison for UOCRs with different pre-trained models.
Electronics 14 00561 g012
Figure 13. Imitator’s input and output under six unit conditions.
Figure 13. Imitator’s input and output under six unit conditions.
Electronics 14 00561 g013
Figure 14. The first channel of the Imitator output in unit output condition 3.
Figure 14. The first channel of the Imitator output in unit output condition 3.
Electronics 14 00561 g014
Figure 15. Data points from six unit conditions used in the experiment and their previous eight steps of output conditions for input to CESG.
Figure 15. Data points from six unit conditions used in the experiment and their previous eight steps of output conditions for input to CESG.
Electronics 14 00561 g015
Figure 16. Iteration process of the genetic algorithm in the CESG experiment across six conditions.
Figure 16. Iteration process of the genetic algorithm in the CESG experiment across six conditions.
Electronics 14 00561 g016
Figure 17. Comparison of constraint items between the generated cost-effective CBS and the original strategy across six output conditions: (a) output condition comparison, (b) calorific value comparison, (c) sulfur content comparison, (d) ash content comparison, (e) volatile matter comparison.
Figure 17. Comparison of constraint items between the generated cost-effective CBS and the original strategy across six output conditions: (a) output condition comparison, (b) calorific value comparison, (c) sulfur content comparison, (d) ash content comparison, (e) volatile matter comparison.
Electronics 14 00561 g017
Figure 18. Rate of change between the generated cost-effective CBS and the original strategy across six unit conditions in terms of total coal flow, proportion of A coal, proportion of B coal, proportion of C coal, total cost, coal cost, and other costs.
Figure 18. Rate of change between the generated cost-effective CBS and the original strategy across six unit conditions in terms of total coal flow, proportion of A coal, proportion of B coal, proportion of C coal, total cost, coal cost, and other costs.
Electronics 14 00561 g018
Table 1. Description of input parameter data in the dataset.
Table 1. Description of input parameter data in the dataset.
MeanMinMax
Total Coal Flow Rate (t/h)145.3872.03240.85
A Coal Percentage (%)0.410.210.75
B Coal Percentage (%)0.270.070.61
Main Steam Pressure (MPa)2.701.923.47
Main Steam Flow Rate (t/h)785.83534.331038.70
Main Steam Temperature (°C)567.21547.52581.52
Feedwater Pressure (MPa)22.0916.0026.38
Feedwater Temperature (°C)269.27251.43283.59
Feedwater Flow Rate (t/h)772.52540.431042.08
Table 2. Coal quality and price data in the dataset.
Table 2. Coal quality and price data in the dataset.
Quality IndicatorsCoal Type
ABC
Sulfur Content (%)1.17 1.22 1.26
Ash Content (%)29.26 28.63 30.79
Volatile Matter (%)17.42 15.90 16.33
Calorific Value (Kcal/kg)5200.00 4576.81 3205.24
Price (CNY/t)843.31 725.77 650.28
Table 3. Hyperparameter settings for the unit output recognition module.
Table 3. Hyperparameter settings for the unit output recognition module.
Power Output Situation Label NumberInput Steps lengthTrain Hyperparameter
EpochInitial Learning Rate of ImitatorInitial Learning Rate of Pre-Trained ModelInitial Learning Rate of Classifier
6925 1 × 10−41 × 10−51 × 10−4
Table 4. Parameter counts, sources of pre-trained weights, and batch size settings for different pre-trained image classification models.
Table 4. Parameter counts, sources of pre-trained weights, and batch size settings for different pre-trained image classification models.
Number of ParametersSource of Pre-Trained WeightsBatch Size
ResNet1811.7M[24]256
ResNet3421.8M256
ResNet5025.6M200
ResNet10144.5M100
ResNeXt50-32X4D25.0M128
ResNeXt101-32X4D88.8M64
Vit-Base86.4M[25]64
Table 5. Coal quality constraint indicator settings.
Table 5. Coal quality constraint indicator settings.
MinimumMaximum
Sulfur Content (%)-1.5
Ash Content (%)-30
Volatile Matter (%)-20
Calorific Value (Kcal/kg)4200
Table 6. Genetic algorithm hyperparameter settings.
Table 6. Genetic algorithm hyperparameter settings.
Population SizeCrossover RateMutation RateMax Generations
100 0.80 0.01 20
Table 7. Comparison of training error, training accuracy, and test accuracy for UOCRs using different pre-trained models.
Table 7. Comparison of training error, training accuracy, and test accuracy for UOCRs using different pre-trained models.
Training LossTraining AccuracyTest Accuracy
ResNet18 1.164 × 10−195.94%80.21%
ResNet341.024 × 10−195.94%79.95%
ResNet507.723 × 10−296.95%85.20%
ResNet101 4.925 × 10−298.08%85.20%
ResNeXt50-32X4D6.914 × 10−297.20%81.25%
ResNeXt101-32X4D1.129 × 10−195.46%76.46%
ViT-Base2.914 × 10−189.94%75.11%
Table 8. Experimental results of the auxiliary cost surrogate model.
Table 8. Experimental results of the auxiliary cost surrogate model.
Train RMSETest RMSETest MAPE
WeightedEnsemble1.625 × 10−33.706 × 10−34.312%
LightGBMXT1.694 × 10−33.800 × 10−34.428%
CatBoost1.706 × 10−33.728 × 10−34.386%
LightGBM1.708 × 10−33.776 × 10−34.356%
LightGBMLarge1.780 × 10−33.678 × 10−34.277%
ExtraTrees1.812 × 10−33.416 × 10−33.990%
KneighborsDist1.857 × 10−34.080 × 10−34.683%
NeuralNetTorch1.927 × 10−33.855 × 10−34.324%
RandomForest1.937 × 10−33.643 × 10−34.246%
KneighborsUnif1.970 × 10−34.066 × 10−34.663%
XGBoost2.006 × 10−33.563 × 10−34.173%
NeuralNetFastAI2.348 × 10−33.338 × 10−33.831%
Table 9. Comparison between the generated cost-effective CBS and the original strategy across six unit conditions in terms of total coal flow, proportion of A coal, proportion of B coal, proportion of C coal, total cost, coal cost, and other costs.
Table 9. Comparison between the generated cost-effective CBS and the original strategy across six unit conditions in terms of total coal flow, proportion of A coal, proportion of B coal, proportion of C coal, total cost, coal cost, and other costs.
Total Coal Flow Rate
(t/h)
A Coal PercentageB Coal PercentageTotal Cost
(CNY/
(kW·h))
Coal Cost
(CNY/
(kW·h))
Other
Cost
(CNY/
(kW·h))
Cond. 1 Orig.88.80359.53%12.82%0.4570.3820.075
Cond. 1 Gen.89.62046.41%12.23%0.4310.3560.075
Cond. 2 Orig.122.52144.47%13.43%0.5200.4400.080
Cond. 2 Gen.118.31849.01%7.83%0.5010.4260.075
Cond. 3 Orig.147.90635.48%19.04%0.5620.4870.075
Cond. 3 Gen.135.08046.05%11.57%0.5260.4520.074
Cond. 4 Orig.164.57344.22%19.42%0.5790.5090.070
Cond. 4 Gen.168.00540.64%17.60%0.5610.4930.069
Cond. 5 Orig.174.44443.09%37.92%0.5810.5120.069
Cond. 5 Gen.186.07748.02%8.71%0.5780.5110.067
Cond. 6 Orig.200.38543.91%38.05%0.5970.5270.070
Cond. 6 Gen.197.12945.76%37.75%0.5920.5210.071
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Wu, S.; Wang, T.; Ding, J. A Surrogate-Assisted Intelligent Adaptive Generation Framework for Cost-Effective Coal Blending Strategy in Thermal Power Units. Electronics 2025, 14, 561. https://doi.org/10.3390/electronics14030561

AMA Style

Wang X, Wu S, Wang T, Ding J. A Surrogate-Assisted Intelligent Adaptive Generation Framework for Cost-Effective Coal Blending Strategy in Thermal Power Units. Electronics. 2025; 14(3):561. https://doi.org/10.3390/electronics14030561

Chicago/Turabian Style

Wang, Xiang, Siyu Wu, Teng Wang, and Jiangrui Ding. 2025. "A Surrogate-Assisted Intelligent Adaptive Generation Framework for Cost-Effective Coal Blending Strategy in Thermal Power Units" Electronics 14, no. 3: 561. https://doi.org/10.3390/electronics14030561

APA Style

Wang, X., Wu, S., Wang, T., & Ding, J. (2025). A Surrogate-Assisted Intelligent Adaptive Generation Framework for Cost-Effective Coal Blending Strategy in Thermal Power Units. Electronics, 14(3), 561. https://doi.org/10.3390/electronics14030561

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop