Next Article in Journal
Applications of Virtual Reality to Support Social Communication in Individuals with Autism: A Systematic Review of Immersive Interventions
Previous Article in Journal
ZeroDay-LLM: A Large Language Model Framework for Zero-Day Threat Detection in Cybersecurity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrated Energy Short-Term Adaptive Load Forecasting Method Based on Coupled Feature Extraction

School of Information Science and Engineering, Northeastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Information 2025, 16(11), 940; https://doi.org/10.3390/info16110940 (registering DOI)
Submission received: 16 September 2025 / Revised: 26 October 2025 / Accepted: 27 October 2025 / Published: 29 October 2025
(This article belongs to the Section Artificial Intelligence)

Abstract

Integrated energy load forecasting plays a crucial role in optimizing the operation and economic dispatch of integrated energy systems. Its forecasting accuracy is not only time-dependent but also influenced by the coupling characteristics among energy sources. Solely relying on time-scale training methods cannot adequately capture the strong correlations among multiple energy sources. To address challenges in extracting coupled load forecasting features, obtaining periodic characteristics, and setting model network structures, this paper proposes an Integrated Energy Short-Term Adaptive Load Forecasting Method Based on Coupled Feature Extraction (AP-CFE). This approach integrates high-dimensional coupling features and periodic temporal features effectively using ensemble algorithms. To prevent overfitting or underfitting issues, an Adaptive learning algorithm (AP) is introduced. The load demonstrates highly stochastic behavior in response to external factors, resulting in rapid, volatile fluctuations in grid demand. The strategy of employing sparse self-attention to approximate the residual terms effectively mitigates this issue. Simulation results using comprehensive energy load data from Australia demonstrate that the proposed model outperforms existing models, achieving better capture of energy coupling characteristics with average absolute percentage errors reduced by 20.75%, 28.48%, and 21.64% for electricity, heat, and gas loads, respectively.

1. Introduction

The tight coupling of multiple energy sources across production, distribution, storage, and consumption endows integrated energy systems with operational characteristics fundamentally distinct from those of standalone systems. These systems are characterized by multi-physics coupling, dynamic interrelationships across multiple timescales, strong nonlinearity, and inherent uncertainty [1,2]. Consequently, integrated energy systems have attracted considerable global research interest, encompassing critical areas such as planning and design [3], load forecasting [4], energy management [5], and monitoring and maintenance [6].
Among these, load forecasting serves as the foundation for operational optimization and economic dispatch in integrated energy systems, and Load forecasting has been a hot research topic in integrated energy systems in recent years [7,8]. Currently, the mainstream approaches can be categorized into two types: those that solely consider the temporal variation patterns and those that account for the characteristic coupling of multiple energy sources. At present, the former category has several implementation approaches, which can be further divided based on whether the temporal characteristics of the data are decomposed. For example, in references [9,10], the historical load data is not decomposed, but instead, traditional fuzzy theory is employed for long-term forecasts of planning issues in the transmission network and various scenarios.
With the continuous development of signal processing techniques and neural networks, researchers aim to extract different frequency band state information by analyzing the fluctuation characteristics of historical load data [11,12,13,14,15]. In reference [11], wavelet packet decomposition is applied to daily maximum load values and seasonal data, separating high-frequency and low-frequency feature data, and neural networks are used to predict each decomposition frequency band. In reference [12], a household-level short-term load forecasting model is proposed to understand residents’ daily routine. This method utilizes Long Short-Term Memory (LSTM) networks and wavelet packet decomposition of load nodes and neighboring nodes to capture highly correlated historical and future information, enhancing the prediction accuracy of the model. However, the use of fixed wavelet packet decomposition basis functions limits the adaptability of this forecasting model. References [13,14] address the non-periodic, nonlinear, volatile, and abrupt stochastic nature of the load data by employing the empirical mode decomposition method to obtain regular and stationary intrinsic mode components. However, this method lacks a solid theoretical derivation and fails to solve the endpoint effect problem. Reference [15] incorporates a vector feature reconstruction process, which effectively addresses the impact of outliers on the overall model training. However, this leads to extreme values in model accuracy, making the prediction errors unable to continuously decrease.
However, there is currently relatively limited research on the impact of considering the coupling characteristics among multiple energy sources on the accuracy of load forecasting in integrated energy systems. References [16,17,18] apply Convolutional Neural Network (CNN) to load forecasting and extract the spatial characteristics of the data. However, directly using a CNN for feature extraction can disrupt the original temporal features [16], leading to a decrease in prediction accuracy. Reference [17] employs multi-task learning to predict the load in integrated energy systems, considering both the coupling characteristics among multiple energy sources and the temporal features. Although a temporal model is incorporated after feature extraction, it does not effectively address the issue of disrupted temporal features. Reference [18] utilizes ensemble learning to weight the predictions of different models. However, due to differences in the training data of each model and the lack of mutual constraints among the models, the prediction accuracy of the models is easily influenced by the features of the training set.
A common limitation in prior load forecasting studies for integrated energy systems is their frequent failure to capture the inherent long-term oscillatory behavior of the data. However, accurate extraction of periodic information not only reveals the patterns in the current data but also enables the learning of future data trends. Therefore, this paper proposes a short-term adaptive load forecasting method for integrated energy systems based on coupled feature extraction. This method considers both the coupling and periodicity of the data, analyzing the intrinsic correlations among different energy sources while preserving the temporal dependence on the data timeline. Additionally, the adaptive algorithm is proposed to enhance the network’s generalization ability, addressing overfitting issues when training on small datasets and improving the model’s generalization capability.

Contribution

This architecture is designed to enable precise forecasting of integrated energy systems. Among these: The two-dimensional feature extraction model is used for extracting temporal features and coupled features, thereby reducing feature loss; AP enable the adaptive adjustment of neural network parameters; Finally, the residuals are mitigated by employing a sparse self-attention mechanism. Our main contributions are summarized as follows:
(1)
In order to investigate the coupling characteristics of multi-energy systems, Dynamic Time Warping (DTW) is employed to analyze temporal delays and oscillatory behaviors within load data.
(2)
Periodic characteristics are systematically quantified through multi-timescale information fusion. A two-dimensional feature extraction model is then employed to enhance the correlation of the coupled features within their respective cycles.
(3)
AP accommodate dynamic network structures, thereby mitigating issues of overfitting and underfitting.

2. Analysis of Integrated Energy System Characteristics

In this paper, the analysis is conducted using the load data from the winter and summer seasons in New South Wales, Australia. The energy infrastructure in this region includes combined heat and power plants, electrical conversion devices, thermal storage systems, gas storage facilities, cooling stations, and energy transmission networks, which collectively fulfill the electricity, heating, and gas supply demands throughout the region. Figure 1 illustrates the coupling relationships among various energy transmission systems.
Data from integrated energy systems exhibit both strong temporal dynamics and significant spatial correlations. Therefore, in analyzing load variation patterns, both periodic characteristics and inter-energy coupling properties are of paramount importance. Utilizing the correlations among various loads leads to improved accuracy in integrated energy load forecasting.

3. Analysis of Integrated Energy System Coupling Characteristics

In an integrated energy system, thermal and gas loads exhibit inertia in comparison to electrical loads, meaning they experience delayed fluctuations. In other words, the fluctuations in thermal and gas loads are influenced by the fluctuation patterns of electrical loads, and the correlation of the entire sequence evolves over time. Additionally, the vibration of thermal and gas loads is stable, and they do not experience drastic load fluctuations in the short term. DTW accounts for the delayed oscillation patterns in load data. Offline training is employed to preserve the coupling characteristics of the sequence as much as possible.
DTW works by identifying corresponding points of fluctuation between time series and calculating a relative distance to measure their correlation, enabling better handling of time-shifted relationships in time series data. In most cases, the oscillations of integrated energy loads appear very similar as a whole, but their initial starting points occur at different times, exhibiting a certain degree of delay. Therefore, before calculating the correlation between two time series, one or more of the time series are distorted or adjusted, either compressed or stretched, along the time axis in order to compute the correlation of the time series.
Given two time series, P n = p 1 , p 2 , p n and Q m = q 1 , q 2 , q m , to compute their distance, we first need to generate an m × n matrix, denoted as w , where each element w i j represents the Euclidean distance between Q i and P j , as specified in Equation (1).
w i j = ( Q i P j ) 2
DTW finds the shortest distance path from one sequence to another in the distance matrix w . This step can be solved using dynamic programming (DP), summing up the shortest path from the top-left corner of the matrix to the bottom-right corner. Assuming the matrix is w , and the shortest path from the top-left corner 1 , 1 to any coordinate point i , j is L m i n ( i , j ) , the optimal path can be obtained through recursive Equation (2).
L m i n ( i , j ) = m i n L m i n ( i , j 1 ) , L m i n ( i 1 , j ) , L m i n ( i 1 , j 1 ) + W ( i , j )
Employing DTW on multi-energy data to investigate system coupling, the optimal dynamic warping paths were derived and are visualized in Figure 2. The shortest distances d between electricity, heat, and gas are, respectively, illustrated in Table 1. From Figure 2, it can be observed that both heat and gas loads exhibit significant delays compared to electricity loads, with heat loads being delayed relative to gas loads. The relatively low DTW distances between electricity-heat and heat-gas loads, presented in Table 1, indicate a strong coupling relationship after temporal warping.

4. Two-Dimensional Feature Extraction Network

4.1. Multi-View Coupling Feature Extraction Model

Accounting for the nonlinear characteristics of multi-energy coupling, temporal data coupling analysis is performed using sliding convolutional windows. In practical multi-energy systems, coupling effects inherently exhibit temporal delays, the duration of which varies significantly across different operational contexts, time periods, and energy types. A key limitation of the single-layer sliding window approach is its fixed-size receptive field, which is inadequate for capturing the complex coupling characteristics of multi-energy systems. Under a fixed kernel size, the receptive field (RF) can be continuously expanded by adding more convolutional layers. The coupling feature extraction model is illustrated in Figure 3.
The different field coupling features mapped by multiple layers of sliding convolutional windows can be analyzed using the following formula:
R n = R n 1 + ( k n 1 1 ) × S n
where R n represents the field of view of the nth layer coupling feature; k n 1 denotes the convolution kernel size of the n − 1 layer; S n represents the sliding window stride of the nth layer. In the multi-field coupling feature extraction model proposed in this paper, the convolution kernel and stride are set to fixed values of 3 × 3 and 1, respectively.
After multiple iterations of sliding windows, the result of the nth layer iteration is as follows:
R n = R n 1 + ( k 1 ) × S R n 1 = R n 2 + ( k 1 ) × S R 1 = 1 + ( k 1 ) × S
R n = n × S × ( k 1 ) = 1 + 2 n
To reasonably allocate the weights of coupling features under different fields of view, the associated distance L i of the i layer iteration result is calculated through DTW, followed by s o f t m a x activation function. The specific formula is as follows:
α i = s o f t m a x ( 1 / L i )
a = i = 1 n α i × c i
where s o f t m a x = e x i / ( j e j ) . α i represents the weight vector of coupling features; c i denotes the result of the i layer iteration; a represents the coupling feature vector.

4.2. Two-Dimensional Temporal-Coupling Feature Extraction Model

Data used for load forecasting incorporates both coupled relationships across variables and critical temporal information. The two-dimensional feature extraction network (TDN) nonlinearly fuses temporal features and coupling features through multiple gating units. This gate-mediated recombination enables the simultaneous selection of spatiotemporal characteristics and historical patterns. The specific structural diagram principle is shown in Figure 4.
The integration of two-dimensional input features is achieved through the actions of two gating units, with the specific calculation formula as follows:
n τ = σ ( W n [ a τ a τ ] ) + b n
X τ = t a n h ( W c [ a τ a τ ] ) + b X
X τ = n τ X τ
where σ ( ) represents the sigmoid activation function; b denotes the network bias; a τ and a τ respectively represent the coupling features and temporal features; W signifies the network weight; tanh is the activation function; denotes the element-wise Hadamard product.
In temporal feature extraction models, information transmission is constrained by distance, leading to feature loss throughout the training and learning process. To store long-distance dependency information, multiple time-scale hidden state fusion is adopted to enhance the extraction of long-distance features for downstream tasks.
The feature deletion and memory of the TDN are achieved using backpropagation neural networks, which possess memory, retention, forgetting, and fusion capabilities, making them more suitable for handling temporally related tasks, particularly those associated with periodicity. Its forward propagation mechanism is analogous to that of LSTM networks, leveraging both memory cells and hidden states to preserve historical sequence information. The memory cells, controlled by three gating units, store the historical information of the sequence data. The formula is as follows:
f τ = σ ( W f h τ 1 x τ ) + b f
i τ = σ ( W i h τ 1 x τ ) + b i
o τ = σ ( W o h τ 1 x τ ) + b o
C τ = t a n h ( W c h τ 1 x τ ) + b c
C τ = f τ C τ 1 + i τ C τ
h τ = o τ t a n h C τ
y = h τ W h + b h
The equation comprises: C τ as the candidate memory unit used for updating the memory unit; h τ represents the hidden state information; and y stands for the output prediction result.
To address the issue of loss of periodic information due to long-distance data transmission, a feature periodic correlation vector is constructed, as shown in the formula below:
H = [ h τ 1 , h τ , , h τ + ( t 1 ) t + 1 ]
β τ 1 = ( h τ 1 h τ + t ) / i = τ 1 τ + ( t 1 ) β i
w = [ β τ 1 , β τ , , β τ + ( t 1 ) t + 1 ]
y τ + t = H w T = i = τ 1 τ + ( t 1 ) β i h i
where H denotes the hidden state information matrix; β τ 1 signifies the weight parameter of the hidden state feature; w represents the weight matrix; and y τ + t denotes the model’s output result.
The principles of two-dimensional temporal feature extraction retain the network’s memory function, where each time step’s hidden state possesses all previous time step features and oscillation patterns. The correlation of load at different positions within the sequence is enhanced, corresponding to an increase in the correlation of hidden state information within the respective period. Therefore, through the fusion of information across multiple time scales, it is relatively easy to compute periodic features, thereby better addressing the long-distance dependency issue in temporal networks and resolving the network’s dependence on the last time information. Obtaining global information enables the model to efficiently integrate multidimensional features.
Embedding encoding is employed to transform categorical features, such as weather, holidays, and electricity prices, into a spatial mapping, thereby enhancing the representation of inter-feature correlations. Model employs a sparse self-attention mechanism that amalgamates the expansion and Topk approaches. This integration allows for the simultaneous capture of global and salient information while effectively alleviating the computational burden associated with the self-attention’s spatial complexity.

4.3. Adaptive Learning Algorithm

In recent years, amidst the continuous evolution of neural networks, there has been a growing focus on adjusting network structures to achieve greater reductions in loss functions; Enhancing optimization algorithms to accelerate the rate of reduction in model loss functions; And handling input data to mitigate the impact of dimensionalities on network training. While existing methods enhance model accuracy and convergence rates, they lack mechanisms for dynamic network architecture adaptation.
AP effectively mitigates these limitations. At its core, AP employs a neural self-growth mechanism that enables dynamic adaptation of network size based on training data volume. The implementation of AP algorithms improves both the transferability and structural adaptability of network models. The structural principles of AP algorithms are depicted in Figure 5.
The essence of the adaptive algorithm lies in activating certain neurons through a mechanism when the training data volume reaches a certain upper limit, enabling their participation in subsequent network training. In summary, the formula for the network structure employing the adaptive algorithm is as follows:
r ( l ) = r e n e w p ( l )
r e n e w p ( l ) = r min   x d a t a x min       r min + x d a t a x min p x min < x d a t a < x max r max   x max x d a t a
where r ( l ) denotes the number of neurons in l layer; p represents the threshold of data volume for adding neurons, i.e., every layer adds 16 neurons when the additional training data volume reaches p; r min and r max respectively represent the minimum and maximum network structures; x d a t a represents the real-time incoming training data volume, x min is the maximum data volume that the minimum network structure can be trained on, and x max represents the maximum data volume that the network can handle. In this paper, based on existing research [19,20], p is set to 600, r min is set to 16, and x min is set to 600, while x max is determined by the maximum neuron structure of the network.
ω ˜ ( l ) = R ( l ) ω ( l )
R ( l ) = 1 , 1 , , 1 , r ( l ) 0 , , 0 R 1 × r max
z ( l + 1 ) = y ( l ) × ( 1 m ω ( l ) ) + b ( l + 1 )
y ( l + 1 ) = f z ( l + 1 )
where ω ( l ) R 1 × r max represents the weight of layer l , ω ( l ) is the weight of layer l after the adaptive algorithm, y ( l ) R m represents the output of layer l , b ( l ) R r max is the bias of layer l , z ( l + 1 ) is the output of layer l + 1 before the activation function f · , 1 m is an m-dimensional column vector with all elements being 1, and represents the Kronecker product.
The pseudocode for AP-CFE is shown in Algorithm 1. To achieve full-view observation for convolutions, compute the number of convolutional layers L. Across all training datasets, the AP algorithm and matrix multiplication are utilized to perform forward propagation. The backpropagation algorithm is then applied to minimize the error, leading to iterative updates of all convolutional kernel parameters. Embedding encoding maps the features of external factors (weather, holidays, electricity prices) into dense vector representations, which effectively captures the implicit relationships between them. Sparse self-attention mechanisms enable dense operations without compromising performance, significantly reducing runtime and minimizing memory usage. The training process was terminated after 100 epochs when the model’s validation error stabilized, indicating convergence.
Algorithm 1: AP-CFE Algorithm
Information 16 00940 i001

5. Case Analysis

5.1. Data Preprocessing

In a Python 3.11.8 simulation environment, the network architecture was constructed using TensorFlow 2.8. The experiment is conducted within the context of Denmark and New South Wales, Australia. Data samples are sourced from the Australian Energy Market Operator (https://www.aemo.com.au/) and the Australian Bureau of Statistics(https://www.energy.gov.au/). Due to the varying coupling relationships among different energy sources, a singular linear relationship cannot be employed for qualitative and quantitative analysis. Therefore, a multi-feature extraction approach is utilized to incorporate nonlinear computations, thereby delineating multidimensional characteristics among energy sources. The study employs load data from 1 August 2020, to 1 September 2020, with a sampling interval of 30 min (yielding 1536 data points). Input data comprise historical load data from the preceding 10 h. The data is partitioned into training, validation, and test sets in a ratio of 6:3:1, respectively. Preprocess data using maximum-minimum normalization [21]. The data preprocessing process is outlined as follows:
x s t d = x x min x max x min
where z represents the maximum sample data within each dataset, x min denotes the minimum sample data, and x s t d signifies the normalized value.

5.2. Evaluation Metrics

The predictive performance of the load forecasting model is influenced by both predictive accuracy and prediction time. This study adopts the Mean Absolute Percentage Error (MAPE), Root Mean Square Error (RMSE) and Standard deviation (S) as evaluation metrics for the model’s predictive accuracy. The definitions of MAPE, RMSE and S are as follows:
M A P E = 1 n i = 1 n y i y ^ i y i × 100 %
R M S E = 1 n i = 1 n ( y i y ^ i ) 2
S = i = 1 n ( X i X ¯ ) 2 n
A c c u r a c y = 1 M A P E
where n represents the number of predicted samples, y i and y ^ i respectively denote the true value and the predicted value of the i th sample, X i represents prediction error, X ¯ represents the average prediction error and Accuracy denotes the model’s accuracy.

5.3. Evaluation Metrics Optimization Analysis of LSTM, CNN, TDN, and AP-CFE

To validate the contributions of coupled feature extraction, AP algorithms, and periodic feature extraction, we compare the LSTM, CNN, TDN, and the proposed AP-CFE model. The comparison results of the four models are shown in Figure 6.
Overall, in Figure 6, the electric load fluctuates most significantly in the multi-energy system, while the thermal load exhibits the most stable changes. The coupled features in the multi-energy system lead to delays in the thermal and gas loads compared to the electric load. Short-term downward trends in the electric load affect the changes in the thermal and gas loads. Among them, in the multi-energy system, the short-term variation range of the thermal load is the smallest, and the electric load is significantly affected by various factors, showing pronounced oscillations.
From Figure 6a, it can be observed that during the decline of the electric load, there is a significant decrease in load at the 20th sampling point, followed by a trend of oscillation and then an increase in load. In the 134 sampling points of the testing dataset, the span of the electric load’s extremes is 30 MW. Figure 6b shows the variation in the thermal load, which remains relatively stable overall. Only at the 55th sampling point, there is a downward trend in load, which lags behind the electric load. The fluctuation pattern of the electric load affects the oscillation of the thermal load, making it the slowest-changing load among the multi-energy loads. The energy conversion and transmission of electric and thermal loads are implemented by CHP (Combined Heat and Power), while heating equipment such as air conditioning is electrically driven, increasing the correlation between electric and thermal loads. The predicted results of the gas load are shown in Figure 6c, with the overall fluctuation ranging from 75 MW to 96 MW, and the inflection point of load decline occurring at the 35th sampling point, between the electric and thermal loads.
Among the four prediction models, the CNN model exhibits the most pronounced lag phenomenon, with the lowest accuracy, largest mean error, and dispersed distribution. The TDN model greatly improves the decrease in prediction accuracy caused by LSTM’s inability to extract coupled and periodic features. However, the inclusion of the CNN model leads to more complex backpropagation computations in the TDN model, requiring longer model training times.
The high coupling and periodicity of electricity loads result in a significant decrease in errors when employing the ensemble TDN. The relatively stable fluctuation patterns of heat and gas loads can reduce the complexity of model learning, with the LSTM model performing best in fitting the low-coupling gas load. From Figure 6 and Table 2, it can be observed that all four models can reasonably predict the approximate variation curves of the three loads. However, the proposed model (AP-CFE) can better describe the load fluctuation and anomaly features, with a more concentrated error distribution and relatively smaller MAPE and RMSE.
The AP-CFE enhanced by an adaptive learning algorithm with neuron self-growth mechanisms, enables optimization training that better conforms to the structural features of the data, while saving more time for training. Compared to the LSTM model, the proposed AP-CFE model reduces MAPE values by 34.03%, 34.68%, and 36.36%.
Table 3 shows the computational efficiency of the models. Given the same amount of data, LSTM typically train the fastest and consume the least memory; CNN train the slowest and consume the most memory; TDN and AP-CFE fall between the two.

5.4. Model Comparison

To validate the superiority of the proposed method in terms of prediction accuracy and speed, the AP-CFE method is compared with methods from reference [18] (Bi-LSTM), ref. [16] (PatchTST), ref. [15] (FEDFormer), ref. [17] (CNN-GRU), and ref. [14] (EMD). The parameter settings for the five comparison prediction models are listed in Table 4.
Figure 7 illustrates the prediction curves of electricity, heat, and gas loads for each model. To evaluate the prediction accuracy of the models, the percentage error is used to analyze the error distribution of load prediction models, as shown in Figure 7.
From Figure 7 and Table 5, it can be observed that the FEDFormer model has a good fitting effect; the Bi-LSTM has the worst fitting effect because the fully connected layer cannot extract temporal or spatial features effectively; the EMD model decomposes and reconstructs the input data, ignoring the changing patterns of anomalies and outliers, which leads to inaccurate predictions for certain special values; the prediction effect of the CNN-GRU model deteriorates in the gas load, mainly due to the decreased coupling of gas load compared to electricity load.
Additionally, the energy conversion between electricity, heat, and gas greatly increases the difficulty of feature extraction for electricity load, resulting in lower overall prediction accuracy for electricity load. The error of the comparison model differs significantly from that of AP-CFE, indicating that the predictive accuracy of the comparison model needs improvement.
To evaluate the model’s generalization capability, additional simulation comparison experiments were conducted in the Danish region. The results are shown in Figure 8 and Table 6. The p-value of less than 0.05 indicates that the error of the AP-CFE model differs significantly from that of other models. The AP-CFE model also demonstrated the best predictive performance in the Danish region. The predictive performance of Bi-LSTM outperforms that of EMD, which differs from the forecast results in New South Wales, Australia. Meanwhile, PatchTST’s predictive performance is the closest to AP-CFE.

6. Conclusions

This study integrates a multi-modal feature extraction framework with an AP algorithm for comprehensive energy load forecasting. The proposed approach aims to enhance the prediction accuracy of electricity, heat, and gas loads while improving model adaptability and computational efficiency.
The TDN architecture enhances the network’s capacity for nonlinear modeling, enabling the effective extraction of high-dimensional temporal, periodic, and coupling features from load data. Training three sets of networks with identical structures but different weights and biases enables the model to meet the requirements of different load types. Each network extracts both temporal variations and spatial coupling features. Sparse self-attention effectively reduces the spatial computational complexity of self-attention.
Compared to existing load forecasting models, the proposed AP-CFE model more effectively captures the dynamic patterns of integrated energy loads and provides enhanced characterization of multi-energy temporal dynamics and coupling relationships. This enables adaptive updates of the network structure, reducing both training and prediction times, and enhancing the model’s stability, accuracy, and adaptability. Load is influenced by multiple factors including weather, holidays, market electricity prices, and renewable energy output. Future model architectures must adopt a unified framework capable of naturally processing multimodal inputs while simultaneously accomplishing multiple related tasks.

Author Contributions

Y.Q. conceived the idea for the manuscript, B.H. (Bonan Huang), Y.Z. (Yameng Zhang), L.W. (Luyuan Wang) and J.T. (Jiaqi Tian) wrote the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 52377079.

Data Availability Statement

Data downloads are available at links (https://www.aemo.com.au/) and (https://www.energy.gov.au/). Data processing are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, L.; Song, H.; Yu, Q. The Application of Hybrid Deep Learning Method of Improved Transformer Model in Load Forecasting Based on Future Network and Edge Computing. Int. J. High Speed Electron. Syst. 2025, 2540848. [Google Scholar] [CrossRef]
  2. Xu, B.; Zhang, Y. Development Path of Regional Integrated Energy System; IOP Publishing: Bristol, UK, 2019. [Google Scholar]
  3. Liu, L.; Wang, D.; Hou, K. Region model and application of regional integrated energy system security analysis. Appl. Energy 2025, 260, 114268. [Google Scholar] [CrossRef]
  4. Liu, X.; Wu, K.; Qi, L. Multi-Objective Optimization Design of Regional Integrated Energy System; IOP Publishing: Bristol, UK, 2020. [Google Scholar]
  5. Lorenzo, B.; Stefano, C.; Vincenzo, M. Short-term forecasting method to improve the performance of a model predictive control strategy for a residential hybrid renewable energy system. Energy 2019, 172, 997–1004. [Google Scholar]
  6. Andersen, F.; Larsen, H.; Gaardestrup, R. Long term forecasting of hourly electricity consumption in local areas in Denmark. Appl. Energy 2013, 110, 147–162. [Google Scholar] [CrossRef]
  7. Duan, Q.; Liu, J.; Zhao, D. Short term electric load forecasting using an automated system of model choice. Int. J. Electr. Power Energy Syst. 2017, 91, 92–100. [Google Scholar] [CrossRef]
  8. Moazzami, M.; Khodabakhshlan, A.; Hooshmand, R. A new hybrid day-ahead peak load forecasting method for Iran’s National Grid. Appl. Energy 2013, 101, 489–501. [Google Scholar] [CrossRef]
  9. Maryam, I.; Hassan, G. Residential load forecasting using wavelet and collaborative representation transforms. Appl. Energy 2019, 253, 113505. [Google Scholar] [CrossRef]
  10. Mohan, N.; Soman, K.; Kumar, S. A data-driven strategy for short-term electric load forecasting using dynamic mode decomposition model. Appl. Energy 2018, 232, 229–244. [Google Scholar] [CrossRef]
  11. Yang, D.; Guo, J.; Sun, S. An interval decomposition-ensemble approach with data-characteristic-driven reconstruction for short-term load forecasting. Appl. Energy 2022, 306, 117992. [Google Scholar] [CrossRef]
  12. Yu, F.; Xu, X. A short-term load forecasting model of natural gas based on optimized genetic algorithm and improved BP neural network. Appl. Energy 2014, 134, 102–113. [Google Scholar] [CrossRef]
  13. Yin, L.; Xie, I. Multi-temporal-spatial-scale temporal convolution network for short-term load forecasting of power systems. Appl. Energy 2021, 283, 116328. [Google Scholar] [CrossRef]
  14. Wang, X.; Wang, S.; Zhao, Q. A multi-energy load prediction model based on deep multi-task learning and ensemble approach for regional integrated energy systems. Int. J. Electr. Power Energy Syst. 2021, 126, 106583. [Google Scholar]
  15. Jin, X.; Pan, T.; Yu, H. Electricity Load Forecasting Method Based on the GRA-FEDformer Algorithm. Energies 2025, 18, 4057. [Google Scholar] [CrossRef]
  16. Gong, M.; Yang, Y.; Huang, J.; Cui, H.; Jing, S.; Zhang, F. A novel wind power prediction model based on PatchTST and temporal convolutional network. Environ. Prog. Sustain. Energy 2025, 44, e14584. [Google Scholar] [CrossRef]
  17. Yuan, Q.; Li, H.; Pan, Y. Research on Hybrid Model for Short-term Power Load Forecasting Based on Error Compensation. IAENG Int. J. Comput. Sci. 2025, 52, 2719–2725. [Google Scholar]
  18. Wen, C.; Guang, H.; Hong, Z.; Lyuchao, L. Short-Term Load Forecasting with an Ensemble Model Using Densely Residual Block and Bi-LSTM Based on the Attention Mechanism. Sustainability 2022, 14, 16433. [Google Scholar]
  19. Zhong, C.; Liu, S.; Lu, Q.; Zhang, B.; Wang, J.; Xu, Q. Topological structural analysis based on self-adaptive growing neural network for shape feature extraction. Neurocomputing 2022, 503, 219–235. [Google Scholar] [CrossRef]
  20. Mazurov, M.; Mikryukov, A.; Gorbatih, M.; Bergaliev, T.; Shchukina, N. Information and neural educational system for training standard and selective neural network technologies in universities. CEUR Workshop Proc. 2021, 2919, 161–173. [Google Scholar]
  21. Viet, D.; Phuong, V.; Duong, M.; Tran, Q. Models for Short-Term Wind Power Forecasting Based on Improved Artificial Neural Network Using Particle Swarm Optimization and Genetic Algorithms. Energies 2020, 13, 2873. [Google Scholar] [CrossRef]
Figure 1. Illustrates the energy transmission relationships in an integrated energy system.
Figure 1. Illustrates the energy transmission relationships in an integrated energy system.
Information 16 00940 g001
Figure 2. Dynamic programming paths for multi-energy DTW.
Figure 2. Dynamic programming paths for multi-energy DTW.
Information 16 00940 g002
Figure 3. Coupling Feature Extractor.
Figure 3. Coupling Feature Extractor.
Information 16 00940 g003
Figure 4. Two-dimensional feature extraction network.
Figure 4. Two-dimensional feature extraction network.
Information 16 00940 g004
Figure 5. Adaptive algorithm structure. In this diagram, red and yellow represent two distinct layers of neurons, while gray denotes neurons that are not activated.
Figure 5. Adaptive algorithm structure. In this diagram, red and yellow represent two distinct layers of neurons, while gray denotes neurons that are not activated.
Information 16 00940 g005
Figure 6. Electric, Thermal, and Gas Load Prediction Curves.
Figure 6. Electric, Thermal, and Gas Load Prediction Curves.
Information 16 00940 g006aInformation 16 00940 g006b
Figure 7. Comparative curves of multi-energy Loads Forecasting in Australia.
Figure 7. Comparative curves of multi-energy Loads Forecasting in Australia.
Information 16 00940 g007
Figure 8. Comparative curves of multi-energy Loads Forecasting in Denmark.
Figure 8. Comparative curves of multi-energy Loads Forecasting in Denmark.
Information 16 00940 g008
Table 1. Multi-energy DTW distances.
Table 1. Multi-energy DTW distances.
Distance (d)Electrical LoadHeat LoadGas Load
Electrical load04.4116.947
Heat load4.41104.280
Gas load6.9474.2800
Table 2. Evaluation Metrics of Four Prediction Models.
Table 2. Evaluation Metrics of Four Prediction Models.
ModelLoad
PowerHeatGas
MAPE (%)RMSE (KW)MAPE (%)RMSE (KW)MAPE (%)RMSE (KW)
CNN2.343.012.242.842.393.10
LSTM1.912.431.732.171.652.07
TDN1.411.761.351.681.251.55
AP-CFE1.261.561.131.391.051.28
Table 3. Computational Efficiency of Four Models.
Table 3. Computational Efficiency of Four Models.
Data VolumeModel TypeFLOPsMemoryTraining Time
1500LSTM2.97 TFLOPs5.28 MB5.61 s
1500CNN3.52 TFLOPs6.88 MB6.45 s
1500TDN3.22 TFLOPs6.25 MB5.62 s
1500AP-CFE3.39 TFLOPs6.32 MB5.88 s
Table 4. Parameter Settings of Comparative Models.
Table 4. Parameter Settings of Comparative Models.
ModelParametersSettings
Bi-LSTMLayers2
Number(60, 100, 80)
FunctionRelu
PatchTSTpatches6
Step length10
Patch length10
CNN-GRULayers4
CNN-Number(64, 128, 256, 64)
GRU-Number(60, 100, 80)
FunctionRelu
FEDFormerNum attention heads8
Layers4
FunctionRelu
EMDLayers2
Number(60, 100, 80)
FunctionAdam/Relu
Table 5. Evaluation Metrics of Prediction Models in Australia.
Table 5. Evaluation Metrics of Prediction Models in Australia.
ModelCNN-GRUEMDBi-LSTMFEDFormerPatchTSTAP-CFE
PowerMAPE (%)1.792.592.231.591.701.26
RMSE (KW)2.253.302.832.002.131.56
S (KW)3.764.824.253.443.583.07
P0.0210.0090.0120.0360.022
HeatingMAPE (%)1.642.472.161.581.601.13
RMSE (KW)2.063.142.741.992.001.39
S (KW)3.524.654.103.403.452.81
P0.0260.0100.0170.0380.035
GasMAPE (%)1.812.582.301.341.531.05
RMSE (KW)2.323.282.921.671.911.28
S (KW)3.774.774.393.153.312.69
P0.0200.0090.0110.0420.029
Table 6. Evaluation Metrics of Prediction Models in Denmark.
Table 6. Evaluation Metrics of Prediction Models in Denmark.
ModelCNN-GRUEMDBi-LSTMFEDFormerPatchTSTAP-CFE
PowerMAPE (%)1.802.192.491.521.771.32
RMSE (KW)2.262.783.201.922.211.63
S (KW)3.704.284.733.363.663.00
P0.0210.0130.0090.0330.025
HeatingMAPE (%)1.672.202.461.631.661.20
RMSE (KW)2.122.793.122.062.091.48
S (KW)3.594.244.273.503.552.82
P0.0280.0190.0110.0320.030
GasMAPE (%)1.602.302.561.371.551.08
RMSE (KW)2.112.923.231.721.951.21
S (KW)3.574.394.763.163.402.61
P0.0310.0130.0090.0400.032
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, Y.; Huang, B.; Wang, L.; Tian, J.; Zhang, Y. Integrated Energy Short-Term Adaptive Load Forecasting Method Based on Coupled Feature Extraction. Information 2025, 16, 940. https://doi.org/10.3390/info16110940

AMA Style

Qin Y, Huang B, Wang L, Tian J, Zhang Y. Integrated Energy Short-Term Adaptive Load Forecasting Method Based on Coupled Feature Extraction. Information. 2025; 16(11):940. https://doi.org/10.3390/info16110940

Chicago/Turabian Style

Qin, Yidan, Bonan Huang, Luyuan Wang, Jiaqi Tian, and Yameng Zhang. 2025. "Integrated Energy Short-Term Adaptive Load Forecasting Method Based on Coupled Feature Extraction" Information 16, no. 11: 940. https://doi.org/10.3390/info16110940

APA Style

Qin, Y., Huang, B., Wang, L., Tian, J., & Zhang, Y. (2025). Integrated Energy Short-Term Adaptive Load Forecasting Method Based on Coupled Feature Extraction. Information, 16(11), 940. https://doi.org/10.3390/info16110940

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop