Next Article in Journal
Codepoietic Generation of Meaningful Information in the Evolving Biosphere
Previous Article in Journal
Normal Variance Mixture with Arcsine Law of an Interpolating Walk Between Persistent Random Walk and Quantum Walk
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Step Natural Gas Load Forecasting Incorporating Data Complexity Analysis with Finite Features

1
School of Management, Xi’an University of Architecture and Technology, Xi’an 710055, China
2
School of Economics and Management, Chongqing Normal University, Chongqing 401331, China
3
School of Mechanical and Electrical Engineering, Shaanxi University of Science and Technology, Xi’an 710021, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(7), 671; https://doi.org/10.3390/e27070671
Submission received: 7 May 2025 / Revised: 18 June 2025 / Accepted: 20 June 2025 / Published: 23 June 2025
(This article belongs to the Section Signal and Data Analysis)

Abstract

Data complexity directly affects the dynamics of complex systems, which in turn influences the accuracy and robustness of forecasting models. However, the load data exhibit complex features such as self-similarity, long-term memory, randomness, and chaos. This study aims to quantify and evaluate the complexity features of natural gas loads and to develop a multi-step-ahead forecasting model that integrates data decomposition and ensemble deep learning while considering these complexity features. Firstly, the complexity features of the series are quantified by rolling the fractal dimension, Hurst exponent, sample entropy, and maximum Lyapunov exponent. The analysis contributes to understanding data characteristics and provides information on complex features. Secondly, the ensemble learning eXtreme Gradient Boosting (XGBoost) can effectively screen complexity features and meteorological factors. Concurrently, variational mode decomposition (VMD) provides frequency-domain decomposition capability, while the gated recurrent unit (GRU) captures long-term dependencies. This synergy enables effective learning of local features and long-term temporal patterns, resulting in precise predictions. The results indicate that compared to other models, the proposed method (XGBoost-VMD-GRU considering complex features) demonstrates superior performance in forecasting, with R2 of 0.9922, 0.9860, and 0.9679 for one-step, three-step, and six-step prediction, respectively. This study aims to bring innovative ideas to load forecasting by integrating complex features into the decomposition forecasting framework.

1. Introduction

At present, the global energy emission reduction task is very urgent; the total global energy carbon emissions increased by 2%, exceeding 40 billion tons of carbon dioxide for the first time [1]. As an important means to mitigate climate change and reduce carbon emissions, the research and application of clean energy can promote sustainable economic development and enhance the security and stability of energy systems. In the energy transition, new energy sources, such as wind and solar energy, have the characteristics of instability, while natural gas has become an ideal transition energy to reduce carbon emissions and achieve net-zero emissions [2]. Natural gas can provide a reliable supplement to clean energy and make up for the instability of new energy sources. Its low carbon emissions, efficient combustion, flexible transportation modes, and abundant reserves make it play an important role in the energy transition [3]. According to the 73rd edition of the Statistical Yearbook of World Energy, global primary energy consumption reached 620 EJ in 2023, with fossil energy sources still dominating, accounting for 81.5% of the global primary energy structure. Although global natural gas demand remained relatively stable, with an increase of only 0.02%—possibly due to pandemic-related policies—natural gas consumption exhibited a steady growth trend until 2022. However, natural gas still plays an important role in the global energy structure as a less carbon-intensive fossil fuel [1]. Energy issues span from macro to micro levels, impacting national economic sustainability and the stability of people’s livelihoods. Accurately forecasting natural gas load (NGL) can optimize energy utilization, enhance supply security, reduce production costs, facilitate the formulation of sound energy policies, and promote economic development. Simultaneously, for industries and enterprises, forecasting aids in adjusting production and supply strategies to maintain market supply-demand equilibrium.
NGL forecasting models typically encompass statistical models, machine learning (ML) models, grey models, fuzzy logic models, and their combinations. Statistical models and ML models are widely used for short- to medium-term forecasting, leveraging their capacity to manage intricate market dynamics and fluctuations in supply and demand effectively [4]. Statistical models offer advantages in prediction tasks due to their rapid computational speed and ability to capture temporal trends [5,6]. However, they exhibit relatively weaker predictive capabilities when faced with complex data or nonlinear relationships [7]. In contrast, ML models can handle nonlinear and volatile data effectively, possessing strong generalization abilities and robustness against noise. Wei et al. utilized support vector regression (SVR) for short-term natural gas consumption (NGC) forecasting and optimized the model’s hyperparameters using a genetic algorithm (GA) [8]. Rehman et al. compared the performance of stepwise multiple linear regression, multilayer perceptron, and long short-term memory (LSTM) models in hourly natural gas demand forecasting. They emphasized LSTM’s advantages in predictive accuracy and management of extreme events [9]. Wang et al. applied the LMDI method to analyze influencing factors and developed a PSO-LSTM model to forecast natural gas consumption in China. The model demonstrated high prediction accuracy, and the results indicate a continued increase in future energy demand [10].
Deep learning techniques demonstrate powerful predictive capabilities in time series forecasting. However, when confronted with complex, highly volatile, and stochastic data, individual prediction algorithms often prove inadequate, and other techniques are needed to assist prediction to address these challenges effectively. Decomposition algorithms emerged from the need for complex data structures. The original load data are decomposed into simpler, more interpretable components. This facilitates more accurate identification of underlying patterns and trends. Ding et al. introduced the Dual Convolution with a Seasonal Decomposition Network, employing multiple seasonal decompositions using loss to address common seasonal variations and periodic fluctuations in NGC data. Experimental results indicate significant potential for the practical application of this approach [11]. Peng et al. combined local mean decomposition, wavelet threshold denoising, and LSTM in various configurations, discussing their application in NGC forecasting [12]. Jiang et al. utilized variational mode decomposition (VMD), wavelet packet decomposition, and LSTM to forecast complex nonlinear natural gas data. The model exhibited outstanding performance in experiments [13]. Wu and Wang significantly improved the quality of raw data preprocessing by using a two-stage decomposition strategy that combines EMD and EEMD. This approach further enhanced the extraction of valuable predictive information [14]. Zhao et al. proposed a deep learning model based on VMD-CNN-LSTM with a self-attention mechanism for natural gas load interval forecasting. They also developed a hierarchical early warning system to enhance prediction-based risk management [15].
In natural gas demand forecasting, it is crucial to consider the influence of external factors on prediction outcomes. Common short-term influencing factors include temperature, holidays, seasons, and others, which can significantly influence consumption patterns and production demands of natural gas. However, an excess of external factors may lead to increased redundancy in the model, thereby reducing the accuracy and interpretability of predictions. Therefore, effective feature selection is necessary. The most common method for feature selection is correlation analysis [16,17]. However, its effectiveness in complex real-world scenarios is often less than satisfactory. In contrast, feature selection based on embedding methods integrates the feature selection process into model training. By ranking the importance of features, it can effectively identify critical features [18]. Rao et al. employed random forest analysis to assess the importance of factors influencing China’s energy demand, effectively addressing data redundancy and multicollinearity, thereby offering a reliable method for energy demand forecasting [19]. Zeng et al. integrated the CatBoost algorithm into a hybrid prediction framework, significantly enhancing predictive performance by synthesizing the fluctuation parameters of GARCH models and the importance of meteorological conditions and other influencing factors [20].
With the advancement of data science and machine learning technologies, integrating complex features for NGL forecasting has emerged as a new research direction. Current research in natural gas forecasting primarily focuses on combining different models, with data processing limited to basic operations such as preprocessing and feature analysis of raw data. However, due to the inherent complexity and variability of data, relying solely on collected influencing factors may fail to capture all the information necessary for accurate load forecasting. Therefore, achieving more accurate predictions necessitates a deeper focus on the relationship between the complexity and predictability of time series data [21]. Self-similarity can characterize the analogous patterns of load data across various temporal scales. Randomness describes the degree of irregularity present in the load data. Karaca et al. integrated fractal analysis and entropy analysis in their predictive experiments using stock index datasets, achieving higher accuracy in stock data forecasts. This approach provides a representation of self-similarity and complexity for stock prediction [22]. Similarly, Kim et al. combined effective transfer entropy with machine learning algorithms to enhance the prediction of stock prices [23]. Long-term memory describes the persistent influence of past values on future values within load data. Chaos is utilized to assess the degree of chaotic behavior within a system, capturing the unpredictability of the data. Raubitzek and Neubauer found that the integration of complexity analysis results, such as approximate entropy, sample entropy, and fractal dimension, with machine learning outcomes significantly enhances the identification of predictable periods. They also explored the incorporation of complexity analysis from chaos theory to improve the accuracy of machine learning predictions [21]. In the field of natural gas, Bai et al. proposed a complexity measurement method that integrates coefficient of variation analysis, local fluctuation coefficient analysis, and kurtosis analysis. This experiment utilizes the comprehensive complexity of data to assess the impact of various factors on prediction accuracy [24]. Wei et al. employed the analytic hierarchy process to comprehensively address irregularities, complex cyclical variations, and volatility in consumption data. They established complexity measurement indicators to assess the relationship between complexity and predictive performance [25]. However, the above studies have not directly integrated specific complexity features of the data into the forecasting framework, thereby limiting their practical contributions to enhancing predictive performance. They have mainly focused on assessing the correlation between data of varying complexities and predictive performance, while a detailed exploration of the specific data characteristics of complexity features remains to be thoroughly investigated.
In summary, current research lacks a thorough exploration of how to organically integrate complexity features into forecasting frameworks. A thorough exploration and precise quantification of intrinsic complexity characteristics, such as self-similarity, long-term memory, randomness, and chaos in data, are still lacking. Considering these complex features comprehensively becomes a crucial factor in enhancing predictive accuracy in the study. Although NGL forecasting techniques have become increasingly mature, their predictive performance often remains unsatisfactory when handling data with varying levels of complexity. Moreover, these models tend to underperform in capturing long-term dependencies and volatile patterns, posing a critical challenge that requires further investigation.
To achieve more precise predictions, it is essential to delve into the complex features of historical sequences and their impact on predictive performance. By integrating embedding-based feature engineering with load decomposition techniques, predictive models effectively identify inherent data patterns and dynamic trends. This framework significantly improves forecasting accuracy and robustness. In this work, an innovative load-forecasting model that takes into account the characteristics of complexity is proposed. The main contributions of this study are as follows:
(1)
This study utilized fractal dimension, Hurst exponent, sample entropy, and maximum Lyapunov exponent to analyze the self-similarity, long-term memory, randomness, and chaotic characteristics of historical NGL data, thereby delving into the complexity of time series. These analytical findings not only contribute to a comprehensive understanding of data properties but also furnish crucial complexity feature information for predictive models.
(2)
This study represents the first attempt to organically integrate complexity features into a load-forecasting model, establishing a multi-step prediction model for daily NGC. By integrating GRU with XGBoost and VMD, significant improvements in prediction accuracy and robustness were achieved. Feature importance ranking using XGBoost confirms the significance of complexity features and certain meteorological factors in forecasting. In practical applications, multi-step preprocessing and prediction strategies are designed for complex long-term prediction needs.
(3)
This study rigorously verifies the effectiveness and feasibility of the proposed method using detailed NGL data collected from real-world sites spanning from 2016 to 2022. Additionally, to assess its robustness in practical applications, extensive simulation experiments were conducted by introducing various noise disturbances with different distributions into the original data. These experiments confirmed the superior performance of the method in real-world scenarios and demonstrated its strong adaptability to complex data and noise interference.
The rest of the paper is organized as follows. Section 2 describes the complexity features and the related methods of the proposed model. Section 3 elaborates on the main structure of the proposed hybrid model. Section 4 provides a detailed description of the experiments, focusing on feature importance, data decomposition, complexity feature analysis, and comparative result analysis. Finally, Section 5 summarizes the conclusions and proposes future directions.

2. Methods

2.1. Analysis of Data Complexity

2.1.1. Fractal Dimension

The fractal dimension (FD) is a geometric metric used to quantify the complexity of an object. It is commonly used to characterize the geometric properties of both natural and artificial complex structures. The FD method in time series data primarily characterizes the complexity of signals [26]. FD analyzes the internal structure of time series data within the time domain without requiring phase space reconstruction. Its objective is to reveal self-similarity and long-term dependencies in data, thereby further understanding the dynamic characteristics and behavioral patterns of the data.
This study employs the method proposed by Higuchi to compute the FD. The principle is analogous to fractal geometry [27], where complex geometric structures are decomposed into simpler substructures, and their complexity is described through scale transformations. In one-dimensional time series analysis, the FD is estimated by assessing the relationship between the scale length and the corresponding curve length across multiple resolutions. The FD is calculated by averaging these measurements, providing a quantitative description of the complexity of the time series. The underlying principle is as follows.
Step 1: Split the original sequence data in different degrees to form a new subsequence, X k m .
X k m = X ( m ) , X ( m + k ) , X ( m + 2 k ) , , X m + ( N m ) k k
where m refers to the initial time point, m = 1 ,   2 ,   ,   k ; k is the delay interval; and N is the length of the time series.
Step 2: Compute the curve length, L m k , of the subsequence X k m .
L m ( k ) = 1 k i = 1 N m k | X ( m + i k ) X ( m + ( i 1 ) k ) | N 1 [ N m k ] k
Step 3: Compute the average L k by averaging across different m values for the same k .
L ( k ) = 1 k   m = 1 k L m ( k )  
Step 4: Fit different values of k in a logarithmic coordinate system, where the slope represents the FD.
F D = ln ( L ( k ) ) ln ( 1 k )

2.1.2. Hurst Exponent

At different time scales, the long-term correlations of sequences vary. To assess the long-term memory of time series, this study employs rescaled range analysis (R/S analysis) to calculate the Hurst exponent (HE) of the time series [28].
The R/S measure is a classical and effective method used to assess long-term correlations by comparing the variations of time series across different scales, facilitating a deeper understanding of the underlying trends in time series data. The principle behind it is as follows.
Step 1: Divide the original sequence into i non-overlapping subseries X i j of length j .
Step 2: Compute the mean of each subgroup, X a ¯ , where a = 1 ,   2 , ,   i , and b = 1 ,   2 , ,   j .
X a ¯ = 1 j   b = 1 j X a b
Step 3: Within subsequence a , sequentially calculate the cumulative deviations Y a b for the first b points of the subseries.
Y a b = k = 1 b ( X a k X a ¯ )
Step 4: Calculate the range R a of subsequence a , representing the fluctuation amplitude of the sequence.
R a = max ( Y a b ) min ( Y a b )
Step 5: Compute the standard deviation S a of subsequence a .
S a = b = 1 j ( X a b X a ¯ ) 2 j 1
Step 6: Standardize the range by using the standard deviation to obtain the re-scaled sequence R S a = R a S a . Then compute the average ( R / S ) j of R S a for each subsequence. Repeat the above steps for subsequences of different lengths to obtain ( R / S ) j at different lengths j .
Step 7: Calculate the slope of the fitting curve l o g ( R S j ) ~ l o g ( j ) to obtain HE.

2.1.3. Sample Entropy

Entropy quantifies the disorder level within a system. Higher entropy values correspond to increased disorder, consequently reducing the predictability of system states. Sample entropy (SE) is a computational method in information theory that improves upon approximate entropy calculations. Richman and Moorman introduced this method for measuring time series complexity, aiming to address the limitations of approximate entropy that are dependent on data length [29]. In NGL prediction, the complexity of load data is characterized by computing SE, which further reflects the predictability of the data, describing its randomness and uncertainty. The underlying principle is as follows.
Step 1: For a time series of n elements [ x 1 , x 2 , , x n ] , generate a set of vector sequences X m i   = ( x i , x i + 1 , ,   x i + m 1 ) with subset dimension m, where i  = 1, 2, …, n m .
Step 2: Define a distance function d [ X m i , X m j ] , where j i .
d [ X m ( i ) , X m ( j ) ] = max k = 0 , 1 , 2 , , m 1 | X ( i + k ) X ( j + k ) |
Step 3: Count the number of indices j ( 1 ,   n m + 1 ) for which the distance between X m i   and X m j   is less than or equal to r, denoted as the approximate count ψ i m , r . Then, calculate the approximate ratio φ i m , r and define ϕ i m , r as the probability of matching m points between two sequences under a similarity threshold r.
φ i m , r = 1 n m + 1 ψ i m , r
ϕ m , r = 1 n m i = 1 n m φ i m , r
Step 4: Increase the dimension to m + 1 and repeat the aforementioned process to obtain ϕ m + 1 , r .
ϕ m + 1 , r = 1 n m i = 1 n m φ i m + 1 , r
Step 5: Calculate the SE.
S E = ln ( ϕ m + 1 , r ϕ m , r )

2.1.4. Maximum Lyapunov Exponent

The Maximum Lyapunov Exponent (MLE), introduced by Lyapunov [30], quantifies the exponential growth rate of nearby trajectories in a system. It characterizes the system’s sensitivity to initial conditions and serves as a measure of predictability in time series [31]. A larger value of the Lyapunov exponent indicates greater chaos in the time series and, correspondingly, lower predictability. This study exclusively computes MLE because it represents the maximum systemic chaos and sensitivity of sequences [32]. The commonly used algorithms for this purpose are the Wolf algorithm [33] and the Rosenstein algorithm [34]. A positive MLE indicates that the system exhibits hyperchaos [35,36], characterized by significant disorder and complexity. A negative value of MLE means that the system is stable. Since the research object of this paper is a set of NGL data, it is more robust to use Rosenstein’s algorithm to calculate MLE for longer time series. The specific calculation formula is as follows.
Step 1: For a given sequence data x i = [ x 1 , x 2 , x n ] , reconstruct the sequence points as X m τ ( i ) using attractors.
X m τ ( i ) = [ x i , x i + τ , , x ( i + ( m 1 ) τ ) ]
Step 2: Calculate the distance d i 0 between two neighboring points on the trajectory.
d i ( 0 ) = min X m τ ( i ) d ( X m τ ( j ) , X m τ ( i ) )
Step 3: When identifying nearest neighbors, the constraint requires that the time interval between them should exceed the average period of the time series. Here, λ represents the MLE, while C j denotes the initial interval.
d j ( i ) C j e λ ( i Δ t )
Step 4: Take the logarithm of both sides of the above equation.
ln ( d j ( i ) ) ln C j + λ ( i Δ t )
Step 5: The above equation represents a set of similarly parallel lines with slopes proportional to λ . Estimate the MLE by fitting the average line using the least squares method, where < > denotes the average of all j points.
y ( i ) = 1 Δ t ln ( d j ( i ) )
In summary, Table 1 comprehensively outlines the four complexity analysis metrics, detailing their theoretical definitions, typical value ranges, and interpretive significance for practical applications. The listed typical value ranges represent the commonly observed behavior of the four complexity analysis metrics in time series applications. They should not be regarded as their complete value domains under broader mathematical definitions.

2.2. Extreme Gradient Boosting

From an ensemble learning perspective, eXtreme Gradient Boosting (XGBoost) utilizes a combination strategy of multiple base learners, continuously optimizing the overall performance of the model through weighted strategies in each iteration [37]. Compared to other tree models, XGBoost demonstrates greater flexibility in feature engineering and it incorporates regularization techniques to enhance model generalization capabilities. This study utilizes XGBoost to rank feature importance, thereby effectively providing critical input features for subsequent prediction methods. During the process of tree splitting, the calculation of G a i n is shown in Equation (19):
G a i n = 1 2 G L 2 H L + λ + G R 2 H R + λ ( G L + G R ) 2 H L + H R + λ γ
Here, G L and G R denote the sum of first-order derivatives of the loss function for the left and right nodes, respectively. H L and H R denote the sum of second-order derivatives of the loss function for the left and right nodes, respectively. Specifically, j represents the node and i represents the sample of each node, G j = ( i I j ) g i , H j = ( i I j ) h i . g i represents the first-order derivative of the loss function, and h i represents the second-order derivative of the loss function. λ and γ are regularization parameters.

2.3. Variational Mode Decomposition

Short-term load data are inherently complex due to their diverse cyclic patterns, underlying trends, and the presence of noise. Therefore, accurate prediction requires careful analysis to effectively capture the relationships among these features. This paper introduces the variational mode decomposition (VMD) technique to handle complex NGL data, which effectively captures trend information within the data and mitigates noise interference on prediction outcomes. Moreover, the VMD offers controllable mode decomposition and flexible parameter adjustment capabilities, allowing adaptation to specific requirements. Furthermore, it demonstrates relatively high computational efficiency [38]. The detailed principles are outlined as follows.
The VMD transforms the decomposition process into an optimization problem, expressed through the following constraints:
min { u k } , { ω k } k = 1 k t [ ( δ ( t ) + j π t ) u k ( t ) ] e j w k t 2 2 u k = f ( t )
in which u k = { u 1 , u 2 u k } represents the k decomposed mode components; ω k = { ω 1 , ω 2 ω k } denotes the central frequencies of each mode; δ ( t ) is the Dirac distribution; the operator e j w k t can shift the central frequency of the mode component to the baseband; and u k = f ( t ) indicates the sum of all mode components is equal to the original signal.
Introducing the quadratic penalty factor, α, and the Lagrange penalty operator, σ ( t ) , improves the solution of the aforementioned variational constraints optimally. The expanded Lagrange expression is given by the following equation:
L ( { u k ( t ) } , { ω k ( t ) } , σ ( t ) ) = α k = 1 k t [ ( δ ( t ) + j π t ) u k ( t ) ] e j w k t 2 2 + f ( t ) k = 1 k u k ( t ) 2 2 + [ σ ( t ) , f ( t ) k = 1 k u k ( t ) ]
Applying the alternating direction method of multipliers to address the variational problem described above involves iteratively updating u k n + 1 ( t ) , ω k n + 1 t , and σ k n + 1 ( t ) . This approach ultimately yields the saddle point of the unconstrained model, following these steps:
Step 1: Initialize u ~ k 1 , ω ~ k 1 , σ ~ k 1 , and n.
Step2: Increment n = n + 1 and k = k + 1 iteratively until the specified number of layers is reached, terminating the loop. Update formulas for mode components and central frequencies as follows:
u ~ k ( n + 1 ) = f ~ ( ω ) i k u ~ i n + 1 ( ω ) i k u ~ i n ( ω ) + σ ~ n ( ω ) 2 ( 1 + 2 α ( ω ω k n ) 2 )
where u ~ k n + 1 , f ~ ω , and σ ~ n ω , respectively, denote the Fourier transforms of u k n ( t ) , f t   , and σ k n ( t ) ; and u ~ i n + 1 ω represents the result of f ~ ω i k u ~ i t ω passed through a wiener filter. The algorithm re-estimates central frequencies based on the power spectra of each component.
ω k + 1 n = 0 ω u ~ k n + 1 ( ω ) 2 0 u ~ k n + 1 ( ω ) 2
Step 3: Update σ with a precision threshold ξ ; terminate the iteration when k u ~ k n + 1 u ~ k n 2 2 u ~ k n 2 2 ξ .
σ ~ n + 1 σ ~ n + 1 ( ω ) + τ ( f ~ ( ω ) u ~ k n + 1 ( ω ) )

2.4. Gated Recurrent Unit

The gated recurrent unit (GRU) is a deep learning model for sequence modeling and prediction originally proposed by [39]. LSTM addresses the gradient vanishing problem of RNNs by employing gated units designed to capture long-term dependencies. However, LSTM incorporates input gates, forget gates, and output gates, rendering its structure more complex. The GRU further simplifies the internal model structure based on LSTM, including the update gate and the reset gate. The GRU utilizes an update gate to regulate how much information from the previous unit should be retained, while the reset gate controls whether the stored information from the preceding unit should be combined with the input to the next unit. The design of GRU aims to simplify the gating mechanism while enhancing computational efficiency without compromising performance. The specific structure is illustrated in Figure 1.
The GRU neuron comprises a reset gate r t , an update gate z t , a candidate hidden state h ~ t , and the final hidden state h t , where t denotes the current time step. Detailed expressions are given by Equations (25)–(28), in which W denotes the weight matrix, b represents the bias term, σ denotes the sigmoid activation function, tanh denotes the hyperbolic tangent activation function, and represents the Hadamard product.
r t = σ ( W r [ h t 1 , x t ] + b r )
z t = σ ( W z [ h t 1 , x t ] + b z )
h ~ t = tanh ( W h [ r t h t 1 , x t ] + b h )
h t = ( 1 z t ) h t 1 + z t h ~ t
The reset gate r t , determined by the sigmoid function, regulates how much information from the previous unit’s hidden state h t 1 flows into the candidate hidden state h ~ t . It controls how the hidden state from the previous timestep influences the candidate’s hidden state at the current timestep. The update gate z t , determined by the sigmoid function, regulates how much information from the input variable x t and the previous time step’s hidden state h t 1 output flows into the current hidden state h t . The h ~ t is the candidate hidden state updated based on r t , incorporating information from the reset gate and the current input. Based on r t and the previous timestep’s hidden state h t 1 , computed using the Hadamard product, the resulting vector is added to the current timestep input x t . This combined result undergoes a tanh activation function to yield h ~ t . The final hidden state h t , determined by the update gate, combines useful information from the current candidate state h ~ t and the previous state h t 1 . It is a weighted combination resulting from the update gate. It is worth mentioning that the activation function can also be other functions such as rectified linear unit (ReLU) or scaled exponential linear unit (SELU). The sigmoid and tanh functions are generally used.

3. Main Structure of the Proposed Hybrid Model

The complexity of historical NGL data requires a deeper understanding of its features of the historical series and their impact on forecasting to achieve more accurate forecasts. This paper innovatively proposes a load-forecasting model based on data decomposition and ensemble deep learning by incorporating complex features. This approach establishes a comprehensive and efficient forecasting framework that not only enhances prediction accuracy but also strengthens the model’s adaptability and robustness to future variations. The process of the hybrid forecasting framework is illustrated in Figure 2, providing a clear depiction of the entire workflow from data collection to final prediction. This ensures systematic and methodological rigor in NGL forecasting.
The proposed framework includes five core components: ① Data collection involves obtaining raw data from multiple sources to ensure comprehensiveness and accuracy. ② This study constructs complexity features that capture NGL’s self-similarity, long-term memory, randomness, and chaos to augment the predictive model’s performance. ③ The contribution of meteorological and complexity features to the forecasting model is assessed using ensemble learning with XGBoost technology. ④ Since direct modeling on highly complex data often leads to suboptimal results, the VMD method is used to decompose historical load data into several intrinsic mode functions (IMFs). ⑤ Finally, the selected key features are combined with the decomposed subcomponents as inputs to the GRU forecasting model to collectively perform the multi-step prediction task.
To meet the requirements of real-world applications, this study conducted multi-step forecasting of NGL by utilizing a multi-feature dataset and comparative modeling experiments. The results provide decision-making support for developing sustainable gas supply scheduling strategies.

4. Experiment

4.1. Dataset Description and Preprocessing

The raw data used in this study primarily originates from a branch of a natural gas station in Hanzhong, China, spanning historical daily load data from 2016 to 2022. Hanzhong City is situated in southern China. Characterized by a mild and humid climate, it occupies a unique geographical position that links the Guanzhong Plain urban cluster and the Chengdu–Chongqing urban cluster, playing a crucial role in the regional economy. To accurately predict NGL, this study also collected meteorological data from the Hanzhong station of the National Centers for Environmental Information in the United States, located at coordinates (107.02, 33.04). The specific factors include average temperature (TEMP), average dew point (DEWP), average visibility (VISIB), average wind speed (WDSP), maximum sustained wind speed (MXSPD), maximum temperature (MAXT), minimum temperature (MINT), and precipitation (PRCP). Additionally, this study integrates date type (DT) and seasonal type (ST) factors.
Due to objective reasons causing data gaps, this study employed linear spline interpolation to fill in missing values. Distinct linear interpolation strategies were applied to intervals with different missing values to achieve smooth data completion. The processed historical daily load data, as shown in Figure 3, yielded a total of 2557 samples. The descriptive statistical analysis of all data is presented in Table 2, encompassing maximum values, minimum values, mean values, standard deviations, medians, variances, kurtosis, skewness, and coefficients of variation. The standard deviation and variance of NGL data are large, indicating a high degree of dispersion and variability among the data points. A kurtosis above 3 indicates that the data are distributed more sharply. With a skewness of 2.066, the data exhibit a rightward deviation from its mean. The use of the augmented Dickey–Fuller test further confirms the stationarity of NGL data, providing a reliable foundation for subsequent time series analysis, as depicted in Table 3. The p-value of the series is 0.096, which is greater than 0.05, indicating that the series has a unit root and is non-stationary.

4.2. Analysis of Complexity Features

This paper analyzes the complex features of historical data, including FD, HE, SE, and MLE. These features capture the self-similarity, long-term memory, randomness, and chaos features in NGL.
Appropriate sliding windows facilitate efficient computation and extraction of internal information patterns within the data. The computation of complex data features typically requires long time series data, taking into account the seasonality and periodicity of the data. The window size should exceed the average autocorrelation time of the data to fully account for its long-term correlations. Natural gas historical load data exhibit an annual cyclic pattern with similarities between consecutive years. Hence, this study sets the window size to 365 data points. The specific results of complexity feature calculations are illustrated in Figure 4.
Figure 4a shows that the sliding window calculation of the FD yields a maximum value of 1.9179 and a minimum value of 1.6633, indicating a relatively high level of FD. The results indicate that the sliding time series exhibits strong self-similarity and complex dynamic behavior, characterized by multi-scale features arising from fluctuations at various time scales. Complex models are required for this type of data to capture data characteristics and predict future changes.
Figure 4b presents the sliding-window calculation results of the HE, with a maximum value of 0.9450 and a minimum of 0.5760. An HE value approaching 0.5 is typically interpreted as indicative of a random sequence. Values greater than 0.5 indicate positive long-term correlations in the sequence, reflecting varying degrees of long-term memory in the time series. In this scenario, the data exhibits some trendiness, suggesting that current changes are likely to persist into the future for some time.
As shown in Figure 4c, the maximum sliding value of SE is 1.1703, while the minimum is 0.3383. A high SE value indicates that the data has a high degree of irregularity and randomness. Low SE values indicate data with high regularity and relative simplicity. In the natural gas daily load forecasting data analyzed in this study, a value around 0.3383 is relatively low. Values approaching or exceeding 1.0 are considered high SE. Different time scales reveal varying levels of complexity and patterns of change, indicating stochastic variability in the data’s patterns.
Figure 4d presents the results of the sliding MLE calculation. The maximum value observed is 0.0318, while the minimum is −0.0204. A positive MLE indicates that the data may exhibit chaotic properties, while a negative value suggests that the sequence tends toward stability. For NGL, positive values indicate that the system exhibits uncertainty and variability during certain periods. Although the rate of change is relatively low, it still exerts a noticeable influence on the long-term dynamics of the series. In this context, NGL may be influenced by factors such as seasonality, supply and demand dynamics, or operational activities at facilities. Negative values indicate that the NGL exhibits stability and convergence over a certain period. At this point, the system can cope with external influences and can regulate.
Additionally, this section also calculates the overall complexity characteristics of historical NGL. The result for FD is 1.753, indicating that the load exhibits similar statistical distribution properties over time scales. The HE value of 0.954 indicates that the load tends to maintain similar trends over a longer time range, suggesting a certain level of correlation between successive states. The SE value of 0.5965 indicates a significant degree of randomness and uncertainty within the time series. The MLE is 0.0118, a tiny positive value indicating that the system may have some dynamic changes and chaos. These results provide significant statistical references for the characteristics of daily NGL.

4.3. Ranking of Feature Importance

Feature importance ranking holds significant importance in predictive tasks. By analyzing and determining the extent to which each feature contributes to the predicted outcome, one can gain a deeper understanding of the critical driving factors within the data. It is important to note that for data with significant volatility, the influence of nonlinear factors needs to be paid attention to. Tree-based ensemble embedding methods are effective in conducting nonlinear correlation analysis and have demonstrated notable success in feature selection tasks [40]. These methods are capable of outputting feature importance rankings upon completion of model training [17]. Previous studies on NGL forecasting have not considered data features such as FD, HE, SE, and MLE. The results of the complexity analysis are fused into the traditional feature information in this section to constitute a new feature dataset, allowing the forecasting model to have sufficient auxiliary information in understanding the load data. The total feature set can be found in Supplementary Materials Table S1.
This section focuses on conducting importance ranking of TEMP, DEWP, VISIB, WDSP, MXSPD, MAXT, MINT, PRCP, DT, ST, FD, HE, SE, and MLE using XGBoost. The results of feature importance are presented in Figure 5. The importance of complexity features is observed to be second only to ST and TEMP. Based on the calculated feature importance results, this section proposes to select features with importance metrics exceeding 1.5% for further in-depth investigation as key features (marked in blue). Specifically, these include ST, TEMP, FD, HE, SE, MLE, MAXT, and MINT. Using XGBoost for feature importance ranking across a series of external influencing factors and data complexity characteristics ensures the effectiveness of input features, mitigates redundancy, and enhances predictive performance.

4.4. Data Decomposition

This section utilizes VMD to decompose the complex load data into several intrinsic mode functions (IMF0 to IMF9). The VMD method provides a clearer and more interpretable modal decomposition, as shown in the detailed results in Figure 6. IMF0 to IMF9 represent distinct frequency components of NGL fluctuation patterns. The decomposition results reveal distinct fluctuation characteristics of each modal component in the time series, arranged from high to low frequencies. This reflects the complexity and diversity of fluctuations in NGL data. Each modal component exhibits pronounced periodic or trend-based variations, indicating the temporal dynamics of NGL data across different time scales.

4.5. Evaluation Indicators

This study analyzes the performance of load forecasting using four widely adopted evaluation metrics [4]: mean absolute percentage error (MAPE), root mean squared error (RMSE), mean absolute error (MAE), and coefficient of determination (R2). The specific definitions are presented in Table 4.

4.6. Experimental Results and Discussions

4.6.1. Application and Comparison of Complexity Features

This study incorporates complexity analysis into the scope of feature considerations, with a detailed feature set outlined in Supplementary Materials Table S1. These features encompass not only traditional meteorological factors but also complex features such as self-similarity, long-term memory, randomness, and chaos features.
As indicated in Section 4.2, key features identified through XGBoost for feature importance ranking include ST, TEMP, FD, HE, SE, MLE, MAXT, and MINT. This analysis considers four complexity features, each offering valuable insights into the underlying dynamics of the time series. Furthermore, they offer theoretical support for forecasting and managing future load demands, bearing significant relevance in predicting future trends.
To validate the feature screening effect of the XGBoost model, this study designed multiple distinct feature datasets as inputs for detailed comparative analysis. As shown in Table 5, these include the unfiltered feature set, the set comprising only complexity analysis features, the set containing only meteorological factors, and the crucial feature set identified through XGBoost selection.
The combined predictive performance of hybrid models integrating various feature sets with the VMD-GRU model is illustrated in Figure 7a. The figure demonstrates that the predictive trends of the All-VMD-GRU, Complexity-VMD-GRU, Meteorological-VMD-GRU, and XGBoost-VMD-GRU hybrid models closely resemble the actual values, indicating that VMD-GRU effectively learns from historical load data. The XGBoost-VMD-GRU model, which considers complexity features, exhibits the highest degree of overlap with the actual values, indicating its superior predictive capability. The error results shown in Figure 7b indicate that the red line (XGBoost-VMD-GRU) fluctuates slightly around zero, suggesting relatively low prediction errors.
Furthermore, prediction performance metrics were calculated to assess the model’s accuracy. As illustrated in Figure 7c–f, the proposed model outperforms the comparative models in terms of RMSE, MAE, MAPE, and R2. Evaluating predictive performance metrics can reflect the prediction accuracy of the models. As shown at the bottom of Figure 7, the proposed models demonstrate the highest predictive accuracy. Additionally, Table 6 illustrates significant differences in predictive performance among different feature sets. Overall, the XGBoost-VMD-GRU model that incorporates complexity features exhibited optimal predictive performance with an RMSE of 447.914292 Nm3, MAE of 321.934698 Nm3, MAPE of 4.200876%, and R2 of 0.992201. In contrast, the All-VMD-GRU model shows the lowest level of predictive accuracy. Complexity-VMD-GRU and Meteorological-VMD-GRU produce comparable results; however, both remain less accurate than the XGBoost-VMD-GRU model, which demonstrates notably higher performance. Experimental results demonstrate that integrating complexity features into the feature set and efficiently selecting the relevant data through XGBoost significantly improves prediction accuracy during the forecasting process.

4.6.2. Comparison of Different Models

To further prove the validity of the proposed hybrid model, multiple ablation experiments are conducted in this section for a comprehensive comparison. Specifically, this section employs extreme learning machine (ELM), LSTM, and GRU as foundational prediction models, with CEEMDAN and VMD utilized as signal decomposition methods. For the prediction of daily NGL, three sets of experiments were designed. Group one includes ELM, LSTM, and GRU. Group two consists of CEEMDAN-ELM, CEEMDAN-LSTM, and CEEMDAN-GRU. Group three comprises VMD-ELM, VMD-LSTM, and VMD-GRU. Note that this section uses feature data filtered by XGBoost and historical loadings for validation. The forecasting results in Figure 8 illustrate the outcomes of the three experimental groups. The three plots in Figure 8a–c clearly illustrate the predictive trends. In Group Three, where VMD is used for decomposition and GRU for forecasting, the predicted curve shows strong agreement with the actual values.
The detailed prediction evaluation results are presented in Table 7, illustrating that the decomposed prediction models outperform single models across multiple metrics. The adoption of decomposition methods significantly enhances prediction accuracy, with the decomposition effectiveness of VMD notably superior to that of CEEMDAN. VMD effectively extracts periodic and trend components from the data, providing more accurate input information for the model. For instance, the GRU model enhanced by VMD demonstrates a reduction of 17.8570% in MAPE compared to the standalone GRU model. Similarly, RMSE and MAE metrics also exhibit reductions. Notably, although the R2 value of the single model did not exceed 90%, all decomposition-based prediction models outperformed this benchmark. In particular, the VMD-GRU model achieved a remarkably high R2, exceeding 99%. Thus, the decomposed prediction model, particularly VMD-GRU, demonstrates significant competitiveness and superiority across multiple evaluation metrics. The detailed experimental results can be found in Supplementary Materials Table S2. Additionally, Section S2 of the Supplementary Materials presents a comparative discussion between the proposed model and three recent time series forecasting models.

4.6.3. Comparison of Different Decomposition Approaches

The intrinsic mode number of VMD in this study is determined with reference to the adaptive decomposition results of CEEMDAN, and through multiple experiments, an optimal number of 10 layers has been identified. To validate the prediction effectiveness of the VMD_IMF10 method, this section contrasts various decomposition methods and IMF quantities. The specific decomposition methods include no decomposition, CEEMDAN, VMD_IMF4, VMD_IMF6, VMD_IMF8, VMD_IMF10, VMD_IMF12, and VMD_IMF14. Based on the key features selected by XGBoost discussed in Section 4.6.1, experiments were conducted to compare prediction and error results obtained from different combinations of decomposition methods with GRU, as illustrated in Figure 9. Figure 9a shows that the prediction results obtained using VMD_IMF10 are closest to the actual values. Figure 9b illustrates that the error lines are distributed around zero, indicating that decomposition with VMD_IMF10 achieves the best performance.
Figure 10 presents the prediction performance results of this experiment. Subplots (a), (b), and (c) compare the RMSE, MAE, and MAPE metrics under different decomposition methods combined with XGBoost and GRU. The results indicate that VMD_IMF10 achieves the lowest values across all three metrics. In addition, Figure 10d–k present scatter plots of predicted versus actual values. Among them, the results from the VMD_IMF10 decomposition show the best fit, as indicated by the R2 values closest to 1. Therefore, decomposing the sequence into ten IMFs using the VMD method is identified as the optimal approach. The corresponding predictive evaluation results are shown in Table 8. The results indicate that the optimal VMD_IMF10 achieves an RMSE of 447.914291, an MAE of 321.934698, a MAPE of 4.2008763%, and an R2 value of 0.992201.

4.6.4. Discussion of Different Prediction Steps

The selection of prediction time steps directly influences the forecast accuracy and practical utility of the model in predicting future load variations. To evaluate the multi-step forecasting performance of the proposed model, this section configures the prediction input for 12 steps with output steps of 6. The code execution records predictions at intervals of one, three, and six steps, as illustrated in Figure 11. The experiments indicate that the one-step forecasting results closely match the actual data. According to the data in Table 9, the R2 for one-step prediction is 0.992201, for three-step prediction is 0.986031, and for six-step prediction is 0.967999. Overall, the results are above 0.95, indicating that the proposed target model predicts well.
In this section, in addition to conducting multi-step prediction experiments on the proposed model, we also perform multi-step natural gas daily load forecasting experiments using different feature datasets proposed in this paper (All, Complexity analysis only, Meteorological factor only, and XGBoost screened) and multiple comparative models (ELM, LSTM, GRU, CEEMDAN-ELM, CEEMDAN-LSTM, CEEMDAN-GRU, VMD-ELM, VMD-LSTM, and VMD-GRU). Supplementary Materials Table S3 presents the performance evaluation results of RMSE, MAE, MAPE, and R2 for all experiments at one-step, three-step, and six-step predictions. The experimental results indicate significant variations in performance across different models, datasets, and prediction horizons.
To enhance visual clarity, the predictive performance of all experiments is summarized in Figure 12. This figure consists of 16 subplots arranged in a 4 × 4 grid. Each column represents one of the four evaluation metrics: RMSE, MAE, MAPE, and R2. Each row corresponds to a different feature selection method used to construct the input dataset for the forecasting tasks. Within each subplot, the performance of the nine models considered in this study is presented under one-step, three-step, and six-step prediction scenarios. Generally, multi-step forecasting requires consideration of increased uncertainty in future periods compared to one-step forecasting, thereby being more susceptible to cumulative error and presenting greater forecasting challenges. Therefore, in terms of overall predictive performance, one-step > three-step > six-step. In comparing the predictive performance across multiple models, decomposition prediction models outperform single prediction models, with the VMD-GRU model showing particularly superior performance. For the four datasets, the complexity features and meteorological features screened by XGBoost combined with the load data had the best prediction performance.
In practical applications, one-step forecasting provides immediate and precise load prediction results, which are crucial for natural gas scheduling and market transactions. However, for decisions such as long-term resource planning and gas supply arrangements, three-step and six-step forecasts become more critical. These forecasts capture trends over longer time spans, aiding in the formulation of more sustainable strategies to meet evolving natural gas demands and market conditions.

4.6.5. Robustness Analysis

To explore the robustness of XGBoost-VMD-GRU in real-world applications, this section conducts robustness analysis by introducing various noise distributions to the original test set. By comparing the performance of the proposed models under various noise conditions, one can assess their robustness against anomalous data and thereby enhance the reliability of these models in practical applications. The high robustness of the model implies its ability to maintain stable performance in the presence of disturbances, thereby enhancing the reliability of its applications [41].
In this experiment, Gaussian distribution N ( 0 , σ 2 ) , Poisson distribution P ( λ ) , uniform distribution U( a , b ) , and exponential distribution E X P ( λ ) were introduced as noise into the data to simulate various disturbances that may occur in real-world scenarios. Meanwhile, different parameters were tuned for each type of noise to evaluate the performance of the model under different levels of interference, and the specific results are shown in Table 10, Table 11, Table 12 and Table 13. Experimental results demonstrate that, despite the introduction of noise into the data, the predictive performance of the XGBoost-VMD-GRU model that incorporates complexity features experiences only a slight decline. The overall R2 remains above 0.986, indicating its exceptional predictive capability. The most significant performance degradation was observed under Gaussian noise with a variance of 2. Even in this case, the R2 value decreased by only 0.005996, indicating strong robustness of the proposed model.

4.6.6. Statistical Tests

In this section, the Diebold–Mariano test was employed to further validate the scientific rigor and reliability of the proposed model. The DM test is fundamentally a t-test [42]. It is used to determine whether there exists a statistically significant difference in prediction errors between two forecasting models. In contrast to traditional error evaluation metrics, it provides a rigorous statistical comparison based on empirical errors, offering a more comprehensive assessment of the relative strengths and weaknesses of different models in prediction. The optimal dataset filtered by XGBoost was used for the experiments in this section (see Section 4.6.1). The predictive effectiveness of VMD-GRU, accounting for data complexity, was compared with eight models (ELM, LSTM, GRU, CEEMDAN-ELM, CEEMDAN-LSTM, CEEMDAN-GRU, VMD-ELM, and VMD-LSTM) using the DM test. Detailed results are presented in Table 14. In all cases, the null hypothesis assumed that the two models have identical predictive performance. However, all p-values were less than 0.01, indicating rejection of the null hypothesis at the 1% significance level. This suggests significant differences in predictive performance between VMD-GRU and the eight benchmark models. Furthermore, all DM values were negative, indicating that VMD-GRU exhibited significantly lower prediction errors compared to the other models. The extremely small p-values further confirm that the superiority of VMD-GRU observed in the experiments is not incidental but attributable to significant differences in its design and performance. Therefore, it can be concluded that VMD-GRU outperforms the compared models significantly in terms of performance.

5. Conclusions and Future Directions

Historical loadings of natural gas are characterized by complexity. Achieving more accurate forecasts requires a deep understanding of the complex characteristics of historical sequences and their impact on predictability. However, existing research rarely incorporates the complex features of historical load data into natural gas forecasting frameworks. Therefore, this paper proposes a load-forecasting model based on data decomposition and ensemble deep learning that combines data complexity features.
This study employs FD, HE, SE, and MLE to analyze the self-similarity, long-term memory, randomness, and chaotic features of historical NGL data, addressing the data complexity features and exploring the complexities of time series. Next, employing XGBoost, we conducted an importance ranking of four complexity features and ten collected meteorological features, identifying eight key features. Subsequently, VMD decomposed the load data into ten intrinsic mode functions (IMFs). These modes were integrated with selected key features as input vectors for the GRU network to perform multi-step forecasting. Furthermore, this study discusses the predictive performance of various feature datasets and different combinatorial models. The results further validate the efficiency of the hybrid XGBoost-VMD-GRU model considering data complexity features in multi-step NGL forecasting. To better meet practical application needs, this study conducts multi-step forecasting experiments for daily NGL using various feature datasets and comparative models. This approach aids in formulating sustainable strategies for natural gas resource planning and supply scheduling. Finally, this study performs robustness analysis and the DM test on the proposed methods, comprehensively demonstrating their superiority across various metrics.
The hybrid predictive model that accounts for data complexity features can be applied in energy load forecasting. This study validates the reliability of the method using data from Hanzhong, Shaanxi. The research findings will aid in the optimization of the current energy forecasting framework and provide valuable insights for energy demand management and market operations. Additionally, this research can serve as a theoretical reference for other fields, particularly those involving the analysis and forecasting of complex time series data, such as financial markets and environmental science.
Despite the strong performance of the proposed model, several limitations remain to be addressed. (1) Potential overfitting risk: although the model achieves high accuracy on the current dataset, it may overfit when trained on limited data. Future work will explore regularization techniques and cross-validation strategies to reduce this risk. (2) Limited regional transferability: the model has not been extensively tested in regions with different natural gas consumption patterns. Its generalizability across regions needs further validation. Subsequent research will conduct transferability assessments incorporating multi-source data to expand the model’s application scope. (3) Sensitivity to noise: this study conducted an initial robustness analysis by introducing Gaussian, Poisson, uniform, and exponential noise, demonstrating the model’s stability under certain perturbations. However, the model may still exhibit performance fluctuations when dealing with unstructured or extremely anomalous data. Subsequent work will consider integrating robust optimization methods or anomaly detection mechanisms to improve adaptability under complex data environments. These limitations suggest clear directions for future research and can support more reliable and scalable applications of the model in real-world energy systems.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/e27070671/s1. References [43,44,45] are cited in the Supplementary Materials.

Author Contributions

Conceptualization, H.Z.; methodology, N.T., H.Z., and M.R.; validation, M.R.; investigation, X.Z.; resources, B.S.; data curation, N.T.; writing—original draft preparation, N.T.; writing—review and editing, B.S. and H.Z.; visualization, W.Z. and X.Z.; supervision, S.W.; funding acquisition, B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Chongqing Municipal Education Commission Humanities and Social Sciences Research Base Project (No. 25SKJD068), the Youth Project for Humanities and Social Sciences Research (No. 25SKGH057), the National Natural Science Foundation of China (No. 62072363), and the Yulin City Science and Technology Plan Project (No. CXY-2020-063). This research was also supported by the Institute of Information Technology and Management, Xi’an University of Architecture and Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors thank the editor and the anonymous reviewers for their valuable comments.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Nomenclature

Abbreviations
CEEMDANComplete Ensemble Empirical Mode Decomposition with Adaptive Noise
DTDate Type
DEWPAverage Dew Point
ELMExtreme Learning Machine
FDFractal Dimension
GAGenetic Algorithm
GRUGate Recurrent Unit
HEHurst Exponent
IMFsIntrinsic Mode Functions
LSTMLong Short-Term Memory
MAEMean Absolute Error
MAPEMean Absolute Percentage Error
MAXTMaximum Temperature
MINTMinimum Temperature
MLMachine Learning
MLEMaximum Lyapunov Exponent
MXSPDMaximum Sustained Wind Speed
NGLNatural Gas Load
PRCPPrecipitation
ReLURectified Linear Unit
RMSERoot Mean Squared Error
R2Coefficient of Determination
R/S analysisRescaled Range analysis
SESample Entropy
SELUScaled Exponential Linear Unit
STSeasonal Type
SVRSupport Vector Regression
TEMPAverage Temperature
VISIBAverage Visibility
VMDVariational Mode Decomposition
WDSPAverage Wind Speed
XGBoosteXtreme Gradient Boosting
Variables
X k m A subsequence set partitioned from the starting point m with a delay interval k.
L m k Curve length of the subsequence starting at an initial time point m with a time interval k.
L k The mean curve length at identical time intervals k.
X a ¯ The mean value of subsequence a.
Y a b For subsequence a, the cumulative deviations are calculated progressively for the initial b data points.
R a Range of subsequence a.
S a Standard deviation of subsequence a.
u k k-th intrinsic mode function.
ω k Central frequency of the k-th mode.
r t Reset gate of the GRU.
z t Update gate of the GRU.
h ~ t Candidate hidden state of the GRU.
h t Final hidden state of the GRU.

References

  1. Energy Institute. Energy Institute Statistical Review of World Energy 2024. London, June 2024. Available online: https://www.energyinst.org/statistical-review (accessed on 10 July 2024).
  2. Levi, M. Climate consequences of natural gas as a bridge fuel. Clim. Change 2013, 118, 609–623. [Google Scholar] [CrossRef]
  3. Zeng, H.; Shao, B.; Dai, H.; Tian, N.; Zhao, W. Incentive-based demand response strategies for natural gas considering carbon emissions and load volatility. Appl. Energy 2023, 348, 121541. [Google Scholar] [CrossRef]
  4. Tian, N.; Shao, B.; Bian, G.; Zeng, H.; Li, X.; Zhao, W. Application of forecasting strategies and techniques to natural gas consumption: A comprehensive review and comparative study. Eng. Appl. Artif. Intel. 2024, 129, 107644. [Google Scholar] [CrossRef]
  5. Hussain, A.; Memon, J.A.; Murshed, M.; Alam, M.S.; Mehmood, U.; Alam, M.N.; Rahman, M.; Hayat, U. A time series forecasting analysis of overall and sector-based natural gas demand: A developing South Asian economy case. Environ. Sci. Pollut. Res. Int. 2022, 29, 71676–71687. [Google Scholar] [CrossRef] [PubMed]
  6. Cihan, P. Impact of the COVID-19 lockdowns on electricity and natural gas consumption in the different industrial zones and forecasting consumption amounts: Turkey case study. Int. J. Electr. Power. Energy. Syst. 2022, 134, 107369. [Google Scholar] [CrossRef]
  7. Marziali, A.; Fabbiani, E.; Nicolao, G.D. Ensembling methods for countrywide short-term forecasting of gas demand. Int. J. Oil. Gas. Coal. Technol. 2021, 26, 184–201. [Google Scholar] [CrossRef]
  8. Wei, N.; Li, C.; Li, C.; Xie, H.; Du, Z.; Zhang, Q.; Zeng, F. Short-term forecasting of natural gas consumption using factor selection algorithm and optimized support vector regression. J. Energy Resour. Technol. 2019, 141, 032701. [Google Scholar] [CrossRef]
  9. Rehman, A.; Zhu, J.J.; Segovia, J.; Anderson, P.R. Assessment of deep learning and classical statistical methods on forecasting hourly natural gas demand at multiple sites in Spain. Energy 2022, 244, 122562. [Google Scholar] [CrossRef]
  10. Wang, Q.; Suo, R.; Han, Q. A study on natural gas consumption forecasting in China using the LMDI-PSO-LSTM model: Factor decomposition and scenario analysis. Energy 2024, 292, 130435. [Google Scholar] [CrossRef]
  11. Ding, J.; Zhao, Y.; Jin, J. Forecasting natural gas consumption with multiple seasonal patterns. Appl. Energy 2023, 337, 120911. [Google Scholar] [CrossRef]
  12. Peng, S.; Chen, R.; Yu, B.; Xiang, M.; Lin, X.; Liu, E. Daily natural gas load forecasting based on the combination of long short term memory, local mean decomposition, and wavelet threshold denoising algorithm. J. Nat. Gas. Sci. Eng. 2021, 95, 104175. [Google Scholar] [CrossRef]
  13. Jiang, S.; Zhao, X.T.; Li, N. Predicting the monthly consumption and production of natural gas in the USA by using a new hybrid forecasting model based on two-layer decomposition. Environ. Sci. Pollut. Res. Int. 2023, 30, 40799–40824. [Google Scholar] [CrossRef]
  14. Wu, B.; Wang, L. Two-stage decomposition and temporal fusion transformers for interpretable wind speed forecasting. Energy 2024, 288, 129728. [Google Scholar] [CrossRef]
  15. Zhao, W.; Shao, B.; Tian, N.; Zhang, W.; Zhao, X.; Wu, S. Early warning study of field station process safety based on VMD-CNN-LSTM-self-attention for natural gas load prediction. Sci. Rep. 2025, 15, 6360. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, Z.; Liu, X. Prediction of long-term gas load based on particle swarm optimization and gray neural network model. Adv. Mech. Eng. 2017, 9, 1687814017711081. [Google Scholar] [CrossRef]
  17. Li, F.; Zheng, H.; Li, X.; Yang, F. Day-ahead city natural gas load forecasting based on decomposition-fusion technique and diversified ensemble learning model. Appl. Energy 2021, 303, 117623. [Google Scholar] [CrossRef]
  18. Hou, H.; Liu, C.; Wang, Q.; Wu, X.; Tang, J.; Shi, Y.; Xie, C. Review of load forecasting based on artificial intelligence methodologies, models, and challenges. Electr. Pow. Syst. Res. 2022, 210, 108067. [Google Scholar] [CrossRef]
  19. Rao, C.; Zhang, Y.; Wen, J.; Xiao, X.; Goh, M. Energy demand forecasting in China: A support vector regression-compositional data second exponential smoothing model. Energy 2023, 263, 125955. [Google Scholar] [CrossRef]
  20. Zeng, H.; Shao, B.; Dai, H.; Yan, Y.; Tian, N. Prediction of fluctuation loads based on GARCH family-CatBoost-CNNLSTM. Energy 2023, 263, 126125. [Google Scholar] [CrossRef]
  21. Raubitzek, S.; Neubauer, T. An exploratory study on the complexity and machine learning predictability of stock market data. Entropy 2022, 24, 332. [Google Scholar] [CrossRef]
  22. Karaca, Y.; Zhang, Y.D.; Muhammad, K. Characterizing complexity and self-similarity based on fractal and entropy analyses for stock market forecast modelling. Expert Syst. Appl. 2020, 144, 113098. [Google Scholar] [CrossRef]
  23. Kim, S.; Ku, S.; Chang, W.; Song, J.W. Predicting the direction of US stock prices using effective transfer entropy and machine learning techniques. IEEE Access 2020, 8, 111660–111682. [Google Scholar] [CrossRef]
  24. Bai, S.; Huang, X.; Luo, M.; Su, J. Deep hybrid models for daily natural gas consumption forecasting and complexity measuring. Energy Sci. Eng. 2023, 11, 654–674. [Google Scholar] [CrossRef]
  25. Wei, N.; Yin, L.; Li, C.; Liu, J.; Li, C.; Huang, Y.; Zeng, F. Data complexity of daily natural gas consumption: Measurement and impact on forecasting performance. Energy 2022, 238, 122090. [Google Scholar] [CrossRef]
  26. Raubitzek, S.; Neubauer, T. A fractal interpolation approach to improve neural network predictions for difficult time series data. Expert Syst. Appl. 2021, 169, 114474. [Google Scholar] [CrossRef]
  27. Higuchi, T. Approach to an irregular time series on the basis of the fractal theory. Phys. D Nonlinear Phenom. 1988, 31, 277–283. [Google Scholar] [CrossRef]
  28. Hurst, H.E. Long-term storage capacity of reservoirs. Trans. Am. Soc. Civil. Eng. 1951, 116, 770–799. [Google Scholar] [CrossRef]
  29. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol.-Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  30. Lyapunov, A.M. The general problem of the stability of motion. Int. J. Control. 1992, 55, 531–534. [Google Scholar] [CrossRef]
  31. Eckmann, J.P.; Kamphorst, S.O.; Ruelle, D.; Ciliberto, S. Liapunov exponents from time series. Phys. Rev. A. 1986, 34, 4971. [Google Scholar] [CrossRef] [PubMed]
  32. Raubitzek, S.; Neubauer, T. Taming the chaos in neural network time series predictions. Entropy 2021, 23, 1424. [Google Scholar] [CrossRef]
  33. Wolf, A.; Swift, J.B.; Swinney, H.L.; Vastano, J.A. Determining Lyapunov exponents from a time series. Phys. D Nonlinear Phenom. 1985, 16, 285–317. [Google Scholar] [CrossRef]
  34. Rosenstein, M.T.; Collins, J.J.; De Luca, C.J. A practical method for calculating largest Lyapunov exponents from small data sets. Phys. D Nonlinear Phenom. 1993, 65, 117–134. [Google Scholar] [CrossRef]
  35. Zengrong, L.; Liqun, C.; Ling, Y. On properties of hyperchaos: Case study. Acta. Mech. Sin. 1999, 15, 366–370. [Google Scholar] [CrossRef]
  36. Henriques, T.; Ribeiro, M.; Teixeira, A.; Castro, L.; Antunes, L.; Costa-Santos, C. Nonlinear methods most applied to heart-rate time series: A review. Entropy 2020, 22, 309. [Google Scholar] [CrossRef]
  37. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  38. Liu, J.; Huang, X.; Li, Q.; Chen, Z.; Liu, G.; Tai, Y. Hourly stepwise forecasting for solar irradiance using integrated hybrid models CNN-LSTM-MLP combined with error correction and VMD. Energ. Convers. Manag. 2023, 280, 116804. [Google Scholar] [CrossRef]
  39. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  40. Tuv, E.; Borisov, A.; Runger, G.; Torkkola, K. Feature selection with ensembles, artificial variables, and redundancy elimination. J. Mach. Learn. Res. 2009, 10, 1341–1366. [Google Scholar]
  41. Dai, H.; Huang, G.; Zeng, H.; Zhou, F. PM2.5 volatility prediction by XGBoost-MLP based on GARCH models. J. Clean. Prod. 2022, 356, 131898. [Google Scholar] [CrossRef]
  42. Zhao, Z.; Yun, S.; Jia, L.; Guo, J.; Meng, Y.; He, N.; Li, X.; Shi, J.; Yang, L. Hybrid VMD-CNN-GRU-based model for short-term forecasting of wind power considering spatio-temporal features. Eng. Appl. Artif. Intel. 2023, 121, 105982. [Google Scholar] [CrossRef]
  43. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar] [CrossRef]
  44. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  45. Wu, H.; Hu, T.; Liu, Y.; Zhou, H.; Wang, J.; Long, M. Timesnet: Temporal 2d-variation modeling for general time series analysis. arXiv 2022, arXiv:2210.02186. [Google Scholar] [CrossRef]
Figure 1. Structure diagram of GRU.
Figure 1. Structure diagram of GRU.
Entropy 27 00671 g001
Figure 2. A load-forecasting framework based on data decomposition and ensemble deep learning for incorporating complex features.
Figure 2. A load-forecasting framework based on data decomposition and ensemble deep learning for incorporating complex features.
Entropy 27 00671 g002
Figure 3. Natural gas daily load data.
Figure 3. Natural gas daily load data.
Entropy 27 00671 g003
Figure 4. Sliding fractal dimension, Hurst exponent, sample entropy, and maximum Lyapunov exponent of NGL.
Figure 4. Sliding fractal dimension, Hurst exponent, sample entropy, and maximum Lyapunov exponent of NGL.
Entropy 27 00671 g004
Figure 5. Ranking of feature importance.
Figure 5. Ranking of feature importance.
Entropy 27 00671 g005
Figure 6. Intrinsic mode functions (IMF0~IMF9).
Figure 6. Intrinsic mode functions (IMF0~IMF9).
Entropy 27 00671 g006
Figure 7. Comparison of the forecasting performance of hybrid models combining different feature sets with the VMD-GRU model. The checkmark (√) in (cf) indicates the model with the best prediction performance.
Figure 7. Comparison of the forecasting performance of hybrid models combining different feature sets with the VMD-GRU model. The checkmark (√) in (cf) indicates the model with the best prediction performance.
Entropy 27 00671 g007
Figure 8. Results of predictions for multiple comparative experiments.
Figure 8. Results of predictions for multiple comparative experiments.
Entropy 27 00671 g008
Figure 9. Prediction and error results for different decomposition methods combined with XGBoost and GRU.
Figure 9. Prediction and error results for different decomposition methods combined with XGBoost and GRU.
Entropy 27 00671 g009
Figure 10. Evaluation results of the prediction performance of different decomposition methods. Subfigures (ac) compare the RMSE, MAE, and MAPE metrics under different decomposition methods combined with XGBoost and GRU. Subfigures (dk) present scatter plots of predicted versus actual values.
Figure 10. Evaluation results of the prediction performance of different decomposition methods. Subfigures (ac) compare the RMSE, MAE, and MAPE metrics under different decomposition methods combined with XGBoost and GRU. Subfigures (dk) present scatter plots of predicted versus actual values.
Entropy 27 00671 g010
Figure 11. Diagram of multi-step prediction results of XGBoost-VMD-GRU considering complexity features.
Figure 11. Diagram of multi-step prediction results of XGBoost-VMD-GRU considering complexity features.
Entropy 27 00671 g011
Figure 12. The RMSE, MAE, MAPE, and R2 metrics for one-step, three-step, and six-step forecasts are presented for various comparative models (including ELM, LSTM, GRU, CEEMDAN-ELM, CEEMDAN-LSTM, CEEMDAN-GRU, VMD-ELM, VMD-LSTM, and VMD-GRU) across different feature datasets.
Figure 12. The RMSE, MAE, MAPE, and R2 metrics for one-step, three-step, and six-step forecasts are presented for various comparative models (including ELM, LSTM, GRU, CEEMDAN-ELM, CEEMDAN-LSTM, CEEMDAN-GRU, VMD-ELM, VMD-LSTM, and VMD-GRU) across different feature datasets.
Entropy 27 00671 g012
Table 1. Summary of methods for data complexity analysis.
Table 1. Summary of methods for data complexity analysis.
MethodsAbbreviationsDescriptionTypical Value RangePractical Implications
Fractal DimensionFDIt quantifies the self-similarity of a sequence to facilitate a deeper comprehension of the dynamic characteristics and behavioral patterns inherent in the data.(1, 2)A larger value indicates a rougher and more complex sequence.
Hurst ExponentHEIt quantifies long-term memory and persistence characteristics within the sequence.(0, 1)HE < 0.5 indicates an anti-persistent sequence; HE > 0.5 signifies a persistent sequence; and HE = 0.5 denotes a random walk sequence.
Sample EntropySEIt quantifies the degrees of randomness and uncertainty inherent in the data.[0, +∞)A larger value indicates a higher degree of disorder and greater unpredictability within the sequence.
Maximum Lyapunov ExponentMLEIt captures the system’s sensitivity to initial condition perturbations, characterizing chaotic behavior.(−∞, +∞)A positive value of the data suggests the potential presence of chaotic characteristics, whereas a negative value indicates that the sequence tends toward stability.
Table 2. Descriptive statistical analysis of the dataset.
Table 2. Descriptive statistical analysis of the dataset.
NO.DataSample SizeMaximumMinimumMeanStandard DeviationMedianVarianceKurtosisSkewness
1TEMP255793.827.160.95615.54561.1241.647−1.158−0.031
2DEWP255780.4−2.151.82115.24253.3232.328−0.883−0.272
3VISIB255718.60.911.1064.95611.324.564−1.129−0.171
4WDSP25578.70.22.1930.7812.10.617.8631.877
5MXSPD255715.51.43.7051.4833.92.1993.7621.22
6MAXT2557104.728.469.13516.82169.3282.952−1.069−0.057
7MINT255783.817.253.6815.08854.1227.639−1.168−0.084
8PRCP25574.1100.110.31800.10137.7825.312
9DT2557200.3910.62700.3930.6771.36
10ST25570.80.20.4980.2230.40.05−1.3580.011
11DNGL255723,510.5215486800.3243349.295561411,217,775.414.5862.066
Table 3. Stationarity test.
Table 3. Stationarity test.
tpCritical Value
1%5%10%
DNGL−2.5870.096−3.433−2.863−2.567
Table 4. Evaluation metrics.
Table 4. Evaluation metrics.
MetricsFormulaRules
1MAPE M A P E = 1 N t = 1 N | A t F t A t | × 100 % The smaller, the more accurate.
2RMSE R M S E = 1 N t = 1 N ( A t F t ) 2 The smaller, the more accurate.
3MAE M A E = 1 N t = 1 N | A t F t | The smaller, the more accurate.
4 R 2 R 2 = 1 t = 1 N ( F t A t ) 2 t = 1 N ( A t ¯ A t ) 2 R 2   is   between   0   and   1 ;   the   closer   R 2 is to 1, the more accurate it is.
Notes: N is the length of the data; A t is the actual value; F t is the predicted value; A t ¯ is the average of the actual value; and F t ¯ is the average of the predicted value.
Table 5. Different ways of filtering features and details of feature sets.
Table 5. Different ways of filtering features and details of feature sets.
Feature SetDescriptionDetail
AllA complete set of unfiltered features.TEMP, DEWP, VISIB, WDSP, MXSPD, MAX, MIN, PRCP, DT, ST, HE, SE, FD, MLE
Complexity analysis onlyFeature sets that contain only complexity analysis.HE, SE, FD, MLE
Meteorological factor onlyFeature sets containing only weather, date type, and season type.TEMP, DEWP, VISIB, WDSP, MXSPD, MAX, MIN, PRCP, DT, ST
XGBoost screenedKey feature sets selected by XGBoost.TEMP, MAX, MIN, ST, HE, SE, FD, MLE
Table 6. Results of forecast evaluation of different feature sets combined with VMD-GRU.
Table 6. Results of forecast evaluation of different feature sets combined with VMD-GRU.
ModelRMSEMAEMAPE R 2
All-VMD-GRU674.488371512.5182937.0955660.982316
Complexity-VMD-GRU546.501682389.8777734.8244700.988391
Meteorological-VMD-GRU549.066995411.1107915.5491120.988281
XGBoost-VMD-GRU447.914292321.9346984.2008760.992201
Table 7. Results of predictive evaluation of multiple comparative experiments.
Table 7. Results of predictive evaluation of multiple comparative experiments.
Feature SetMetricsSingle Forecasting ModelDecomposition Forecasting Model
ELMLSTMGRUCEEMDAN-ELMCEEMDAN-LSTMCEEMDAN-GRUVMD-ELMVMD-LSTMVMD-GRU
XGBoost screenedRMSE2086.3315492716.9182232300.9728691418.9539991191.7270771134.9987231005.942378577.876172447.914292
MAE1602.1323182002.2541271651.1148071088.172136861.348212814.350885717.038583398.331519321.934698
MAPE23.15172127.15589122.05792215.81108011.54272311.0443769.8085345.2026874.200876
R 2 0.8401590.7289340.8055780.9263300.9480370.9529190.9606660.9870190.992201
Table 8. Results of prediction evaluation for different decomposition methods.
Table 8. Results of prediction evaluation for different decomposition methods.
MetricsDecomposition Method
No DecompositionCEEMDANVMD_IMF4VMD_IMF6VMD_IMF8VMD_IMF10VMD_IMF12VMD_IMF14
RMSE2300.9728691134.998723738.2005945606.115283576.013733447.914291594.157103663.385375
MAE1651.114807814.350885584.798344453.206701438.779238321.934698454.483538530.699079
MAPE22.05792211.0443767.8341736.3738466.3439594.20087636.4448417.630316
R 2 0.8055780.9529190.9765910.9850630.9865690.9922010.9865130.983301
Table 9. The performance results of XGBoost-VMD-GRU with multi-step prediction considering complexity features.
Table 9. The performance results of XGBoost-VMD-GRU with multi-step prediction considering complexity features.
Step SizeRMSEMAEMAPE R 2
XGBoost-VMD-GRUOne step447.914292321.9346984.2008760.992201
Three steps599.747286414.1082635.2044390.986031
Six steps905.755273630.9435847.7879190.967999
Table 10. Results of prediction performance after adding Gaussian distribution noise.
Table 10. Results of prediction performance after adding Gaussian distribution noise.
N ( 0 , σ 2 ) RMSEMAEMAPE R 2
Normal447.914292321.9346984.2008760.992201
σ   = 0.2492.635181348.1826884.5089060.990566
σ   = 0.4516.884271372.9739684.9096150.989615
σ   = 0.6550.019130413.2439825.5379680.988241
σ   = 0.8561.713045415.6659265.9365310.987735
σ   = 1560.878278414.5571845.5337410.987772
σ   = 2595.731438430.7930725.2822460.986205
Table 11. Results of prediction performance after adding Poisson distribution noise.
Table 11. Results of prediction performance after adding Poisson distribution noise.
P ( λ ) RMSEMAEMAPE R 2
Normal447.914292321.9346984.2008760.992201
λ   = 1451.156046327.5415714.3444490.992088
λ   = 2478.850837364.3276285.1858260.991087
λ   = 4482.163025351.2509704.6935620.990964
λ   = 8506.049992367.7577445.0205580.990046
λ   = 16532.947304395.2114375.4898710.988960
Table 12. Results of prediction performance after adding uniform distribution noise.
Table 12. Results of prediction performance after adding uniform distribution noise.
U   ( a ,   b ) RMSEMAEMAPE R 2
Normal447.914292321.9346984.2008760.992201
a ,   b = ( 0.5, 0.5 ) 509.954158374.7394925.2186190.989892
a ,   b = ( 1, 1 ) 550.655201392.6489434.8929830.988214
a ,   b = ( 2, 2 ) 531.684776400.9000715.7664180.989012
a ,   b = ( 5, 5 ) 530.323694401.8322235.6787570.989068
a ,   b = ( 10, 10 ) 543.076969418.1834546.0509530.988537
Table 13. Results of prediction performance after adding exponential distribution noise.
Table 13. Results of prediction performance after adding exponential distribution noise.
E X P ( λ ) RMSEMAEMAPE R 2
Normal447.914292321.9346984.2008760.992201
λ   = 1492.356910361.2746834.8663600.990577
λ   = 2484.870815349.8393954.6702030.990862
λ   = 4489.147496374.5019245.3560990.990699
λ   = 8495.410941370.4218225.1593850.990460
λ   = 16501.701613361.8870914.7300340.990216
Table 14. Diebold–Mariano test results between VMD-GRU and other models, respectively.
Table 14. Diebold–Mariano test results between VMD-GRU and other models, respectively.
Model (Consider the Complexity of the Data)DM Valuep Value
ELM−12.970941 ***1.992764 × 10−33
LSTM−12.153916 ***5.241625 × 10−30
GRU−10.360662 ***6.044957 × 10−23
CEEMDAN-ELM−11.057308 ***1.308850 × 10−25
CEEMDAN-LSTM−9.867247 ***3.998776 × 10−21
CEEMDAN-GRU−9.127753 ***1.655615 × 10−18
VMD-ELM−8.464992 ***2.772544 × 10−16
VMD-LSTM−4.632283 ***4.603091 × 10−6
*** denotes significance levels at 1%.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tian, N.; Shao, B.; Zeng, H.; Ren, M.; Zhao, W.; Zhao, X.; Wu, S. Multi-Step Natural Gas Load Forecasting Incorporating Data Complexity Analysis with Finite Features. Entropy 2025, 27, 671. https://doi.org/10.3390/e27070671

AMA Style

Tian N, Shao B, Zeng H, Ren M, Zhao W, Zhao X, Wu S. Multi-Step Natural Gas Load Forecasting Incorporating Data Complexity Analysis with Finite Features. Entropy. 2025; 27(7):671. https://doi.org/10.3390/e27070671

Chicago/Turabian Style

Tian, Ning, Bilin Shao, Huibin Zeng, Meng Ren, Wei Zhao, Xue Zhao, and Shuqiang Wu. 2025. "Multi-Step Natural Gas Load Forecasting Incorporating Data Complexity Analysis with Finite Features" Entropy 27, no. 7: 671. https://doi.org/10.3390/e27070671

APA Style

Tian, N., Shao, B., Zeng, H., Ren, M., Zhao, W., Zhao, X., & Wu, S. (2025). Multi-Step Natural Gas Load Forecasting Incorporating Data Complexity Analysis with Finite Features. Entropy, 27(7), 671. https://doi.org/10.3390/e27070671

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop