Next Article in Journal
Research on an Adaptive Hole Layout Method for Bench Blasting Based on Voronoi Diagram
Previous Article in Journal
Usefulness of Flavonoids and Phenolic Acids in Differentiating Honeys Based on Geographical Origin: The Case of Dominican Republic and Spanish Honeys
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Daily Peak Load Prediction Method Based on XGBoost and MLR

1
Inner Mongolia Daqingshan Laboratory Co., Ltd., Hohhot 010020, China
2
College of Electrical Engineering, Zhejiang University, Hangzhou 310027, China
3
Inner Mongolia Power (Group) Co., Ltd., Hohhot 010020, China
4
Inner Mongolia Electric Power Economic and Technical Research Institute Branch, Inner Mongolia Electric Power Group Mengdian Economic and Technical Research Institute Co., Ltd., Hohhot 010020, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(20), 11180; https://doi.org/10.3390/app152011180 (registering DOI)
Submission received: 12 September 2025 / Revised: 11 October 2025 / Accepted: 14 October 2025 / Published: 18 October 2025

Abstract

During the peak load period, there is a high level of imbalance between power supply and demand, which has become a critical challenge, leading to higher operational costs for power grids. To improve the accuracy of peak load forecasting, this study introduces a novel approach based on Extreme Gradient Boosting Trees (XGBoost) and Multiple Linear Regression (MLR) for daily peak load prediction. The proposed methodology first employs an improved version of the Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN) algorithm to decompose the raw load data, subsequently reconstructing each Intrinsic Mode Function (IMF) into high-frequency and stationary components. For the high-frequency components, XGBoost serves as the base predictor within a Bagging-based ensemble structure, while the Sparrow Search Algorithm (SSA) is employed to optimize hyperparameters automatically, ensuring efficient learning and accurate representation of complex peak load fluctuations. Meanwhile, the stationary components are modeled using MLR to provide fast and reliable estimations. The proposed framework was evaluated using actual daily peak load data from Western Inner Mongolia, China. The results indicate that the proposed method successfully captures the peak characteristics of the power grid, delivering both robust and precise predictions. When compared to the baseline model, the RMSE and MAPE are reduced by 54.4% and 87.3%, respectively, underscoring its significant potential for practical applications in power system operation and planning.

1. Introduction

The issue of power supply and demand imbalance during peak load periods has become increasingly prominent, leading to higher operational costs for power grids [1,2]. Accurate peak load forecasting directly impacts capacity planning, equipment maintenance, and emergency response strategies [3,4]. Inaccurate peak load predictions can lead to grid overload or resource waste, and in severe cases, it can even trigger power outages and grid failures. Moreover, peak load forecasting plays a crucial role in calculating reserve capacities, implementing demand response measures, and determining peak and off-peak electricity pricing [5]. This issue is particularly significant during summer and winter peak demand periods, when the gap between supply and demand is substantial [6]. For instance, in September 2021, large-scale power outages occurred in Northeast China and other provinces, primarily due to power shortages that could not meet the rapidly growing electricity demand. Therefore, accurately forecasting peak loads in the coming days would help the grid develop scheduling plans, assist power companies in formulating production plans, and narrow the supply–demand gap. Furthermore, accurate peak load forecasting is vital for ensuring the safe and stable operation of grids in regions with weak grid structures and heavy loads.
Currently, peak load forecasting methods can be broadly categorized into statistical analysis, machine learning, and model-based forecasting using load characteristics analysis. Reference [7] summarizes the impact of the day of the week and month on peak loads and proposes a hybrid forecasting model using the Differencing Integrated Moving Average Regression model to predict grid loads. Reference [8] utilizes robust regression methods to forecast peak loads. Reference [9] introduces a short-term load forecasting method based on improved empirical mode decomposition and neural networks, effectively improving forecasting accuracy. Reference [10] decomposes monthly electricity consumption into non-periodic trend components and seasonal components using seasonal indices to model the time-varying load. Reference [11] analyzes the impact of regional weather variables on grid loads and uses alternating conditional expectation nonparametric simulation for peak load forecasting. Although these methods have yielded good results, they still have some limitations. Firstly, traditional statistical methods are ineffective at capturing dynamic load changes when dealing with large volumes of high-dimensional grid data. Secondly, although machine learning and load characteristic-based methods can theoretically handle complex data relationships, their practical application is often constrained by data quality. Moreover, most of the studies cited above utilize single models, which tend to become trapped in local minima during training. As a result, models that correspond to these local optima often struggle with generalization.
Over the past few years, deep neural network algorithms have been increasingly applied to power grid peak load forecasting. The error backpropagation network in References [12,13] is widely used and capable of implementing complex nonlinear mappings between inputs and outputs. Reference [14] proposes a method for short- and long-term peak load forecasting at the regional level. Reference [15] uses a gray prediction method to forecast short-term peak loads. Reference [16] applies CNN-LSTM models for short-term load forecasting, validated using data from the Bangladesh power system. Reference [17] first decomposes the raw load data into several Intrinsic Mode Functions (IMFs) using EEMD and constructs a hybrid neural network based on DBN and BILSTM to predict daily peak loads for the following month. Reference [18] constructs a deep hybrid network using CNN, GRU, and Fully Connected Layers (FCNs). However, deep learning networks have complex architectures, and the selection of hidden layers and node numbers greatly influences training outcomes. If poorly designed, the network often falls into local minima and overfitting. These characteristics limit the further development of deep learning in load forecasting.
In recent years, substantial advancements have been made in tree ensemble algorithms. Notably, in competitive environments such as the Kaggle data science competition, tree ensemble algorithms, particularly XGBoost, have demonstrated superior performance, surpassing many deep learning models. Reference [19] leverages the XGBoost algorithm to mitigate overfitting, successfully predicting the cooling and heating loads of buildings. In a similar vein, Reference [20] integrates the Bagging algorithm with XGBoost, employing the Particle Swarm Optimization (PSO) algorithm to fine-tune XGBoost’s hyperparameters. Reference [21] adopts genetic algorithms (GAs) to further optimize the hyperparameters of XGBoost. While these intelligent optimization algorithms are effective in refining deep neural networks, they often suffer from slow convergence rates and a propensity to converge to local optima. In contrast, the Sparrow Search Optimization Algorithm (SSA) draws inspiration from the collective behaviors of sparrows during predation and anti-predation processes, utilizing their cooperative and competitive dynamics to explore and identify optimal solutions. SSA excels in both efficiency and precision, particularly in parameter optimization, thus overcoming many of the shortcomings typically encountered in traditional optimization methods. Consequently, SSA has emerged as a promising optimization tool, particularly for enhancing the performance of machine learning and deep learning models [22,23].
To tackle the challenges outlined, we propose a new peak load forecasting model that combines XGBoost with multiple linear regression (MLR). A comparison between our approach and the latest research is presented in Table 1. Initially, historical load data is subjected to decomposition using the ICEEMDAN algorithm, which separates the data into high-frequency and stationary components. To predict the high-frequency components, Bagging-XGBoost is employed, and the Sparrow Search Optimization Algorithm (SSA) is incorporated to optimize the hyperparameters of XGBoost, thereby reducing the computational time required for this process. For the stationary components, MLR is utilized, taking advantage of its robust curve fitting capabilities and computational efficiency to deliver accurate predictions. Subsequently, the output from the first-layer model is fed into the second-layer prediction model, which consists of SSA-XGBoost, for nonlinear fusion and reconstruction. This final step results in the generation of the ultimate peak load prediction.

2. Overall Framework of Peak Load Forecasting Model

Due to factors such as seasonal variations, diurnal differences, and the diversity of load patterns, electrical load data typically exhibit both periodic and fluctuating characteristics [24]. This is especially true for peak load forecasting scenarios over multiple consecutive days or even an entire month. If the predictive model directly learns from the raw peak load sequences, it will be significantly affected by data disturbances during training, leading to poor model generalization and robustness [25]. Therefore, to reduce the nonlinearity in peak load sequences, simplify the prediction task, and improve prediction accuracy, this study proposes a daily peak load forecasting method based on XGBoost and MLR. The overall framework of the two-layer continuous multi-day peak load forecasting model, based on ICEEMDAN-Bagging-XGBoost-MLR, proposed in this paper is shown in Figure 1. The model analysis and algorithm flow will be described in detail below.
In this study, the original peak load sequence is subjected to decomposition using the ICEEMDAN algorithm, which separates the data into distinct high-frequency and stationary components based on the zero-crossing rate of each Intrinsic Mode Function (IMF). For the high-frequency components, which encapsulate the detailed fluctuations in peak load and exhibit considerable randomness, the SSA-Bagging-XGBoost method is utilized for prediction. This methodology effectively mitigates variance and bias throughout the training process, facilitating a more accurate tracking of the load change curve. In contrast, for the low-frequency stationary components, which primarily capture the overall load variation trends, Multiple Linear Regression (MLR) is employed for prediction. The simplicity of the MLR model, combined with its strong curve fitting capability and reduced susceptibility to overfitting, makes it well-suited to this task.
Specifically, we first evaluate the contribution of each input feature to the model using the tree gain metrics derived from the XGBoost and Random Forest (RF) algorithms. These selected features are then used to define the final input set for the prediction model. The high-frequency and stationary components produced by the decomposition process serve as the target outputs. When combined with the identified input features, this forms the dataset required by the model. Next, using this dataset, the SSA optimization algorithm is applied to search for the optimal hyperparameters for XGBoost (such as the number of trees, tree depth, learning rate, and the minimum loss for node splitting). Afterward, under the optimal hyperparameter configuration, the initial training samples are fed into the Bagging-XGBoost model to predict the high-frequency components. In contrast, MLR does not require hyperparameter tuning and can directly predict the stationary components using the validation and test sets after training.

3. Framework Modules

3.1. ICEEMDAN Algorithm Mechanism

The Empirical Mode Decomposition (EMD) method is widely used for processing non-stationary and nonlinear signals. Essentially, it smooths the signal and decomposes the complex original signal into several Intrinsic Mode Functions (IMFs), ordered by their frequencies from high to low. However, a phenomenon known as modal aliasing arises during the decomposition process of the EMD algorithm. Although the Ensemble Empirical Mode Decomposition (EEMD) algorithm addresses the issue of mode aliasing in EMD, it leaves a certain amount of white noise in the IMF components, which increases reconstruction errors. In comparison to EEMD, the Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) algorithm further reduces reconstruction errors by adding adaptive white noise. As an improvement to CEEMDAN, the Improved Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN) algorithm directly introduces Gaussian white noise ( E k w i ) during the decomposition process. It then decomposes the white noise using EMD and selects the k-th IMF component to further eliminate residual noise from the IMF components and reduce the occurrence of spurious modal components. The decomposition process is well-documented in the literature [26].

3.2. XGBoost Algorithm Mechanism

XGBoost is an ensemble learning-based tree boosting model. The idea behind it is to iteratively add different trees to the model, where each tree grows through feature splitting and fits the residuals of the predictions from all previous trees. Its learning mechanism is shown in Figure 2.
For the dataset D = x i , y i , ( D = n ), the ensemble model is expressed as follows.
y i ^ t = k = 1 t f k x i = y i ^ t 1 + f t x i
where y i ^ t is the prediction result, y i ^ t 1 is the sum of the prediction results, and f t x i is the newly added tree.
The loss function of XGBoost consists of two parts: the prediction error and the regularization term.
O b j t = i = 1 n l y i , y i ^ t 1 + f t x i + i = 1 t Ω f t
Ω f t = γ T + 1 2 λ w 2
where γ and λ are the penalty coefficients, T is the number of leaves in the tree, and w is the leaf weight. Since the complexity of the first t − 1 trees is a known constant, Equation (2) can be rewritten as:
O b j t = i = 1 n l y i , y i ^ t 1 + f t x i + Ω f t + c o n s t a n t
The loss function is expanded using a second-order Taylor series, yielding:
O b j t i = 1 n l y i , y i ^ t 1 + g i f t x i + 1 2 h i f t 2 x i + Ω f t + c o n s t a n t
y i ^ 0 = 0 y i ^ 1 = f 1 x i = y i ^ 0 + f 1 x i y i ^ 2 = f 1 x i + f 2 x i = y i ^ 1 + f 2 x i y i ^ t = k = 1 t f k x i = y i ^ t + f t x i
where g i = y i ^ t 1 l y i , y i ^ t 1 , h i = y i ^ t 1 2 l y i , y i ^ t 1 .
The incremental training method adopted by the XGBoost algorithm is that in each iteration, new trees are continuously added to fit the previous error value to reduce the objective function as much as possible, as shown in Equation (7).
O b j t i = 1 n g i f t x i + 1 2 h i f t 2 x i + Ω f t

3.3. Multiple Linear Regression Mechanism

Multiple linear regression (MLR) is a traditional mathematical statistical model with early development, simple model and fast calculation speed, which has a good fitting effect on smooth curves. MLR is very suitable for forecasting data with strong periodicity and relatively stable, and its formula is:
Y = X × β + μ
where Y is the target output matrix; X is the input feature matrix; β is the retrospective coefficient matrix; μ is a constant term. For the regression parameters, the least squares method is used to estimate them, and the regression function can be obtained. The formula is:
β ^ = ( X X ) 1 X Y

3.4. Parallel Ensemble Learning Method of Bagging Mechanism

Bagging is an ensemble learning technique designed to effectively reduce variance. As shown in Figure 3, the algorithm’s architecture follows a self-sampling method (Bootstrap) to generate multiple random datasets. For instance, given a dataset D with m samples, random sampling with replacement is applied to produce a new dataset D i containing the same number of samples as the original dataset, performed m times. This process results in duplicate samples within the new dataset due to the random sampling. After completing n iterations of this self-sampling process, n new datasets D 1 , D 2 , D n 1 , D n are obtained. Each dataset is used to train n independent base learners V 1 , V 2 , V n 1 , V n . In regression tasks, the outputs from the trained base learners are averaged to compute the final result.

3.5. Sparrow Search Optimization Algorithm Mechanism

The Sparrow Search Optimization Algorithm (SSA) is an innovative swarm intelligence optimization technique inspired by the foraging and anti-predation behaviors of sparrows. Within the SSA framework, three distinct roles are defined: the discoverer, the joiner, and the predator. In an N-dimensional search space, the position of the i-th sparrow is represented as follows:
X i t = x i , 1 t , x i , 2 t , , x i , N t , i = 1,2 , m
where t is the current iteration number; m is the population size.
The position update formula for the discoverer within the sparrow population is given by:
x i , j t + 1 = σ exp x w o r s e t x i , j t i 2 , i > n / 2 x b e s t t + 1 + 1 N j N r a n d 1 , 1 · x i , j t x b e s t t + 1 , i n / 2
where x i , j t represents the position of the sparrow, x w o r s e t denotes the worst solution on the j-th dimension of the current population’s search space, and x b e s t t + 1 corresponds to the optimal solution on the j-th dimension in the search space for the current population. When i > n/2, it indicates that the i-th joiner has not found a better solution and decides to explore other areas. Conversely, when i ≤ n/2, the joiner will monitor the discoverer, competing for the best solution, and may replace the discoverer to explore a larger search space.
Additionally, the sparrow search algorithm includes predators, whose positions are determined as follows:
x i , j t + 1 = x b e s t t + β x i , j t x b e s t t , f i f b e s t x i , j t + K x i , j t x w o r s e t f i f w o r s e + μ , f i = f b e s t
where β is a random variable following a standard normal distribution, K [ 1 , 1 ] , and μ is a small constant to prevent the denominator from becoming zero. f w o r s e and f b e s t represent the worst and optimal fitness values of the current population, respectively, while f i denotes the fitness value of the i-th sparrow.

4. Case Analysis

4.1. Research Data and Evaluation Metrics

The experimental dataset used in this study was provided by the Western Inner Mongolia Power Supply Company. For the SSA, a population size of 40 and 1000 iterations were specified. The stopping criterion for decomposition in ICEEMDAN is automatically determined through an iterative process of data decomposition. The raw data comprises load measurements from 48 sampling points per day, covering the period from January 2022 to January 2024. The dataset is divided into training and testing sets at a ratio of 7:3. Additionally, the dataset includes daily average temperature, holiday information, as well as daily maximum and minimum air temperatures, relative humidity, and rainfall data. To eliminate issues with inconsistent dimensional scales, the original relevant factors and daily peak load data are first normalized to the range [0, 1]. The time series for this region are presented in Figure 4. The related model programming is implemented through Matlab 2021a and Python 3.7 environment. The computer’s CPU is an Intel Core i5-13500H with a clock speed of 2.60 GHz, and it is equipped with 16 GB of memory.
In order to comprehensively evaluate the performance of the prediction model in this paper, this paper selects relative mean square error e R M S E , mean absolute error e M A E , and mean absolute error percentage e M A P E .
e M A E = 1 m t = 1 m x t y t
e R M S E = 1 m t = 1 m [ x t y ( t ) ] 2
e M A P E = 1 m t = 1 m x t y ( t ) x ( t ) × 100 %
where x ( t ) and y ( t ) represent the actual value and predicted value at time t , respectively, and m is the number of samples.

4.2. Data Characteristics Analysis

According to Figure 4, the maximum daily electricity consumption in this area gradually increases every year, and the daily peak load pattern is an inverted U shape. This is because in the modern power system, there are many types of electrical appliances that constitute the power load, and the proportion of loads affected by weather conditions such as air conditioners continues to increase, which increases the volatility and randomness of load changes. Figure 5 shows the spatial distribution of daily peak loads and their respective meteorological factors in this region.
As shown in Figure 5, the load patterns on Saturdays and Sundays differ significantly from those observed on weekdays (Monday to Friday). Additionally, the daily maximum electricity consumption is strongly influenced by temperature fluctuations. Among the meteorological factors, although the effect of relative humidity on maximum power consumption is less pronounced than that of temperature, there is a discernible trend: as relative humidity increases, the daily maximum power consumption also tends to rise. Although rainfall is relatively frequent in this region, it has no significant impact on the maximum daily electricity consumption. Figure 6 presents the analysis of input feature contributions.
From the feature contribution scores of XGBoost and RF shown in Figure 6, it is evident that among the meteorological factors, rainfall has minimal impact on the daily maximum load during the forecast period. As a result, the effect of rainfall is excluded from the analysis. The input features for the period from Tuesday to Friday are combined into a single feature, and the meteorological input features include daily maximum temperature, minimum temperature, average temperature, and relative humidity.
While some studies have suggested a correlation between precipitation and power load, it was not included as a significant feature in this research. This decision is primarily due to the specific conditions of our study area—Western Inner Mongolia—where precipitation is frequent yet highly variable, having little direct impact on load, especially in short-term forecasts. Moreover, precipitation often works in tandem with temperature and humidity, occasionally influencing the use of air conditioning and heating systems, which in turn can affect load. However, in this study, temperature (particularly maximum temperature) and humidity were found to play a more substantial role in forecasting power load. As a result, we chose to exclude precipitation from the model, aiming to streamline the analysis, minimize potential noise, and ensure that the focus remains on the most influential meteorological variables for accurate predictions.

4.3. Sequence ICEEMDAN Decomposition

Figure 7 presents the decomposition results of ICEEMDAN applied to the daily peak load sequence. The zero-crossing rate of each IMF component obtained through ICEEMDAN decomposition is calculated, as shown in Figure 8. Based on the zero-crossing rates of the IMF components, high-frequency and stationary components are classified. Specifically, IMF1 to IMF3 are combined to form a new high-frequency component, IMF1′; IMF4 and IMF5 are combined into another high-frequency component, IMF2′; and IMF6 is included in a new high-frequency component, IMF2′. The remaining components are merged into a new stationary component, IMF3′. The high-frequency components, IMF1′ and IMF2′, are predicted using SSA-Bagging-XGBoost, while the stationary component, IMF3′, is predicted using MLR.

4.4. Hyperparameter Search Process

In this paper, the SSA is selected to optimize the hyperparameters of XGBoost. As shown in Figure 9, SSA demonstrates significant advantages in both convergence speed and global optimization capability. Experimental results indicate that SSA converges rapidly over multiple iterations and effectively avoids the issue of local optima, making it particularly effective for solving complex, high-dimensional, and multi-modal optimization problems. Compared to PSO, SSA is able to escape local optima during later iterations, ensuring a more thorough exploration of the global optimum. In comparison to GWO and WOA, SSA exhibits faster convergence and greater stability in multi-objective optimization problems. Moreover, when compared to MFO, SSA shows superior search efficiency and global optimization capability. In conclusion, the use of SSA for hyperparameter tuning of XGBoost in this study can significantly enhance the model’s predictive performance.

4.5. Comparative Analysis of Prediction Models

To validate the predictive performance of the ICEEMDAN-Bagging-XGBoost-MLR model proposed in this study, a comparison is made with both ensemble and single models. The model types and their corresponding prediction results are presented in Figure 10 and Table 2. It is worth noting that, given the significant seasonal variations in load patterns, peak load predictions for March, June, September, and December are selected for comparison.
As illustrated in Figure 10 and Table 2, both SVM and LSTM demonstrate strong predictive accuracy when employed as single models in March and December. For instance, in March, SVM yields RMSE values of 177.15 MW, an MAE of 141.84 MW, and a MAPE of 1.65%. However, during June and September, the performance of SVM declines significantly, with RMSE values rising to 576.27 MW in June and 817.91 MW in September, the MAE reaching 361.58 MW and 547.97 MW, and the MAPE increasing to 3.31% and 5.13%, respectively. Similarly, LSTM exhibits favorable results for March and December but shows substantial error increases in June and September. In June, the RMSE is 920.11 MW, MAE is 773.18 MW, and MAPE is 6.47%; in September, these metrics rise to an RMSE of 1137.34 MW, MAE of 942.41 MW, and MAPE of 8.23%. These findings clearly indicate that the accuracy of single models deteriorates significantly when faced with months characterized by high data fluctuations.
Although the XGBoost ensemble model performs well in March and December, its accuracy in June and September remains markedly lower compared to that in other months. For example, in June, the RMSE is 1167.39 MW, MAE is 1032.67 MW, and MAPE is 8.66%, with even higher values in September—the RMSE reaches 1303.76 MW, the MAE reaches 1198.35 MW, and the MAPE reaches 10.38%. This indicates that XGBoost struggles to effectively handle data disturbances, leading to instability in its performance during months with high volatility.
The Prophet and MLR models, both based on linear regression, also fail to adequately address the complex and nonlinear relationships within the data. For instance, Prophet produces RMSE values of 1235.23 MW, MAE of 1039.90 MW, and MAPE of 15.81% in March. In June, the RMSE increases to 1835.26 MW, MAE to 1644.27 MW, and MAPE to 18.41%; and in September, these values reach RMSE of 1300.98 MW, MAE of 1124.42 MW, and MAPE of 11.26%. These results highlight the limitations of Prophet in capturing data volatility, with its linear assumptions unable to account for the nonlinearities in the data, leading to suboptimal predictive performance. Similarly, the MLR model fails to provide satisfactory results. In March, RMSE is 1039.51 MW, MAE is 950.33 MW, and MAPE is 10.95%; in June, RMSE is 1377.68 MW, MAE is 1194.96 MW, and MAPE is 9.92%; in September, RMSE is 1222.65 MW, MAE is 1089.15 MW, and MAPE is 9.41%; and in December, RMSE is 1158.14 MW, MAE is 1015.69 MW, and MAPE is 11.49%. These results suggest that MLR is unable to capture the nonlinear features of the data, leading to significant prediction errors.
In contrast, the proposed model demonstrates consistently high accuracy across all months, particularly in June and September, when the data exhibits significant volatility. Notably, when compared to Prophet and MLR, the proposed model shows a marked improvement in prediction accuracy during June. In Prophet, the RMSE is 1835.26 MW, MAE is 1644.27 MW, and MAPE is 18.41%; in MLR, the RMSE is 1377.68 MW, MAE is 1194.96 MW, and MAPE is 9.92%. In contrast, the proposed model achieves RMSE values of 836.25 MW, MAE of 739.46 MW, and MAPE of 2.35%. Specifically, the RMSE of the proposed model is reduced by 54.4% compared to Prophet and by 39.3% compared to MLR. Meanwhile, the MAPE is reduced by 87.3% compared to Prophet and by 76.3% compared to MLR. These results clearly demonstrate that the proposed model is capable of mitigating the impact of data disturbances, maintaining high prediction accuracy in volatile months, and showcasing its significant advantage in handling complex and fluctuating data.

4.6. Scalability Verification

Building on the previous research, we initially used data from a specific region in Western Inner Mongolia. However, we recognize that these data may not fully capture the range of peak load patterns found in other areas with varying climate conditions, infrastructure, and load characteristics. To strengthen the validation of our model, we expanded the scope of our analysis by including load data from a city in Northern China and its affiliated county-level State Grid Corporation. This dataset encompasses the total load data from over 30 key users, covering a period of three years, from 25 November 2020, to 25 November 2023, with daily load measurements throughout. Consistent with the previous sections, the data format, model training parameters, and comparison models are retained. The focus of the model’s prediction is the peak load for October 2023, with the results summarized in Table 3.
As presented in Table 3, the proposed model consistently outperforms all comparison models across key error metrics, including RMSE, MAE, and MAPE. With an RMSE of 11.44 MW, MAE of 9.47 MW, and MAPE of just 1.28%, the proposed model demonstrates its exceptional ability to predict peak loads with a high degree of accuracy, even when confronted with the intricate dynamics of regional load behaviors. In contrast, models like RF and XGBoost, though robust, show higher prediction errors. RF results in an RMSE of 34.64 MW, MAE of 28.91 MW, and MAPE of 3.79%, while XGBoost produces even higher errors, with an RMSE of 35.82 MW, MAE of 32.90 MW, and MAPE of 4.35%. These figures suggest that while ensemble models are typically effective, they struggle to capture peak load dynamics as effectively as the proposed model in this particular case. Similarly, SVM and LSTM perform reasonably well but still fall short compared to the proposed model. SVM achieves an RMSE of 22.31 MW, MAE of 18.93 MW, and MAPE of 2.48%, while LSTM records an RMSE of 21.59 MW, MAE of 17.80 MW, and MAPE of 2.40%. While these models offer solid performance, they do not reach the level of accuracy demonstrated by the proposed model, particularly in capturing peak load fluctuations.
The RBFNN and ELM models, though competitive, also yield higher errors, with RBFNN showing an RMSE of 37.11 MW and ELM an RMSE of 25.18 MW. Although CNN-LSTM brings a hybrid approach similar to LSTM, it still lags behind, achieving a MAPE of 2.36%, showing that traditional and hybrid models, while useful, do not match the performance of the proposed model in this context. Models based on time-series forecasting, such as Prophet and ARIMA-ML, also exhibit notable shortcomings. Prophet produces an RMSE of 32.58 MW, MAE of 27.02 MW, and MAPE of 4.21%, while ARIMA-ML shows an RMSE of 28.38 MW, MAE of 24.07 MW, and MAPE of 3.26%. These higher errors highlight the difficulty of time-series models in accurately predicting peak load events in this specific case.
In conclusion, the superior performance of the proposed model can be attributed to its hybrid structure, which effectively integrates the strengths of different algorithms while overcoming the limitations of individual models. The results solidify the proposed approach as a highly effective tool for peak load forecasting, demonstrating its reliability and robustness in managing complex, fluctuating load data.

5. Conclusions

This paper proposes a continuous multi-day peak load prediction model based on ICEEMDAN-Bagging-XGBoost-MLR and incorporates the SSA optimization algorithm to identify the optimal hyperparameters of XGBoost, thus reducing the time required for hyperparameter optimization. In the absence of temperature data for the upcoming month, the proposed model is applied to decompose the temperature series, retaining only the trend component that reflects the overall temperature changes, while discarding components with higher complexity and frequencies of variation. The analysis shows that the smoothed temperature series exhibits higher feature contribution scores, indicating that the overall temperature trend is the primary factor driving daily peak load variations, while the random fluctuations in temperature contribute less to these changes. The smoothing process employed is both logical and highly practical, demonstrating its applicability in similar forecasting scenarios. In addition, the proposed model effectively captures and tracks load variations, even during periods of significant fluctuations. When compared to both the ensemble and individual benchmark models, the proposed model consistently exhibits high and stable prediction accuracy across all months of the year, along with strong resistance to overfitting and impressive generalization capabilities. Of particular note is the fact that the RMSE of the proposed model is reduced by 54.4% compared to that of the Prophet model and by 39.3% compared to that of the MLR model. Additionally, the MAPE shows significant reductions of 87.3% and 76.3%, respectively, relative to that of Prophet and MLR. In further scalability assessment, the proposed model also outperformed all benchmark models, demonstrating its robustness.
The current model primarily focuses on daily peak load forecasting, which is highly effective for short-term predictions but may not yield satisfactory results for long-term forecasting. Incorporating external factors, such as seasonal variations, socio-economic changes, and infrastructure development, in future work could extend the model’s applicability to longer-term forecasts.

Author Contributions

Conceptualization, methodology and validation, B.C., J.Y. and S.H.; software, Y.G.; writing—original draft, S.H., B.C., J.Y. and Y.C.; writing—review and editing, S.H., Y.G. and J.Y.; supervision, X.C., B.C., Y.C., Y.W. and X.L.; project administration, B.C., Q.Z., S.H., Y.C., and Y.G.; funding acquisition, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Technological Support Project of Daqingshan Laboratory 2024KYPT0011.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

Authors Bin Cao, Xiaolei Cheng and Qian Zhang were employed by the company Inner Mongolia Daqingshan Laboratory Co., Ltd. Authors Sile Hu, Yu Guo and Xianglong Liu were employed by the company Inner Mongolia Power (Group) Co., Ltd. Author Yuan Wang was employed by the company Inner Mongolia Electric Power Group Mengdian Economic and Technical Research Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Gao, Y.; Tahir, M.; Siano, P.; Bi, Y.; Hu, S.; Yang, J. Optimization of renewable energy-based integrated energy systems: A three-stage stochastic robust model. Appl. Energy 2025, 377, 124635. [Google Scholar] [CrossRef]
  2. Meng, Q.; Jin, X.; Luo, F.; Wang, Z.; Hussain, S. Distributionally robust scheduling for benefit allocation in regional integrated energy system with multiple stakeholders. J. Mod. Power Syst. Clean Energy 2024, 12, 1631–1642. [Google Scholar] [CrossRef]
  3. Wood, M.; Matrone, S.; Ogliari, E.; Leva, S. Comparing peak electricity load forecasting models for an industrial and a residential building. Math. Comput. Simul. 2026, 240, 303–316. [Google Scholar] [CrossRef]
  4. Gao, Y.; Zhao, Y.; Hu, S.; Tahir, M.; Yuan, W.; Yang, J. A three-stage adjustable robust optimization framework for energy base leveraging transfer learning. Energy 2025, 319, 135037. [Google Scholar] [CrossRef]
  5. Hong, Y.Y.; Apolinario, G.F.D.G.; Cheng, Y.H. Week-ahead daily peak load forecasting using hybrid convolutional neural network. IFAC PapersOnLine 2023, 56, 372–377. [Google Scholar] [CrossRef]
  6. Gao, Z.; Yin, X.; Zhao, F.; Meng, H.; Hao, Y.; Yu, M. A two-layer SSA-XGBoost-MLR continuous multi-day peak load forecasting method based on hybrid aggregated two-phase decomposition. Energy Rep. 2022, 8, 12426–12441. [Google Scholar] [CrossRef]
  7. Lianbing, L.I.; Guoqiang, G.A.; Weiguang, C.H.; Wenjie, F.U.; Chao, Z.H.; Shasha, Z.H. Ultra short term load power prediction considering feature recombination and BiGRU-Attention-XGBoost model. Mod. Electr. Power 2023, 41, 1–11. [Google Scholar]
  8. Cai, Q.; Chao, Z.; Su, B.; Wang, L.; Duan, Q.; Wen, Y.; Li, B. Short-term load forecasting method based on a novel robust loss neural network algorithm. Power Syst. Technol. 2020, 44, 4132–4139. [Google Scholar]
  9. Yu, Y.-L.; Li, W.; Sheng, D.-R.; Chen, J.-H. A hybrid short-term load forecasting method based on improved ensemble empirical mode decomposition and back propagation neural network. J. Zhejiang Univ. Sci. A Appl. Phys. Eng. 2016, 17, 101–114. [Google Scholar] [CrossRef]
  10. Pang, H.; Gao, J.; Du, Y. A short-term load probability density prediction based on quantile regression of time convolution network. Power Syst. Technol. 2020, 44, 1343–1350. [Google Scholar]
  11. Fan, S.; Li, L.; Wang, S.; Liu, X.; Yu, Y.; Hao, B. Application analysis and exploration of artificial intelligence technology in power grid dispatch and control. Power Syst. Technol. 2020, 44, 401–411. [Google Scholar]
  12. Wang, S.; Wang, X.; Wang, S.; Wang, D. Bi-directional long short-term memory method based on attention mechanism and rolling update for short-term load forecasting. Int. J. Electr. Power Energy Syst. 2019, 109, 470–479. [Google Scholar] [CrossRef]
  13. Chang, Y.; Sun, H.; Gu, T.; Du, W.; Wang, Y.; Li, W. Monthly forecast of wind power generation using historical data expansion method. Power Syst. Technol. 2021, 45, 1059–1068. [Google Scholar]
  14. Xu, Y.; Xiang, Y.; Ma, T. VMD-GRU short-term power load forecasting model based on optimized parameters of particle swarm algorithm. J. North China Electr. Power Univ. (Nat. Sci. Ed.) 2023, 50, 38–47. [Google Scholar]
  15. Li, Y.; Liu, X.; Xing, F.; Wen, G.; Lu, N.; He, H.; Jiao, R. Daily peak load prediction based on correlation analysis and bi-directional long short-term memory network. Power Syst. Technol. 2021, 45, 2719–2730. [Google Scholar]
  16. Rafi, S.H.; Al-Masood, N.; Deeba, S.R.; Hossain, E. A short-term load forecasting method using integrated CNN and LSTM network. IEEE Access 2021, 9, 32436–32448. [Google Scholar] [CrossRef]
  17. Tang, X.; Dai, Y.; Liu, Q.; Dang, X.; Xu, J. Application of bidirectional recurrent neural network combined with deep belief network in short-term load forecasting. IEEE Access 2019, 7, 160660–160670. [Google Scholar] [CrossRef]
  18. Afrasiabi, M.; Mohammadi, M.; Rastegar, M.; Stankovic, L.; Afrasiabi, S.; Khazaei, M. Deep-based conditional probability density function forecasting of residential loads. IEEE Trans. Smart Grid 2020, 11, 3646–3657. [Google Scholar] [CrossRef]
  19. Al-Rakhami, M.; Gumaei, A.; Alsanad, A.; Alamri, A.; Hassan, M.M. An ensemble learning approach for accurate energy load prediction in residential buildings. IEEE Access 2019, 7, 48328–48338. [Google Scholar] [CrossRef]
  20. Shi, J.; Ma, L.; Li, C.; Liu, N.; Zhang, J. Peak load forecasting method based on serial-parallel ensemble learning. Chin. J. Electr. Eng. 2020, 40, 4463–4472, 4726. [Google Scholar]
  21. Yu, Y.; Wang, Z.; Chen, X.; Feng, Q. Particle swarm optimization algorithm based on teaming behavior. Knowl. Based Syst. 2025, 318, 113555. [Google Scholar] [CrossRef]
  22. Jin, Z.; Li, X.; Qiu, Z.; Li, F.; Kong, E.; Li, B. A data-driven framework for lithium-ion battery RUL using LSTM and XGBoost with feature selection via Binary Firefly Algorithm. Energy 2025, 314, 134229. [Google Scholar] [CrossRef]
  23. Wang, L.; Peng, L.; Xiong, X.; Li, Y.; Qi, Y.; Hu, X. Research on high-speed constant tension spinning control strategy based on vibration detection and enhanced firefly algorithm based FOPID controller. Measurement 2025, 117, 117789. [Google Scholar] [CrossRef]
  24. Meng, Q.; Xu, J.; Ge, L.; Wang, Z.; Wang, J.; Xu, L.; Tang, Z. Economic optimization operation approach of integrated energy system considering wind power consumption and flexible load regulation. J. Electr. Eng. Technol. 2024, 19, 209–221. [Google Scholar] [CrossRef]
  25. Meng, Q.; Zu, G.; Ge, L.; Li, S.; Xu, L.; Wang, R.; He, K.; Jin, S. Dispatching strategy for low-carbon flexible operation of park-level integrated energy system. Appl. Sci. 2022, 12, 12309. [Google Scholar] [CrossRef]
  26. Liang, B.; Feng, W. Bearing fault diagnosis based on ICEEMDAN deep learning network. Processes 2023, 11, 2440. [Google Scholar] [CrossRef]
Figure 1. ICEEMDAN-Bagging-XGBoost-MLR double-layer peak load forecasting model framework.
Figure 1. ICEEMDAN-Bagging-XGBoost-MLR double-layer peak load forecasting model framework.
Applsci 15 11180 g001
Figure 2. XGBoost algorithm mechanism.
Figure 2. XGBoost algorithm mechanism.
Applsci 15 11180 g002
Figure 3. Bagging parallel ensemble learning architecture.
Figure 3. Bagging parallel ensemble learning architecture.
Applsci 15 11180 g003
Figure 4. Original peak load sequence.
Figure 4. Original peak load sequence.
Applsci 15 11180 g004
Figure 5. Spatial distribution of peak load historical data. (a) Historical data of daily peak load and spatial distribution of average temperature. (b) Historical data of daily peak load and spatial distribution of relative humidity. (c) Historical data of daily peak load and spatial distribution of rainfall.
Figure 5. Spatial distribution of peak load historical data. (a) Historical data of daily peak load and spatial distribution of average temperature. (b) Historical data of daily peak load and spatial distribution of relative humidity. (c) Historical data of daily peak load and spatial distribution of rainfall.
Applsci 15 11180 g005
Figure 6. Input feature contribution analysis. (a) RF feature contribution analysis. (b) XGBoost feature contribution analysis.
Figure 6. Input feature contribution analysis. (a) RF feature contribution analysis. (b) XGBoost feature contribution analysis.
Applsci 15 11180 g006
Figure 7. ICCEMDAN decomposition results.
Figure 7. ICCEMDAN decomposition results.
Applsci 15 11180 g007
Figure 8. Zero-crossing rate of each IMF component.
Figure 8. Zero-crossing rate of each IMF component.
Applsci 15 11180 g008
Figure 9. Test results of different group optimization algorithms.
Figure 9. Test results of different group optimization algorithms.
Applsci 15 11180 g009aApplsci 15 11180 g009b
Figure 10. The absolute error of each model prediction.
Figure 10. The absolute error of each model prediction.
Applsci 15 11180 g010
Table 1. A comparison between our approach and the latest research.
Table 1. A comparison between our approach and the latest research.
Ref.Hybrid ModelTree Ensemble AlgorithmHeuristic Algorithm
[7]××
[8,9,10,11]×××
[12,13]××
[14]×××
[19]×
[20]PSO
[21]GA
ProposedSSA
The cross and checkmark indicate the absence and presence of the method, respectively.
Table 2. Comparison of prediction errors of each algorithm.
Table 2. Comparison of prediction errors of each algorithm.
AlgorithmIndicatorsMarchJuneSeptemberDecember
e R M S E M W 177.15576.27817.91263.61
Proposed e M A E M W 141.84361.58547.97193.36
e M A P E % 1.653.315.132.23
e R M S E M W 207.17920.111137.34296.93
XGBoost e M A E M W 173.76773.18942.41241.15
e M A P E % 1.996.478.232.75
e R M S E M W 731.851167.391303.761021.70
RF e M A E M W 666.011032.671198.35940.47
e M A P E % 7.558.6610.3810.54
e R M S E M W 182.65732.981041.73241.24
SVM e M A E M W 139.56575.15769.36174.64
e M A P E % 1.605.026.851.99
e R M S E M W 220.55922.711317.93316.98
LSTM e M A E M W 190.11800.631090.02202.57
e M A P E % 2.206.829.402.30
e R M S E M W 716.191205.611603.76886.89
RBFNN e M A E M W 658.711120.361463.08781.68
e M A P E % 7.489.5912.588.80
e R M S E M W 529.591262.321123.58668.41
ELM e M A E M W 430.841113.24934.17570.69
e M A P E % 4.869.378.206.43
CNN-LSTM e R M S E M W 205.16836.251003.27289.24
e M A E M W 177.77739.46830.53221.85
e M A P E % 1.842.353.272.64
Prophet e R M S E M W 1235.231835.261300.981653.01
e M A E M W 1039.91644.271124.421265.07
e M A P E % 15.8118.4111.2615.28
ARIMA-ML e R M S E M W 976.531022.18989.361021.02
e M A E M W 836.66898.4749.1815.38
e M A P E % 9.539.829.669.91
e R M S E M W 1039.511377.681222.651158.14
MLR e M A E M W 950.331194.961089.151015.69
e M A P E % 10.959.929.4111.49
Table 3. Comparison of prediction errors of each algorithm.
Table 3. Comparison of prediction errors of each algorithm.
Algorithm e R M S E M W e M A E M W e M A P E %
Proposed11.449.471.28
RF34.6428.913.79
XGBoost35.8232.904.35
SVM22.3118.932.48
LSTM21.5917.802.40
RBFNN37.1132.964.34
ELM25.1821.502.94
CNN-LSTM20.1817.232.36
Prophet32.5827.024.21
ARIMA-ML28.3824.073.26
MLR30.6723.493.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, B.; Chen, Y.; Hu, S.; Guo, Y.; Liu, X.; Wang, Y.; Cheng, X.; Zhang, Q.; Yang, J. Daily Peak Load Prediction Method Based on XGBoost and MLR. Appl. Sci. 2025, 15, 11180. https://doi.org/10.3390/app152011180

AMA Style

Cao B, Chen Y, Hu S, Guo Y, Liu X, Wang Y, Cheng X, Zhang Q, Yang J. Daily Peak Load Prediction Method Based on XGBoost and MLR. Applied Sciences. 2025; 15(20):11180. https://doi.org/10.3390/app152011180

Chicago/Turabian Style

Cao, Bin, Yahui Chen, Sile Hu, Yu Guo, Xianglong Liu, Yuan Wang, Xiaolei Cheng, Qian Zhang, and Jiaqiang Yang. 2025. "Daily Peak Load Prediction Method Based on XGBoost and MLR" Applied Sciences 15, no. 20: 11180. https://doi.org/10.3390/app152011180

APA Style

Cao, B., Chen, Y., Hu, S., Guo, Y., Liu, X., Wang, Y., Cheng, X., Zhang, Q., & Yang, J. (2025). Daily Peak Load Prediction Method Based on XGBoost and MLR. Applied Sciences, 15(20), 11180. https://doi.org/10.3390/app152011180

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop