Short-Term Electrical Load Forecasting Using an Enhanced Extreme Learning Machine Based on the Improved Dwarf Mongoose Optimization Algorithm

: Accurate short-term electrical load forecasting is crucial for the stable operation of power systems. Given the nonlinear, periodic, and rapidly changing characteristics of short-term power load forecasts, this paper introduces a novel forecasting method employing an Extreme Learning Machine (ELM) enhanced by an improved Dwarf Mongoose Optimization Algorithm (Local escape Dwarf Mongoose Optimization Algorithm, LDMOA). This method addresses the significant prediction errors of conventional ELM models and enhances prediction accuracy. The enhancements to the Dwarf Mongoose Optimization Algorithm include three key modifications: initially, a dynamic backward learning strategy is integrated at the early stages of the algorithm to augment its global search capabilities. Subsequently, a cosine algorithm is employed to locate new food sources, thereby expanding the search scope and avoiding local optima. Lastly, a “madness factor” is added when identifying new sleeping burrows to further widen the search area and effectively circumvent local optima. Comparative analyses using benchmark functions demonstrate the improved algorithm’s superior convergence and stability. In this study, the LDMOA algorithm optimizes the weights and thresholds of the ELM to establish the LDMOA-ELM prediction model. Experimental forecasts utilizing data from China’s 2016 “The Electrician Mathematical Contest in Modeling” demonstrate that the LDMOA-ELM model significantly outperforms the original ELM model in terms of prediction error and accuracy.


Introduction
Accurate electric load forecasting is crucial for the planning and reliable economic operation of power systems.It not only ensures the normal electricity usage of consumers but also reduces costs and guarantees the safety of power systems [1].However, challenges in predicting electric load demand have increased sharply due to factors such as global climate change, energy supply constraints, increasing numbers of electricity users, and the integration of new energy device loads into the grid [2].
In the realm of electric load forecasting, researchers have employed time series regression models [3] and fuzzy linear regression models [4] to predict load, focusing on the temporal characteristics of load data and providing strong interpretability of the models.However, these methods have limitations in forecasting non-linear load data.In recent years, with the development of intelligent optimization algorithms, an increasing number of researchers have started incorporating these algorithms into electric load forecasting.Han M. C. enhanced the capability and prediction accuracy of capturing characteristics in load data by optimizing LSTM hyperparameters through a sparrow optimization algorithm that integrates Cauchy mutation and inverse learning strategies [5].Zhang Z. C. quantified the behavior of dragonflies in the dragonfly algorithm to boost the search capability, and Symmetry 2024, 16, 628 2 of 15 utilized an adaptive noise complete empirical mode decomposition method for preprocessing raw data, thereby improving the prediction accuracy of SVR in load forecasting [6].Ge Q. B. employed K-means clustering to categorize data and then used a combined predictive algorithm of reinforcement learning and particle swarm optimization along with the least squares support vector machine to predict different types of data [7].Fan G. F. developed a new model combining the random forest model and mean-generating function model, significantly enhancing the prediction accuracy of peaks and troughs in highly volatile data [8].Additionally, Xu R. [9] noted that extreme learning machines offer faster learning speeds and less human intervention, and are easier to implement.Deng B. [10] argued that compared to support vector machines, extreme learning machines have milder optimization constraints and quicker learning speeds.Some researchers have also applied optimized ELMs to electric load forecasting.For instance, Wang Tong utilized an improved artificial hummingbird algorithm for optimizing parameters in Extreme Learning Machines (ELM), significantly enhancing prediction accuracy [11].Long Gan and others have used an improved multiverse algorithm to optimize the input layer weights and thresholds of ELMs, thereby improving their prediction accuracy [12].Wang Z-X.employed an adaptive evolutionary ELM for data prediction, integrating a chaos-adapted whale optimization algorithm based on a firefly perturbation strategy and a chaotic sparrow search algorithm, which exhibited outstanding performance [13].Additionally, Zhang S. proposed an ELM model under a moth flame optimization algorithm based on Tsne dimensionality reduction and visualization analysis, which achieved higher prediction accuracy than the original ELM model [14].
Accurate electric load forecasting can impact related decisions in power systems, such as generation control, economic dispatch, and maintenance scheduling.Therefore, to achieve high-precision short-term electric load forecasting, this paper proposes an ELM prediction model based on the improved Dwarf Mongoose Optimization Algorithm.Applied to short-term electric load forecasting, experimental results demonstrate that this model achieves higher accuracy compared to other ELM models.

Extreme Learning Machine
The Extreme Learning Machine (ELM) is a type of Single-hidden Layer Feedforward Neural Network (SLFN) algorithm, introduced by Professor Guang-Bin Huang and others based on the theory of the Moore-Penrose pseudoinverse [15].This algorithm was developed to address several issues inherent in SLFNs, such as slow learning rates, long iteration times, and the traditional need to preset learning rates and step sizes.Unlike conventional neural network learning algorithms, the ELM requires only the appropriate setting of hidden layer node numbers.It autonomously generates all necessary parameters for the hidden layer and determines the final output layer weights through the least squares method.Due to its superior learning and nonlinear approximation capabilities compared to traditional machine learning algorithms, researchers have applied the ELM across a broad range of fields, including fault diagnosis [16], load forecasting [17], and feature recognition [18].
The ELM algorithm operates with a single hidden layer, where each layer from input to output comprises independent neurons, all interconnected in a fully connected manner.The network structure of the ELM is illustrated in Figure 1.Assuming there are N arbitrary samples (  ,  ),   = [ 1 ,  2 … ,   ]  ∈   and   = [ 1 ,  2 , ⋯ ,   ]  ∈   .This can be represented by a single hidden layer neural network with L hidden nodes, as illustrated in Figure 1 and described by Equation (1).
Here, () is the activation function,   = [ 1 ,  2 , ⋯ ,   ]  are the input weights,   = [ 1 ,  2 , ⋯ ,   ]  ∈   are the output weights, and   is the bias of the ith hidden layer unit, with   •   being the dot product between them.
Extreme Learning Machines, as a type of single hidden layer neural network, have an output error that asymptotically approaches zero, as shown in Equation (2).
Equation ( 2) can be represented using matrices, as shown in Equation (3): Here,  denotes the hidden layer output matrix,  represents the output weights, and  is the target output.The matrix  can be expressed by Equation (4).

Dwarf Mongoose Algorithm
The Dwarf Mongoose Algorithm is an intelligent optimization algorithm inspired by the social behavior of dwarf mongoose groups.This algorithm consists of three parts: the Alpha Group, the Scout Group, and the Babysitter Group.The Alpha Group produces a female leader to guide the group in foraging.The Scout Group is responsible for finding new locations for sleeping mounds, while the Babysitter Group influences the performance of the algorithm through its numbers.

Alpha Group
The population is initialized as shown in Equation ( 6), Assuming there are N arbitrary samples (x i , t i ), x i = x i1 , x i2 . . ., x in ] T ∈ R n and This can be represented by a single hidden layer neural network with L hidden nodes, as illustrated in Figure 1 and described by Equation (1).
Here, g(x) is the activation function, are the output weights, and b i is the bias of the ith hidden layer unit, with w i •x i being the dot product between them.
Extreme Learning Machines, as a type of single hidden layer neural network, have an output error that asymptotically approaches zero, as shown in Equation (2).
Equation ( 2) can be represented using matrices, as shown in Equation (3): Here, H denotes the hidden layer output matrix, β represents the output weights, and T is the target output.The matrix H can be expressed by Equation (4).
In Equation ( 5), H T represents the Moore-Penrose pseudoinverse of the matrix H.

Dwarf Mongoose Algorithm
The Dwarf Mongoose Algorithm is an intelligent optimization algorithm inspired by the social behavior of dwarf mongoose groups.This algorithm consists of three parts: the Alpha Group, the Scout Group, and the Babysitter Group.The Alpha Group produces a female leader to guide the group in foraging.The Scout Group is responsible for finding new locations for sleeping mounds, while the Babysitter Group influences the performance of the algorithm through its numbers.

Alpha Group
The population is initialized as shown in Equation ( 6), where x i,j represents the initial position, LB and UB denote the lower and upper bounds of the solution space, Dim represents the dimension of decision variables, and uni f rnd is a uniformly distributed random number.Furthermore, a female leader emerges within the dwarf mongoose population, as depicted in Equation (7).
Here, f it i represents the fitness value of the ith individual, and N is the number of individuals in the population.The number of individuals in the Alpha Group is the total population N minus the number of individuals in the Babysitter Group, i.e., n = N − bs.
The female leader in the Alpha Group guides the other members to the food source location via calls, as shown in Equation ( 8): Here, X new is the new position of the dwarf mongoose, peep is the calling coefficient, set at peep = 2 in this study, and phi is a uniformly distributed random number within [0, 1].X i is the current position of the female leader, and X k is the position of another random individual in the Alpha Group distinct from the leader.Subsequently, the new position X new undergoes a fitness evaluation to obtain f it i+1 , and the value of the sleeping mound is determined according to Equation (9).
Here, sm i represents the value of the sleeping mound, and the average value of the sleeping mound can be calculated according to Equation (10).

Scout Group
The primary responsibility of the Scout Group is to locate new positions for sleeping mounds, as described by the movement formula in Equation (11).
Here, r is a random number within [0, 1], CF is a mobility parameter for the dwarf mongoose population, and → M is the direction vector determining the mongoose's movement direction.The formulas for calculating CF and → M are given in Equations ( 12) and ( 13), respectively.
Symmetry 2024, 16, 628 Here, Iter is the current iteration number, and MaxIt is the maximum number of iterations.The movement parameter CF linearly decreases as the number of iterations increases.

Babysitter Group
When the timing parameter is greater than or equal to the exchange parameter, i.e., C ≥ L, the Babysitter Group assumes that the Alpha Group's foraging capability is weak.At this point, the Babysitter Group will swap roles with the Alpha Group, and the dwarf mongoose community will begin searching for a new sleeping mound.

Enhancements to the Dwarf Mongoose Algorithm
To address the Dwarf Mongoose Algorithm's tendency to fall into local optima and its weak global search performance, this paper proposes the incorporation of a reverse learning strategy to enhance the algorithm's exploratory capability, thereby improving its global search performance.Additionally, the inclusion of a craziness operator factor and the sine-cosine algorithm expands the local search range of the algorithm, helping to circumvent issues of local optima.

Dynamic Reverse Learning Strategy
The reverse learning strategy [19] is a common perturbation tactic that expands the algorithm's exploration range to find better solutions.In the reversed learning strategy, the new position generated is symmetrical to the original position at the point X i +x i 2 .This measure enhances the exploratory nature of the algorithm, allowing it to search in the opposite direction for improved population individual positions.Moreover, the new position is compared with the original in terms of fitness, and the individual with the optimal fitness is selected as the population individual position.The reverse learning strategy is depicted in Equation ( 14).
In the formula, X i represents the population individual after reverse learning, with LB and UB denoting the lower and upper limits of the solution space, and x i indicating the original position of the individual within the population.To enhance the search for optimal solutions within the solution space, a random factor is included in the reverse learning strategy, further diversifying the population within the solution space.The dynamic reverse learning strategy introduced in the early stages of the algorithm iteration is shown in Equation (15).
Here, r is a random number within (0, 1).

Sine-Cosine Algorithm
During the search process, the sine-cosine algorithm [20] conducts searches in the form of sine and cosine waveforms.This method enhances the search capabilities of the algorithm, enabling it to avoid becoming trapped in local optima.Additionally, the search process of the algorithm exhibits point symmetry characteristics typical of sine and cosine functions.While the Alpha Group, led by the female leader, is searching for new food sources, this process can easily become trapped in local optima.To avoid such outcomes, this study incorporates the sine-cosine algorithm to expand the search range of the algorithm.The formula for searching new food sources, updated with the sine-cosine algorithm, is shown in Equation ( 16).
In Equation ( 16), r and r 1 are random numbers within (0, 1), r 2 is a random number within (0, 2π), X i is the position of the female leader, and X k is the position of another individual distinct from the leader.

Craziness Factor
In the later stages of the algorithm, when dwarf mongoose individuals seek a sleeping mound, the group tends to converge on this mound, which could lead to local optima.This paper introduces a craziness operator factor, which perturbs the position of the optimal individual to prevent the algorithm from becoming trapped in local optima in its later iterations.The position of the optimal individual after incorporating the craziness operator factor is illustrated in Equation ( 17).
In Equation ( 17), x i represents the original optimal individual position, and Pc and xcraze are disturbance factors within the craziness operator factor, with xcraze set at 0.0001.Pc and xcraze are described by Equations ( 18) and ( 19), respectively.
In the craziness factor, the sign is determined as either 1 or −1 based on the magnitude of c, exhibiting a kind of symmetry in its values.This method of value assignment can perturb the algorithm, expanding its search range and helping to avoid local optima.In Equation (19), c is a random number within (0, 1), and P r is the preset craziness probability, set at 0.4 in this study.

LDMOA Steps and Process
In the enhanced DMOA, the reverse learning strategy is utilized to expand the algorithm's exploratory capacity and search range, thereby enhancing its global search capabilities.Simultaneously, the introduction of the sine-cosine algorithm and the craziness operator factor enhance the local search capability of the algorithm, effectively avoiding situations of local optima.Figure 2 shows the workflow diagram of the LDMOA, and below is the operational process of the LDMOA.
Step 1: Set the initial parameters of the algorithm, such as population parameters, dimensions of the solution space and its limits, and maximum iteration parameters, and utilize the dynamic reverse learning strategy to expand the search range.
Step 2: Select the female leader according to Equation ( 7) and set the related coefficients.During the process of searching for new food sources by the dwarf mongoose group, incorporate the sine-cosine algorithm to further expand the search for new food source positions.
Step 3: The sleeping mound position is influenced by the optimal position; introduce the craziness operator factor to perturb it, and then determine the sleeping mound position and calculate its average value.
Step 4: Assess whether C ≥ L; when this condition is met, swap the Alpha Group and Babysitter Group, and proceed with the formula to search for new sleeping mounds and forage.
Step 5: Determine whether the algorithm has reached the maximum iteration count; if not, repeat the above steps, and otherwise output the optimal results.

Benchmark Function Testing
To ensure that the enhanced strategy provides positive improvements over the original Dwarf Mongoose Optimization Algorithm (DMOA), we conducted benchmark function tests comparing the modified algorithm with the original.The selected benchmark functions are shown in Table 1.
Functions  1 to  2 are unimodal functions, which test the algorithm's convergence capability.Functions  3 to  6 are multimodal functions, evaluating the algorithm's ability to escape local optima.Functions  7 to  11 are hybrid functions, and  12 to  15 are composite functions, with both sets testing the optimization performance in complex scenarios.Shifted and Rotated Zakharov Function 300

Benchmark Function Testing
To ensure that the enhanced strategy provides positive improvements over the original Dwarf Mongoose Optimization Algorithm (DMOA), we conducted benchmark function tests comparing the modified algorithm with the original.The selected benchmark functions are shown in Table 1.
Functions f 1 to f 2 are unimodal functions, which test the algorithm's convergence capability.Functions f 3 to f 6 are multimodal functions, evaluating the algorithm's ability to escape local optima.Functions f 7 to f 11 are hybrid functions, and f 12 to f 15 are composite functions, with both sets testing the optimization performance in complex scenarios.
The enhanced and original algorithms were tested using the benchmark functions listed in the table.To ensure the accuracy of the benchmark tests, the algorithms were configured with parameters as shown in Table 2, including population initialization size (nPop) and solution space dimensions (Dim).The results are presented in Section 4.6 and Table 3.

Results of Benchmark Function Testing
In the benchmark function test results, the mean and standard deviation reflect the convergence performance and stability of the algorithms, respectively, with the best values highlighted in bold.The enhanced algorithm shows superior convergence performance and stability compared to the original algorithm.As demonstrated in Table 3, the enhanced algorithm outperforms the original in unimodal functions f 1 to f 2 , multimodal functions f 3 to f 6 , hybrid functions f 7 to f 11 , and composite functions f 12 to f 15 .Based on the analyses in Section 4.5, the enhanced algorithm surpasses the original in convergence, avoiding local optima, and handling complex optimization problems.Figure 3

Results of Benchmark Function Testing
In the benchmark function test results, the mean and standard deviation reflect the convergence performance and stability of the algorithms, respectively, with the best values highlighted in bold.The enhanced algorithm shows superior convergence performance and stability compared to the original algorithm.As demonstrated in Table 3, the enhanced algorithm outperforms the original in unimodal functions  1 to  2 , multimodal functions  3 to  6 , hybrid functions  7 to  11 , and composite functions  12 to  15 .Based on the analyses in Section 4.5, the enhanced algorithm surpasses the original in convergence, avoiding local optima, and handling complex optimization problems.

Establishing the LDMOA-ELM Model
In the Extreme Learning Machine (ELM) algorithm, the weights w and biases b significantly influence prediction outcomes.Given that these parameters are randomly generated in ELM, their randomness can significantly affect the model's prediction accuracy.This study uses the LDMOA algorithm to optimize the weights w and biases b in the ELM algorithm, leading to notable improvements in prediction error reduction and accuracy enhancement.The LDMOA-ELM model development process is outlined as follows: Step 1: Import original load data, normalize it, and split it into training and test sets.Step 2: Initialize DMOA parameters and use Equations ( 15)-( 17) to optimize the initial population, the Alpha Group's foraging process, and the process of finding new sleeping mounds, respectively, resulting in the LDMOA.
Step 3: Compute the fitness function, using the MAPE of the ELM training set as the fitness measure.
Step 4: Exit the loop if the maximum number of iterations is reached or accuracy requirements are met; otherwise, repeat Steps 2 and 3.
Step 5: Use the optimized parameters as the input weights and biases for the ELM model, and then perform numerical predictions and output the model evaluation metrics.
The steps for establishing the LDMOA-ELM model are depicted in Figure 4.
significantly influence prediction outcomes.Given that these parameters are randomly generated in ELM, their randomness can significantly affect the model's prediction accuracy.This study uses the LDMOA algorithm to optimize the weights  and biases  in the ELM algorithm, leading to notable improvements in prediction error reduction and accuracy enhancement.The LDMOA-ELM model development process is outlined as follows: Step 1: Import original load data, normalize it, and split it into training and test sets.
Step 2: Initialize DMOA parameters and use Equations ( 15)-( 17) to optimize the initial population, the Alpha Group's foraging process, and the process of finding new sleeping mounds, respectively, resulting in the LDMOA.
Step 3: Compute the fitness function, using the MAPE of the ELM training set as the fitness measure.
Step 4: Exit the loop if the maximum number of iterations is reached or accuracy requirements are met; otherwise, repeat Steps 2 and 3.
Step 5: Use the optimized parameters as the input weights and biases for the ELM model, and then perform numerical predictions and output the model evaluation metrics.
The steps for establishing the LDMOA-ELM model are depicted in Figure 4.

Simulation Experiment
In conducting electric load forecasting experiments for comparison, to ensure the efficacy of the improved algorithm, this study pits the proposed LDMOA-ELM algorithm against both the original ELM algorithm and the ELM algorithm optimized by the original Dwarf Mongoose Algorithm.The parameter settings for both the original and the enhanced Dwarf Mongoose Algorithms are detailed in Table 4, where nPop is the population initialization size, Dim is the dimension of the solution space, LB is the lower bound, and UB is the upper bound of the solution space.
Here, y i represents the hourly forecasted electric load; _ y i is the average of the hourly electric load data; y i is the actual hourly electric load data; and n is the number of data points.
In these metrics, MSE and RMSE values are within '[0, +∞)', where a value closer to 0 indicates perfect model prediction and, conversely, a higher value indicates greater prediction error.The MAE follows the same range and interpretation.An R 2 value closer to 1 indicates a better fit, whereas a lower value indicates a poorer fit.

Forecasting Results Comparison
The paper utilizes the standard dataset provided by the 2016 "The Electrician Mathematical Contest in Modeling" in China, with sampling every 15 min, resulting in 96 samples per day and a total of 35,040 samples.To ensure consistency in experimental results, the number of hidden nodes in the prediction model is uniformly set to 85, with the sample configuration using data from the previous seven days to predict the eighth day, set across 100 sample groups.In this paper, we optimized the ELM model for multi-step-ahead forecasting by dividing 100 sample sets into 99 training sets and one test set.The statistical data of the three methods' predictions are shown in Table 5. Section 6.1 demonstrates that a higher R 2 value indicates better prediction results, and smaller values for other metrics indicate better performance.Here, the LDMOA-ELM model's MAE, MAPE, MSE, and RMSE values were 61.62, 0.0080845, 5953.4,and 77.158, respectively, all of which were significantly reduced compared to the ELM model.Integrating these five evaluation metrics, it is evident that the LDMOA-ELM used in this study exhibits superior performance across all indicators, with a prediction accuracy of 99.80%, which is an improvement of 15% over the ELM model.This demonstrates that the LDMOA-ELM model achieves lower prediction errors and higher prediction accuracy.The results indicate that the improved predictive model enhances the accuracy of forecasts for the experimental data used in this study.As illustrated in Figure 5, LDMOA is more adept at escaping local optima and finding optimal values compared to the original Dwarf Mongoose Algorithm.Figure 6 shows the prediction results graph, indicating that the LDMOA-ELM model's prediction curve best fits the actual value curve.Figure 7 displays the relative prediction errors, revealing that the relative error between the predicted values of the LDMOA-ELM model and the actual values is significantly lower than that of the ELM model.Additionally, by integrating the statistical data from Table 5, it is evident that the LDMOA-ELM model exhibits lower prediction errors and improved accuracy compared to the original ELM model.Particularly in terms of relative prediction errors, as shown in Figure 7, there is a significant difference between the two, with the LDMOA-ELM model outperforming the original ELM model.
improvement of 15% over the ELM model.This demonstrates that the LDMOA-ELM model achieves lower prediction errors and higher prediction accuracy.The results indicate that the improved predictive model enhances the accuracy of forecasts for the experimental data used in this study.As illustrated in Figure 5, LDMOA is more adept at escaping local optima and finding optimal values compared to the original Dwarf Mongoose Algorithm.Figure 6 shows the prediction results graph, indicating that the LDMOA-ELM model's prediction curve best fits the actual value curve.Figure 7 displays the relative prediction errors, revealing that the relative error between the predicted values of the LDMOA-ELM model and the actual values is significantly lower than that of the ELM model.Additionally, by integrating the statistical data from Table 5, it is evident that the LDMOA-ELM model exhibits lower prediction errors and improved accuracy compared to the original ELM model.Particularly in terms of relative prediction errors, as shown in Figure 7, there is a significant difference between the two, with the LDMOA-ELM model outperforming the original ELM model.5 and Figures 5-8, it is evident that the LDMOA-ELM algorithm, in comparison to both the ELM and DMOA-ELM algorithms, achieves the smallest prediction errors and the highest prediction accuracy, with the LDMOA-ELM algorithm achieving a prediction accuracy of 99.80%.Figure 9 presents the statistical graph for the Mean Absolute Percentage Error (MAPE).It is evident that the original ELM model exhibits significantly higher MAPE values compared to the LDMOA-ELM model.This demonstrates that the LDMOA-ELM model achieves lower prediction errors.5 and Figures 5-8, it is evident that the LDMOA-ELM algorithm, in comparison to both the ELM and DMOA-ELM algorithms, achieves the smallest prediction errors and the highest prediction accuracy, with the LDMOA-ELM algorithm achieving a prediction accuracy of 99.80%.Figure 9

Conclusions
In response to the challenges of high randomness and low prediction accuracy in short-term electric load forecasting, this paper introduces a short-term electric load forecasting model that utilizes an enhanced Dwarf Mongoose Algorithm-based ELM.Initially, the Dwarf Mongoose Algorithm was modified by incorporating reverse learning strategies, sine-cosine strategies, and a craziness operator factor, which improved the algorithm's exploratory capabilities, enhanced its global search ability, and enabled it to escape from local optima more effectively.Subsequently, combining the LDMOA with an Extreme Learning Machine, this model was applied to forecast the relevant experimental data and subjected to experimental analysis.The results demonstrate that, compared to the original ELM and DMOA-ELM models, the LDMOA-ELM model exhibits significantly higher accuracy in predicting short-term electric loads.
The LDMOA-ELM model proposed in this paper exhibits a Mean Absolute Error (MAE) of 61.62, which is significantly lower than that of other models, thereby reducing the prediction error of the Extreme Learning Machine to a certain extent.Although the accuracy of the LDMOA-ELM model surpasses that of the original ELM model, the improvement in predictive accuracy is not markedly evident when compared with other models.Future research should focus on further refining the optimization algorithms and selecting more appropriate predictive data to verify the accuracy and applicability of the predictive model.At the same time, subsequent research projects should consider the processing of raw data and comparisons between different methodologies.

Conclusions
In response to the challenges of high randomness and low prediction accuracy in shortterm electric load forecasting, this paper introduces a short-term electric load forecasting model that utilizes an enhanced Dwarf Mongoose Algorithm-based ELM.Initially, the Dwarf Mongoose Algorithm was modified by incorporating reverse learning strategies, sine-cosine strategies, and a craziness operator factor, which improved the algorithm's exploratory capabilities, enhanced its global search ability, and enabled it to escape from local optima more effectively.Subsequently, combining the LDMOA with an Extreme Learning Machine, this model was applied to forecast the relevant experimental data and subjected to experimental analysis.The results demonstrate that, compared to the original ELM and DMOA-ELM models, the LDMOA-ELM model exhibits significantly higher accuracy in predicting short-term electric loads.
The LDMOA-ELM model proposed in this paper exhibits a Mean Absolute Error (MAE) of 61.62, which is significantly lower than that of other models, thereby reducing the prediction error of the Extreme Learning Machine to a certain extent.Although the accuracy of the LDMOA-ELM model surpasses that of the original ELM model, the improvement in predictive accuracy is not markedly evident when compared with other models.Future research should focus on further refining the optimization algorithms and selecting more appropriate predictive data to verify the accuracy and applicability of the predictive model.At the same time, subsequent research projects should consider the processing of raw data and comparisons between different methodologies.
Furthermore, future research in load forecasting could consider incorporating methods such as Variational Mode Decomposition for preprocessing the raw data and compare it with other machine learning prediction methods to highlight the advanced nature of the optimized model.It could also be beneficial to explore the impact of varying the number of nodes in the prediction algorithm on the precision of the forecasts, aiming to achieve better load prediction outcomes.

Figure 1 .
Figure 1.Extreme learning machine network structure.

Figure 1 .
Figure 1.Extreme learning machine network structure.
includes graphical representations of some function iterations: (a) unimodal function, (b) multimodal function, (c) hybrid function, and (d) composite function, all showing improved iterative convergence results for the enhanced algorithm.

Figure 7 .
Figure 7. Relative error between predicted and actual load.

Figure 8 ,
Figure8, the prediction evaluation metrics graph, shows that the ELM model results in the highest error values, while the LDMOA-ELM model yields the smallest error values.Integrating data from Table5and Figures5-8, it is evident that the LDMOA-ELM algorithm, in comparison to both the ELM and DMOA-ELM algorithms, achieves the smallest prediction errors and the highest prediction accuracy, with the LDMOA-ELM algorithm achieving a prediction accuracy of 99.80%.Figure9presents the statistical graph for the Mean Absolute Percentage Error (MAPE).It is evident that the original ELM model exhibits significantly higher MAPE values compared to the LDMOA-ELM model.This demonstrates that the LDMOA-ELM model achieves lower prediction errors.

Figure 7 .
Figure 7. Relative error between predicted and actual load.

Figure 8 ,
Figure8, the prediction evaluation metrics graph, shows that the ELM model results in the highest error values, while the LDMOA-ELM model yields the smallest error values.Integrating data from Table5and Figures5-8, it is evident that the LDMOA-ELM algorithm, in comparison to both the ELM and DMOA-ELM algorithms, achieves the smallest prediction errors and the highest prediction accuracy, with the LDMOA-ELM algorithm achieving a prediction accuracy of 99.80%.Figure9presents the statistical graph for the Mean Absolute Percentage Error (MAPE).It is evident that the original ELM model exhibits significantly higher MAPE values compared to the LDMOA-ELM model.This demonstrates that the LDMOA-ELM model achieves lower prediction errors.

Table 3 .
Function test result.
presents the statistical graph for the Mean Absolute Percentage Error (MAPE).It is evident that the original ELM model exhibits significantly higher MAPE values compared to the LDMOA-ELM model.This demonstrates that the LDMOA-ELM model achieves lower prediction errors.