Next Article in Journal
Risk Analysis of Five-Axis CNC Water Jet Machining Using Fuzzy Risk Priority Numbers
Previous Article in Journal
Symmetry-Informed Optimization and Verification of Loader Working Device Based on Improved Genetic Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimized Multiple Regression Prediction Strategies with Applications

1
College of Artificial Intelligence, Yango University, Fuzhou 350015, China
2
School of Artificial Intelligence, Nanjing University of Information Science and Technology, Nanjing 210044, China
3
Department of Automotive Engineering, Bursa Uludag University, Bursa 16000, Turkey
4
College of Computer Science and Engineering, Shandong University of Science and Technology, Qingdao 266590, China
5
Department of Information Management, Chaoyang University of Technology, Taichung 41349, Taiwan
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(7), 1085; https://doi.org/10.3390/sym17071085
Submission received: 29 April 2025 / Revised: 3 June 2025 / Accepted: 21 June 2025 / Published: 7 July 2025
(This article belongs to the Section Computer)

Abstract

As a classical statistical method, multiple regression is widely used for forecasting tasks in power, medicine, finance, and other fields. The rise of machine learning has led to the adoption of neural networks, particularly Long Short-Term Memory (LSTM) models, for handling complex forecasting problems, owing to their strong ability to capture temporal dependencies in sequential data. Nevertheless, the performance of LSTM models is highly sensitive to hyperparameter configuration. Traditional manual tuning methods suffer from inefficiency, excessive reliance on expert experience, and poor generalization. Aiming to address the challenges of complex hyperparameter spaces and the limitations of manual adjustment, an enhanced sparrow search algorithm (ISSA) with adaptive parameter configuration was developed for LSTM-based multivariate regression frameworks, where systematic optimization of hidden layer dimensionality, learning rate scheduling, and iterative training thresholds enhances its model generalization capability. In terms of SSA improvement, first, the population is initialized by the reverse learning strategy to increase the diversity of the population. Second, the mechanism for updating the positions of producer sparrows is improved, and different update formulas are selected based on the sizes of random numbers to avoid convergence to the origin and improve search flexibility. Then, the step factor is dynamically adjusted to improve the accuracy of the solution. To improve the algorithm’s global search capability and escape local optima, the sparrow search algorithm’s position update mechanism integrates Lévy flight for detection and early warning. Experimental evaluations using benchmark functions from the CEC2005 test set demonstrated that the ISSA outperforms PSO, the SSA, and other algorithms in optimization performance. Further validation with power load and real estate datasets revealed that the ISSA-LSTM model achieves superior prediction accuracy compared to existing approaches, achieving an RMSE of 83.102 and an R 2 of 0.550 during electric load forecasting and an RMSE of 18.822 and an R 2 of 0.522 during real estate price prediction. Future research will explore the integration of the ISSA with alternative neural architectures such as GRUs and Transformers to assess its flexibility and effectiveness across different sequence modeling paradigms.

1. Introduction

Optimization problems have long been a central research focus, and they arise in diverse practical domains, such as image classification, feature selection, autonomous driving, medical diagnosis, and energy management [1,2,3,4,5,6,7,8,9]. Solving complex optimization problems requires analyzing constraints and objectives and optimizing the objective function [10]. These problems are challenging due to multimodality, high dimensionality, and dynamic constraints. Traditional methods like gradient descent and Newton’s method rely on derivatives, are computationally expensive, and often get trapped in local optima, limiting their practical use. With the advancement of research, scholars have developed metaheuristic algorithms, which are inspired by intuitive experience or biological behavior patterns [11]. Metaheuristic optimization algorithms can be applied to optimize both symmetric and asymmetric engineering problems and are characterized by high computational efficiency, ease of implementation, flexibility, and simple frameworks. In recent years, a series of metaheuristics have been proposed, including particle swarm optimization (PSO) [12], ant colony optimization (ACO) [13], the whale optimization algorithm (WOA) [14], the grey wolf optimizer (GWO) [15], Harris hawk optimization (HHO) [16], differential evolution (DE) [17], the bat algorithm (BA) [18], ship rescue optimization (SRO) [19], the Rafflesia optimization algorithm [20], and the goose optimization algorithm (GOOSE) [21]. For example, PSO emulates bird flocking behavior, using particle velocity, position, and individual/global extrema to conduct a collaborative search in a solution space. The WOA innovatively incorporates a bubble-net hunting mechanism, where random or optimal search agents simulate humpback whale behavior to navigate solution spaces. GOOSE leverages the early-warning behavior and balance mechanisms of goose flocks, adaptively adjusting the search resolution and speed to enable precise exploration of multimodal solution spaces.
The sparrow search algorithm (SSA), proposed by Xue J and Shen B in 2020 [22], is a metaheuristic algorithm inspired by the foraging and anti-predation behaviors of sparrow populations. The SSA is characterized by a simple structure, few parameters, and high scalability. However, like other metaheuristic algorithms, it is prone to local optima and suffers from suboptimal optimization accuracy. Several improvements have been proposed to enhance the SSA’s performance. For example, Ouyang et al. [23] introduced a dynamic population partitioning method based on K-medoids clustering to balance local exploitation and global exploration. Su et al. [24] developed an adaptive hybrid SSA framework that incorporates differential evolution operators to improve population diversity and escape local optima. Yan et al. [25] proposed a multi-population collaborative search model based on K-means clustering, which avoids premature convergence while preserving global search capabilities. Huang et al. [26] suggested a chaotic optimization strategy using elite solution sets to guide population evolution. The effectiveness of these improvements has been demonstrated through benchmark testing, showing that the improved SSA algorithms outperform traditional methods in optimization tasks. Based on previous research, this paper proposes an improved sparrow search algorithm (ISSA). This algorithm generates the initial population using a reverse learning strategy to increase diversity, modifies the position update formula to prevent convergence to the origin, dynamically adjusts the step size to improve accuracy, and incorporates Lévy flight for enhanced global search capabilities. Simulations show that ISSA-LSTM outperforms other algorithms on the CEC2005 benchmark functions.
Multiple regression models analyze the relationships between multiple variables and a single response, providing better explanatory power than univariate methods by incorporating multiple factors [27,28]. This is particularly useful in time-series tasks, such as power load forecasting and real estate price prediction, where long-term dependencies and multi-factor interactions dominate the data dynamics [29,30]. However, traditional regression methods struggle with high-dimensional and non-stationary data. In contrast [31], deep learning models, such as LSTM networks, have revolutionized prediction tasks by autonomously extracting features from complex datasets. These models outperform traditional machine learning approaches, which rely on manual feature selection [32,33,34]. Recent research has combined metaheuristic optimization with LSTM networks to further enhance prediction accuracy. For instance, Sun et al. [35] proposed a PSO-LSTM hybrid to optimize hyperparameters, improving failure prediction accuracy. Lu et al. [36] demonstrated that differential evolution-optimized LSTMs outperform classical models in electricity forecasting. Pal et al. [37] showed that Bayesian-optimized shallow LSTMs achieve greater long-term prediction accuracy than deeper models. Furthermore, Cao et al. [38] integrated K-means clustering with IPSO to improve weather-resilient predictions. These advancements illustrate the potential of combining heuristic algorithms with LSTM networks to enhance predictive analytics.
This study introduces an ISSA to optimize LSTM hyperparameters for time-series prediction tasks, including hidden layer nodes, training epochs, and the learning rate, significantly improving prediction accuracy and reliability. Validation on two publicly available datasets—power load forecasting and real estate price prediction—shows that ISSA-LSTM outperforms other LSTM models optimized by alternative methods, with consistently lower error deviations. These results highlight the combined strength of the ISSA’s global search capabilities and LSTM’s sequential modeling.
The structure of this paper is as follows: Section 2 introduces the operational paradigms of the SSA and LSTM networks, providing the theoretical foundations for subsequent developments. Section 3 introduces the ISSA model, detailing its enhanced convergence mechanisms through a mathematical derivation. Section 4 validates the optimization superiority of the ISSA by benchmarking it against eight established metaheuristic algorithms using 10 test functions. In Section 5, an integrated ISSA-LSTM architecture is proposed and its performance is demonstrated through multivariate regression analysis on two publicly available datasets, outperforming eight state-of-the-art models. The conclusion summarizes the main contributions of this research and outlines three key research directions for advancing evolutionary algorithms.

2. Related Works

2.1. Sparrow Search Algorithm

The sparrow search algorithm (SSA) simulates the foraging behavior of sparrows to achieve population-based optimization through collaborative mechanisms. Its core principle involves dynamically dividing the population into discoverers and followers based on mutual observation and information exchange. Discoverers are responsible for locating high-quality solutions and guiding the swarm’s movement, while followers adjust their trajectories by observing the positions of the discoverers. Additionally, some individuals compete for better positions through a competitive mechanism, enhancing the overall allocation of resources. Compared to traditional swarm intelligence algorithms, the SSA exhibits stronger global search capabilities and faster convergence, making it well-suited for solving complex optimization problems.
The mathematical formulation of the SSA consists of solution vectors and corresponding fitness values, which are defined as follows:
X = x 1 , 1 x 1 , 2 x 1 , d x 2 , 1 x 2 , 2 x 2 , d x N , 1 x N , 2 x N , d
F x = f ( [ x 1 , 1 , x 1 , 2 , , x 1 , d ] ) f ( [ x 2 , 1 , x 2 , 2 , , x 2 , d ] ) f ( [ x N , 1 , x N , 2 , , x N , d ] )
Here, X , N , and d represent the individual position matrix, the number of individuals in the population, and the dimension of the population in the SSA, respectively. x i , j is the position of the i-th individual in the j-th dimension. f is the fitness function of the SSA, and the fitness matrix ( F x ) is an N × 1 matrix.
As the population evolves, discoverer individuals with higher fitness guide the global search by exploring promising regions, while followers perform localized refinements within these areas. The SSA achieves behavioral coordination by dividing individuals into three roles, each governed by distinct position update strategies.
The discoverers have higher energy. With the progress of each iteration, the formula for updating their positions is as follows:
X i , j t + 1 = X i , j t exp ( i α t max ) , R 2 < ST   X i , j t + Q L , R 2 ST
In this equation, X i , j t + 1 represents the spatial coordinate of the i t h sparrow in the j t h dimension at the (t + 1)-th iteration; α [ 0 , 1 ] is a stochastic decay coefficient modulating the exploration rate; t max denotes the termination criterion for the evolutionary iterations; S T and R 2 denote the safety threshold and early-warning value, with their value ranges being [0.5, 1] and (0, 1], respectively. The term Q N represents a normally distributed random variable bounded within [−1, 1], and L is a 1 × d matrix.
Followers observe the movements of discoverers to identify promising directions. Once a discoverer is found to occupy a region with higher fitness, followers adjust their positions accordingly and engage in competitive replacement. Successful followers replace inferior individuals; otherwise, they continue to monitor other discoverers. The follower update mechanism is mathematically described as follows:
X i , j t + 1 = Q exp ( X w o r s t t X i , j t i 2 ) , i > n 2 X p t + 1 + X i , j t + 1 X p t + 1 + L A + , i n 2
A + = A T ( A A T ) 1
In this formula, X i , j t + 1 represents the spatial coordinate of the i t h sparrow in the j t h dimension at the (t + 1)-th iteration. X w o r s t t denotes the least-fit agent’s spatial configuration at iteration t; i indexes the fitness hierarchy among the population members; and X p t + 1 is the position of the individual with the best fitness at the t + 1 -th iteration. A constitutes a binary coefficient matrix, and the instances of A take values of either 1 or −1. The term Q N represents a normally distributed random variable bounded within [−1, 1], and L is a 1 × d matrix.
The pseudoinverse A + = A T ( A A T ) 1 is used to calculate an update direction. To ensure the mathematical validity of this operation, the matrix A R d × d is constructed as a full-rank binary matrix, with each element randomly assigned within {1, −1}. In this setting, A A T is symmetric and nonsingular with high probability, thus guaranteeing the existence of the inverse and the correctness of the pseudoinverse expression.
X i , j t + 1 = X b e s t t + β X i , j i X b e s t t , f i > f g X i , j t + K X i , j t + 1 X w o r s t t + 1 f i f w + ε ,   f i = f g  
In this formula, X i , j t + 1 represents the spatial coordinate of the i t h sparrow in the j t h dimension at the (t + 1)-th iteration. X b e s t t encodes the elite agent’s optimal spatial state, where β and K u ( 0 , 1 ) are stochastic scaling factors. ε is negligible and has a constant value; its purpose is to prevent division by zero. f i denotes the fitness of the present individual, while f g (global best) and f w (global worst) are designations used to denote the maximum and minimum levels of fitness that can be achieved, respectively, by any given individual.

2.2. Long Short-Term Memory Neural Network

As an improved architecture of recurrent neural networks (RNNs), the LSTM model effectively mitigates the challenge of learning long-term dependencies in time-series prediction. This is achieved through a gating mechanism composed of input, forget, and output gates, which enables the selective retention and removal of historical information [39]. Compared to standard RNNs, LSTM significantly reduces the problems of gradient vanishing and explosion. The structure of the LSTM network is illustrated in Figure 1.
(1) The forget gate ( f t ) controls which historical states in C t 1 are discarded and transfers the remaining information to the current memory cell ( C t ):
f t = σ ( W f [ h t 1 , x t ] + b f )
In this formula, h t 1 denotes the hidden state at time t − 1 and x t is the input vector at t. W f (weight) and b f (bias) are critical parameters defining the forget gate. The value of f t [ 0 , 1 ] is generated by the sigmoid function. When f t = 0 , all previous states in C t 1 are erased; if f t = 1 , C t 1 is fully preserved.
(2) The input gate ( i t ) determines how much new information from x t is stored in C t :
i t = σ ( W i [ h t 1 , x t ] + b i )
Similar to the forget gate, h t 1 represents the hidden state at the previous time step, and x t is the input vector at the current time step. W i (weight) and b i (bias) are specific to the input gate. The value calculated through the sigmoid activation function ( σ ) is within the interval [0, 1]. When i t = 0 , the gate blocks all input, while when i t = 1 it allows complete storage.
(3) The candidate state ( C ˜ t ) is computed from C t and serves as an intermediate step to update C t 1 .
C ˜ t = tanh ( W c [ h t 1 , x t ] + b C )
Here, W c and b c are the weight matrix and bias parameter of the candidate state ( C ˜ t ), respectively. The value of C ˜ t calculated through the hyperbolic tangent activation function ( tanh ) is within the interval [−1, 1].
(4) Memory Cell Update
The memory cell update rule integrates historical and novel states:
C t = f t C t 1 + i t C ˜ t
The terms f t , i t , and C ˜ t are derived from Equations (7)–(9).
(5) The output gate ( o t ) regulates how much of C ˜ t is propagated to h t at the same time step, and it outputs the filtered information.
o t = σ ( W o [ h t 1 , x t ] + b o )
C t = f t C t 1 + i t C ˜ t
The gate parameters are defined by the weight matrix ( W o ) and the bias vector ( b o ), and o t [ 0 , 1 ] through sigmoidal activation. Equation (12) updates the memory cell by blending C t 1 and C ˜ t .

3. Improved Sparrow Search Algorithm Strategy

3.1. Oppositional Learning Strategy

Oppositional learning [40] enhances population diversity by performing mutation operations on the population after each iteration.
X i = 1 × ( X max + X min ) X i
In this formula, X max and X min denote the upper and lower bounds, respectively. In the i-th iteration, X i represents the individual, whereas X ¯ i signifies the generated adversarial individual. The population size is expanded to 2 M candidate solutions by merging the original individual ( X i ) with the adversarial learning-generated individual ( X ¯ i ). The optimal M solutions are filtered based on fitness ranking to update the location, and the iterative loop continues to optimize the quality of the population.

3.2. Producer Anti-Origin Convergence Strategy

The performance of the SSA was evaluated in Reference [22] through a series of benchmark experiments. The results indicated that when the early-warning variable (R2) exceeds the safety threshold (ST), discoverers update their positions toward the current global optimum using random perturbations following a normal distribution. Conversely, if R2 < ST, the position update mechanism drives individuals to converge toward the origin, which limits the algorithm’s ability to handle optimization tasks with distant optima. As a result, while the SSA performs well on problems with solutions near the origin, it often exhibits insufficient accuracy when addressing complex landscapes where the global optimum is far away. To address this limitation, this study introduces an improved position update rule for discoverers, as shown in Equation (14). The proposed strategy mitigates premature convergence to the origin and enhances the algorithm’s global search capability across diverse optimization scenarios:
X i , j t + 1 = X i , j t ( 1 + Q ) , R 2 < ST   X i , j t + Q L , R 2 ST
In this equation, X i , j t + 1 represents the spatial coordinate of the i t h sparrow in the j t h dimension at the (t + 1)-th iteration; the term Q N represents a normally distributed random variable bounded within [−1, 1]; and L is a 1 × d matrix.

3.3. Lévy Flight Strategy

Lévy flight [41], a non-Gaussian stochastic process with heavy-tailed step-size distributions, exhibits the dual characteristics of frequent small steps for local exploration and occasional large jumps to search globally [42]. The standard SSA tends to fall into local optima, limiting its effectiveness in global optimization tasks. To address this issue, Lévy flight is integrated into the position update mechanism (Equation (6)), enabling a balance between short-range exploitation and long-range exploration. As the algorithm nears the global optimum, Lévy steps enhance fine-grained search capabilities while preventing premature convergence. By utilizing the displacement between the current and best-known solutions, the improved SSA maintains robust local search performance while expanding its global exploration potential. This enhanced update rule is formalized in Equation (15):
X i , j t + 1 = L e v y ( d ) · X b e s t t + β X i , j i L e v y ( d ) X b e s t t , f i > f g X i , j t + K X i , j t + 1 X w o r s t t + 1 f i f w + ε ,   f i = f g  
In this formula, d represents the spatial dimension. The Lévy calculation formulas are shown in Equations (16) and (17):
L e v y ( d ) = 0.01 r 1 σ r 2 1 β
σ = Γ ( 1 + θ ) sin ( π θ / 2 ) Γ [ ( 1 + λ ) / 2 ] θ 2 ( 1 θ ) / 2 1 / θ
In these formulas, Γ represents the gamma function, r 1 and r 2 are random numbers within the interval [0, 1], and θ is a constant.

3.4. Step-Size Factor Adjustment Strategy

In the standard SSA, the step-length control parameters (β and K in Equation (6)) play a crucial role in balancing global exploration and local exploitation. However, due to their inherently random nature, the search process often suffers from poor utilization of the solution space and a tendency to fall into local optima. To overcome this issue, this study proposes an improved strategy where β and K are adaptively adjusted: larger values promote a broader global search, while smaller values support precise local refinement [43]. The modified formulations for β and K are detailed in Equations (18) and (19), respectively:
β = f i t n e s s b e s t ( f i t n e s s b e s t f i t n e s s w o r s t ) · ( T t T ) 1.5
K = ( f i t n e s s b e s t f i t n e s s w o r s t ) · exp ( 20 · tan ( t T ) 2 ) · ( 2 · r a n d 1 )
In these formulas, f i t n e s s b e s t denotes the best fitness value and f i t n e s s w o r s t is used to denote the worst fitness value, while T is used to denote the maximum number of iterations. Equation (18) shows that the improved step-length control parameters exhibit a nonlinear increasing trend. In early SSA iterations, high population diversity implies strong global exploration but weak local exploitation. Therefore, we set the control parameters to smaller values to enhance local exploitation. In later iterations, as sparrows converge to the current global optimum, the risk of premature convergence increases as the search space narrows. Larger control parameter values help avoid sparrows becoming trapped in local optima and facilitate further exploration. Equation (19) reveals biphasic modulation of the step-length factor (K): an ascending phase in the initial iterations promotes exhaustive search space exploration, while subsequent rapid attenuation expedites convergence. This dynamic adjustment of the step-length parameters balances the SSA’s global exploration and local exploitation, improving optimization accuracy while avoiding local optima.

3.5. The Pseudo-Code of the ISSA

To illustrate the improved sparrow search algorithm (ISSA), the pseudo-code is presented in Algorithm 1.
Algorithm 1. The pseudo-code of the ISSA.
Input: The population size (n), the number of discoverers (PD), the number of followers (SD), the early-warning value (R2), the safety threshold (ST), and the maximum number of iterations (T).
Output: The location of a sparrow and its fitness value.
1: Initialize the population using the oppositional learning strategy.
2: Calculate the fitness value of each sparrow and determine the optimal and worst individuals in the current population.
3: While(t < T)
4:    for i = 1:PD
5:          Update the position of the i-th sparrow according to Equation (14) (update the positions of the discoverers)
6:    end for
7:    for i = PD + 1:n
8:          Update the position of the i-th sparrow according to Equation (4) (update the positions of the followers)
9:    end for
10:    for l = 1:SD
11:          Update the position of the l-th sparrow according to Equations (15)–(18) (update the positions of the scouts);
12:    end for
13:    Obtain the updated positions of all sparrows;
14:    If the fitness value corresponding to the new updated position is higher than that of the current position, update the current position;
15:    t = t + 1;
16: end while
17: Return the objective function value.

4. Case Study

4.1. Parameter Settings

Computational experiments were executed on a workstation with a 2.40 GHz Intel i5-1135G7 processor and 16 GB of memory running 64-bit Windows 10 Pro. The ISSA framework was developed using MATLAB R2022a.
To evaluate the effectiveness of the ISSA in solving complex optimization problems, ten benchmark functions from the CEC 2005 test suite were selected for simulation experiments. As detailed in Table 1, these functions are 30-dimensional and fall into three categories: unimodal (F1–F3), multimodal (F4–F7), and composite (F8–F10). Unimodal functions are used to assess the convergence rates of algorithms, multimodal functions test their robustness in navigating local optima, and composite functions examine the balance between exploration and exploitation in intricate problem landscapes.

4.2. Comparative Algorithm Analysis

The ISSA’s performance, evaluated through 30 independent trials per benchmark function, was benchmarked against the SSA, SSA1 (the SSA optimized with Lévy flight), PSO, the GWO, DE, a GA, and GOOSE. All algorithms were initialized with identical configurations: T = 500 (maximum of 500 iterations), N = 30 (30 population members), and dimensionality (d = 30). Three evaluation metrics—the optimal value, mean value, and standard deviation—were derived from the 10 benchmark functions. Given the shared population-based iterative framework, the theoretical time complexity of the ISSA, the SSA, and PSO can be uniformly expressed as O(N × D × T), where N = 30, d = 30, and T = 500. Although the ISSA incorporates enhancements such as reverse learning and Lévy flight, these operations only incur constant-time overhead and do not alter the asymptotic complexity. The simulation results are shown in Table 2.
Table 2 demonstrates that the ISSA, SSA1, and the SSA yielded the best results in solving the unimodal functions, F1, F2, and F3, outperforming traditional algorithms such as PSO, the GWO, DE, the GA, and GOOSE in terms of optimization accuracy. Notably, the ISSA achieved a mean and a standard deviation of zero, reflecting its exceptional consistency and robustness compared to the other methods.
In the case of the multimodal functions, the ISSA surpassed the SSA and SSA1 in optimization accuracy, achieving the best results for F4, F6, and F7. For F5, the ISSA yielded the solution closest to the theoretical optimum (0) among all tested algorithms. In terms of average performance, the ISSA demonstrated the highest accuracy, while the SSA ranked second. Regarding the standard deviation, the ISSA achieved superior consistency across most functions, except for F6 and F7.
Regarding the composite functions, F8–F10, all algorithms—except PSO, the GWO, DE, the GA, and GOOSE—approximated the optimal values to varying degrees. Among them, the ISSA achieved the lowest mean and standard deviation, signifying superior global search performance, robustness, and stability. These results confirm the ISSA’s competitive advantage across all three categories: unimodal, multimodal, and composite functions.
The convergence trajectories of the ISSA and the baseline algorithms for different benchmark functions are visualized in Figure 2. The x-axis represents the iteration number, while the y-axis denotes the best fitness value obtained. The figure clearly illustrates the ISSA’s faster convergence toward the global optimum.
Figure 2 illustrates the convergence trajectories for three function types: unimodal (F1–F3), multimodal (F4–F7), and composite (F8–F10).
Unimodal functions (F1–F3): The ISSA, SSA1, and the SSA showed superior global exploration performance compared to PSO, the GWO, DE, the GA, and GOOSE. Notably, the ISSA achieved rapid early-stage convergence, reaching near-optimal solutions faster, while PSO and the GWO converged more slowly in the initial iterations.
Multimodal functions (F4–F7): The ISSA excelled at avoiding local optima, benefiting from a robust global search mechanism. Its ability to balance exploration and exploitation effectively prevented early stagnation while maintaining solution accuracy.
Composite functions (F8–F10): The ISSA gradually converged toward superior solutions. Specifically, for F10, although PSO performed well initially, the ISSA demonstrated consistent convergence across the iterations and reliably approached the global optimum.
In summary, the ISSA achieves rapid convergence in unimodal tasks through efficient search dynamics. In multimodal problems, it reliably escapes local optima and locates the global optimum with high accuracy. In composite scenarios, the ISSA ensures stable convergence across iterations, showcasing strong adaptability and robustness in complex optimization settings. These strengths make the ISSA a valuable tool for solving high-dimensional predictive and optimization problems.

5. Multiple Regression Prediction Based on ISSA-LSTM

Multiple regression prediction plays a vital role across various domains by integrating multiple influencing factors to forecast outcomes. Optimization of LSTM neural networks via ISSA enhances prediction accuracy and reliability. However, as the number of influencing factors increases, discrepancies between predicted and actual results tend to grow. Therefore, improving the precision of multiple regression prediction has become a significant research focus in recent years.
In this section, the predictive performance of the ISSA-LSTM framework is comparatively evaluated against benchmark models. Empirical results demonstrate its superior accuracy in multivariate regression tasks.

5.1. Dataset Information

This study employed two diverse datasets from Kaggle, representing typical application domains for multivariate regression: an electricity load dataset and a real estate housing price dataset. The electricity load data comprised 625 hourly consumption records collected in June 2020 across three Indian cities (Mumbai, Delhi, and Chennai), featuring 12 variables, including temperature, humidity, and wind speed and direction, as well as temporal indicators like the date and time. Temporal features such as the time of day, the day of the week, and lagged load values were also extracted. Missing values were addressed using linear interpolation, and Min–Max normalization was applied.
The real estate dataset contained 1460 housing transaction records, each characterized by 12 attributes such as the number of bedrooms, the number of bathrooms, the living area, the year the house was built, and location-based parameters. House prices were normalized into units of thousands to facilitate interpretation.
The selection of these datasets ensured a comprehensive evaluation of the ISSA-LSTM model’s adaptability, robustness, and generalization in scenarios with both structured and sequential data.

5.2. ISSA-LSTM Load Forecasting Model

The initial hyperparameter settings of LSTM neural networks often result in suboptimal prediction performance [44]. To overcome this, the ISSA was introduced to dynamically optimize three critical hyperparameters: the number of hidden layer nodes, the number of training epochs, and the learning rate. Specifically, the learning rate controls convergence behavior—a rate that is too low slows training, while a rate that is too high risks entrapment in local optima. Meanwhile, the complexity of the hidden layer and the training duration directly impact predictive accuracy [45]. The ISSA-LSTM model was implemented as follows:
Step 1: Preprocess the data to define the inputs and outputs for the ISSA-LSTM regression model. Feature parameters are extracted from the housing price and electric load datasets as inputs, with corresponding target values as outputs. Due to differences in dimensionality and scale, normalization is applied before splitting the dataset into training and testing subsets at a 4:1 ratio.
Step 2: Configure ISSA parameters including the dimensionality, population size, maximum iterations, and early-warning thresholds. Initialize the sparrow population and establish the LSTM regression model along with the hyperparameter search ranges.
Step 3: Employ the ISSA to iteratively optimize the three hyperparameters: the learning rate, the number of hidden nodes, and the number of epochs. During optimization, experimental data train the model, and the test set RMSE is used as the objective metric. After T iterations, the algorithm converges to an optimal hyperparameter set that minimizes prediction error while balancing the computational cost.
Step 4: Deploy the optimized ISSA-LSTM model for multivariate regression prediction. Validate its performance by comparing the predicted outputs with the ground-truth values of the test set and quantify its accuracy through statistical error metrics to demonstrate improved predictive power and robustness.

5.3. Model Parameter Settings and Evaluation Indicators

This study employed a single-layer LSTM network architecture and selected “Adam” as the optimizer. The gradient clipping threshold was set to 1.0 to avert gradient explosion. All remaining network parameters adopted default configurations, with only the target hyperparameters undergoing optimization.
The basic parameter configurations for each algorithm in the LSTM neural network optimization are shown in Table 3. To ensure experimental fairness, key parameters were kept consistent across the algorithms. Based on the proposed ISSA-LSTM model and the comparative models, multiple regression prediction experiments were conducted on the two datasets described earlier.
The performance assessment metrics employed in this study were as follows:
RMSE quantifies the quadratic mean deviation between model predictions and ground-truth observations:
R M S E = i = n ( y ^ i y i ) 2
The Mean Absolute Error (MAE) is mathematically defined as the average of absolute differences between predicted values and corresponding actual measurements:
M A E = i = 1 n ( y ^ i y i ) n
For the Coefficient of Determination ( R 2 ), the closer its value is to one, the more accurate the results:
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2

5.4. Experimental Results and Analysis of Multiple Regression Prediction

This study employed the ISSA alongside other optimization algorithms to identify the optimal hyperparameter configurations for LSTM prediction models. Experimental validation was performed on models optimized by each algorithm using two datasets.
Figure 3 and Figure 4 illustrate the prediction curves of LSTM models optimized by different algorithms for the electric load and real estate price datasets, respectively. The ISSA-LSTM model demonstrated the closest fit to the actual data, exhibiting minimal deviation in both cases. Table 4 and Table 5 further confirm that ISSA-LSTM achieved the lowest prediction errors across all evaluated metrics.
Building on the critical role of hyperparameter optimization, comparative experiments revealed the superior fitting performance of the ISSA-LSTM model. Table 5 highlights that ISSA-LSTM attained the smallest average prediction errors, with significantly better MAE and RMSE values compared to SSA-LSTM, GWO-LSTM, and other models. Its forecasting deviation was also notably lower than that of the PGOA-LSTM framework [46], underscoring its enhanced prediction accuracy.
Table 5 reinforces these findings, showing that ISSA-LSTM consistently achieved the highest accuracy and the lowest MAE/RMSE values, confirming its robustness and effectiveness.
In summary, the ISSA-LSTM model exhibits outstanding prediction accuracy and a strong generalization ability in electric load and real estate price forecasting, offering a reliable solution for time-series prediction in these domains.

6. Conclusions

This study presented an ISSA framework and assessed its optimization performance against several metaheuristic algorithms through extensive benchmarking on 10 standard test functions. The results show that the ISSA achieved faster convergence on unimodal functions by swiftly approaching the optimal solution. For multimodal functions, it effectively balanced global exploration and local exploitation, improving the search for optimal solutions and reducing the risk of local optima. In composite function optimization, the ISSA demonstrated high stability and efficiency, consistently converging to optimal solutions across multiple trials. By integrating the ISSA with the LSTM neural network, we developed the ISSA-LSTM model for multivariate regression prediction, where the ISSA adaptively optimizes three critical hyperparameters: the number of training epochs, the hidden layer size, and the learning rate. Experimental validations on two datasets revealed that ISSA-LSTM significantly outperformed competing models in prediction accuracy, confirming its effectiveness and superiority in multivariate regression tasks. The model’s enhanced predictive performance arises from the synergy between the ISSA’s strong global optimization capabilities and LSTM’s proficiency in modeling sequential data. The ISSA improves hyperparameter search through reverse learning-based population initialization, adaptive step-size control, and Lévy flight, enabling broader exploration of the parameter space and mitigating premature convergence. This leads to parameter configurations better aligned with data characteristics, allowing the LSTM to capture nonlinear relationships and long-term dependencies more effectively, thereby boosting generalization and prediction stability.
Despite promising results, there remains potential for improvement. This work focused on optimizing three hyperparameters: the number of hidden layer nodes, the number of training epochs, and the learning rate. Nonetheless, other aspects such as the number of layers, the connectivity patterns, and the forget gate thresholds also influence accuracy. Future research will employ the ISSA to further optimize these elements and comprehensively enhance the LSTM model.

Author Contributions

Conceptualization, J.-S.P. and S.-C.C.; Data curation, Y.-M.Z.; Formal analysis, Y.-M.Z.; Investigation, Y.-M.Z. and A.R.Y.; Methodology, Y.-M.Z. and J.-S.P.; Resources, Y.-M.Z. and A.R.Y.; Software, Y.-M.Z.; Validation, Y.-M.Z. and A.R.Y.; Writing—original draft, Y.-M.Z.; Writing—review and editing, J.-S.P. and S.-C.C.; Supervision, J.-S.P. and S.-C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data are contained within this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bharanidharan, N.; Rajaguru, H. Improved chicken swarm optimization to classify dementia MRI images using a novel controlled randomness optimization algorithm. Int. J. Imaging Syst. Technol. 2020, 30, 605–620. [Google Scholar] [CrossRef]
  2. Hu, H.; Shan, W.; Chen, J.; Xing, L.; Heidari, A.A.; Chen, H.; He, X.; Wang, M. Dynamic individual selection and crossover boosted forensic-based investigation algorithm for global optimization and feature selection. J. Bionic Eng. 2023, 20, 2416–2442. [Google Scholar] [CrossRef]
  3. Shalamov, V.; Filchenkov, A.; Shalyto, A. Heuristic and metaheuristic solutions of pickup and delivery problem for self-driving taxi routing. Evol. Syst. 2019, 10, 3–11. [Google Scholar] [CrossRef]
  4. Baghdadi, N.A.; Malki, A.; Abdelaliem, S.F.; Balaha, H.M.; Badawy, M.; Elhosseini, M. An automated diagnosis and classification of COVID-19 from chest CT images using a transfer learning—Based convolutional neural network. Comput. Biol. Med. 2022, 144, 105383. [Google Scholar] [CrossRef] [PubMed]
  5. Daniel, E.; Anitha, J.; Kamaleshwaran, K.K.; Rani, I. Optimum spectrum mask based medical image fusion using Gray Wolf Optimization. Biomed. Signal Process. Control 2017, 34, 36–43. [Google Scholar] [CrossRef]
  6. Chopra, N.; Ansari, M.M. Golden jackal optimization: A novel nature-inspired optimizer for engineering applications. Expert Syst. Appl. 2022, 198, 116924. [Google Scholar] [CrossRef]
  7. Lorestani, A.; Gharehpetian, G.B.; Nazari, M.H. Optimal sizing and techno-economic analysis of energy-and cost-efficient standalone multi-carrier microgrid. Energy 2019, 178, 751–764. [Google Scholar] [CrossRef]
  8. Ma, G.; Wang, Z.; Liu, W.; Fang, J.; Zhang, Y.; Ding, H.; Yuan, Y. Estimating the state of health for lithium-ion batteries: A particle swarm optimization-assisted deep domain adaptation approach. IEEE/CAA J. Autom. Sin. 2023, 1, 1530–1543. [Google Scholar] [CrossRef]
  9. Mohanty, S.; Subudhi, B.; Ray, P.K. A grey wolf-assisted perturb & observe MPPT algorithm for a PV system. IEEE Trans. Energy Convers. 2017, 32, 340–347. [Google Scholar]
  10. Singh, G.; Singh, A. A hybrid algorithm using particle swarm optimization for solving transportation problem. Neural Comput. Appl. 2020, 32, 11699–11716. [Google Scholar] [CrossRef]
  11. Bredael, D.; Vanhoucke, M. A genetic algorithm with resource buffers for the resource constrained multi-project scheduling problem. Eur. J. Oper. Res. 2024, 315, 19–34. [Google Scholar] [CrossRef]
  12. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  13. Dorigo, M.; Birattari, M.; Stutzle, T. Ant colony optimization. IEEE Comput. Intell. Mag. 2007, 1, 28–39. [Google Scholar] [CrossRef]
  14. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  15. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  16. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  17. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  18. Yang, X.S.; He, X. Bat algorithm: Literature review and applications. Int. J. Bio-Inspired Comput. 2013, 5, 141–149. [Google Scholar] [CrossRef]
  19. Chu, S.C.; Wang, T.T.; Yildiz, A.R.; Pan, J.S. Ship rescue optimization: A new metaheuristic algorithm for solving engineering problems. J. Internet Technol. 2024, 25, 61–78. [Google Scholar]
  20. Pan, J.S.; Fu, Z.; Hu, C.C.; Tsai, P.W.; Chu, S.C. Rafflesia optimization algorithm applied in the logistics distribution centers location problem. J. Internet Technol. 2022, 23, 1541–1555. [Google Scholar]
  21. Hamad, R.K.; Rashid, T.A. GOOSE algorithm: A powerful optimization tool for real-world engineering challenges and beyond. Evol. Syst. 2024, 15, 1249–1274. [Google Scholar] [CrossRef]
  22. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  23. Ouyang, C.T.; Zhu, D.L.; Qiu, Y.X. Improved sparrow search algorithm based on clustering algorithm. Comput. Simul. 2022, 39, 392–397. [Google Scholar]
  24. Su, Y.Y.; Wang, S.X. Adaptive hybrid strategy sparrow search algorithm. Comput. Eng. Appl. 2023, 59, 75–85. [Google Scholar]
  25. Yan, S.; Liu, W.; Yang, P.; Wu, F.; Yan, Z. Multi group sparrow search algorithm based on k-means clustering. J. Beijing Univ. Aeronaut. Astronaut. 2022, 50, 508–518. [Google Scholar]
  26. Huang, H.X.; Zhang, G.Y.; Chen, S.Y. WOA based on chaotic weight and elite guidance. Transducer Microsyst. Technol. 2020, 39, 113–116. [Google Scholar]
  27. Licht, M.H. Multiple regression and correlation. In Reading and Understanding Multivariate Statistics; Grimm, L.G., Yarnold, P.R., Eds.; American Psychological Association: Washington, DC, USA, 1995; pp. 19–64. [Google Scholar]
  28. Knofczynski, G.T.; Mundfrom, D. Sample sizes when using multiple linear regression for prediction. Educ. Psychol. Meas. 2008, 68, 431–442. [Google Scholar] [CrossRef]
  29. Zhang, D.; Shen, D. Alzheimer’s Disease Neuroimaging Initiative, Multi-modal multi-task learning for joint prediction of multiple regression and classification variables in Alzheimer’s disease. NeuroImage 2012, 59, 895–907. [Google Scholar] [CrossRef]
  30. Heim, C.; Newport, D.J.; Wagner, D.; Wilcox, M.M.; Miller, A.H.; Nemeroff, C.B. The role of early adverse experience and adulthood stress in the prediction of neuroendocrine stress reactivity in women: A multiple regression analysis. Depress. Anxiety 2002, 15, 117–125. [Google Scholar] [CrossRef]
  31. Chu, C.W.; Zhang, G.P. A comparative study of linear and nonlinear models for aggregate retail sales forecasting. Int. J. Prod. Econ. 2003, 86, 217–231. [Google Scholar] [CrossRef]
  32. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tian, J.H.; Adam, M.; Gertych, A.; Tan, R.S. A deep convolutional neural network model to classify heartbeats. Comput. Biol. Med. 2017, 89, 389–396. [Google Scholar] [CrossRef]
  33. Xue, Y.; Onzo, B.M.; Mansour, R.F.; Su, S. Deep Convolutional Neural Network Approach for COVID-19 Detection. Comput. Syst. Sci. Eng. 2022, 42, 201–211. [Google Scholar] [CrossRef]
  34. Kim, T.Y.; Cho, S.B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  35. Sun, Y.G.; Liu, K.J. Failure rate of avionics system forecasting based on PSO-LSTM network. Aeronaut. Sci. Technol. 2021, 32, 17–22. [Google Scholar]
  36. Peng, L.; Liu, S.; Liu, R.; Wang, L. Effective long short-term memory with differential evolution algorithm for electricity price prediction. Energy 2018, 162, 1301–1314. [Google Scholar] [CrossRef]
  37. Pal, R.; Sekh, A.A.; Kar, S.; Prasad, D.K. Neural network based country wise risk prediction of COVID-19. Appl. Sci. 2020, 10, 6448. [Google Scholar] [CrossRef]
  38. Yan, B.; Peng, H.; Fu, J.; Wang, D.; Lu, H. Learning Spatio-emporal Transformer for Visual Tracking. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 10448–10457. [Google Scholar]
  39. Hochreter, S.; Schmidhuber, R. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  40. Tizhoosh, H.R. Opposition-based learning: A new scheme for machine intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation, and International Conference on Intelligent Agents, Web Technologies and Internet Commerce, Vienna, Austria, 28–30 November 2005; pp. 695–701. [Google Scholar]
  41. Reynolds, A. Liberating Lévy walk research from the shackles of optimal foraging. Phys. Life Rev. 2015, 14, 59–83. [Google Scholar] [CrossRef]
  42. Iacca, G.; Junior, V.C.D.S.; Melo, V.V.D. An improved Jaya optimization algorithm with Lévy flight. Expert Syst. Appl. 2020, 165, 113902. [Google Scholar] [CrossRef]
  43. Gao, L.F.; Rong, X.J. Improved YSGA algorithm combining declining strategy and Fuch chaotic mechanism. J. Front. Comput. Sci. Technol. 2021, 15, 564–576. [Google Scholar]
  44. Ordieres, J.; Vergara, E.; Capuz, R.; Salazar, R. Neural network prediction model for fine particulate matter (PM2. 5) on the US–Mexico border in El Paso (Texas) and Ciudad Juárez (Chihuahua). Environ. Model. Softw. 2005, 20, 547–559. [Google Scholar] [CrossRef]
  45. Lin, S.L.; Huang, C.F.; Liou, M.H.; Chen, C.Y. Improving Histogram-based Reversible Information Hiding by an Optimal Weight-based Prediction Scheme. J. Inf. Hiding Multimed. Signal Process. 2013, 4, 19–33. [Google Scholar]
  46. Pan, J.-S.; Wang, W.; Chu, S.-C.; Shao, Z.-Y.; Yang, H.-M. LSTM Neural Network with Parallel Swarm Optimization Algorithm for Multiple Regression Prediction. J. Internet Technol. 2024, 25, 1035–1049. [Google Scholar] [CrossRef]
Figure 1. The format followed by the schemes.
Figure 1. The format followed by the schemes.
Symmetry 17 01085 g001
Figure 2. Comparison of convergence curves between ISSA and other algorithms. (a) F1. (b) F2. (c) F3. (d) F4. (e) F5. (f) F6. (g) F7. (h) F8. (i) F9. (j) F10.
Figure 2. Comparison of convergence curves between ISSA and other algorithms. (a) F1. (b) F2. (c) F3. (d) F4. (e) F5. (f) F6. (g) F7. (h) F8. (i) F9. (j) F10.
Symmetry 17 01085 g002aSymmetry 17 01085 g002b
Figure 3. Comparison plots of true and predicted values for electric load dataset. (a) LSTM. (b) GWO-LSTM. (c) PSO-LSTM. (d) GOOSE-LSTM. (e) DE-LSTM. (f) GA-LSTM. (g) SSA-LSTM. (h) SSA1-LSTM. (i) ISSA-LSTM.
Figure 3. Comparison plots of true and predicted values for electric load dataset. (a) LSTM. (b) GWO-LSTM. (c) PSO-LSTM. (d) GOOSE-LSTM. (e) DE-LSTM. (f) GA-LSTM. (g) SSA-LSTM. (h) SSA1-LSTM. (i) ISSA-LSTM.
Symmetry 17 01085 g003aSymmetry 17 01085 g003b
Figure 4. Comparison plots of true and predicted values for real estate housing price dataset. (a) LSTM. (b) GWO-LSTM. (c) PSO-LSTM. (d) GOOSE-LSTM. (e) DE-LSTM. (f) GA-LSTM. (g) SSA-LSTM. (h) SSA1-LSTM. (i) ISSA-LSTM.
Figure 4. Comparison plots of true and predicted values for real estate housing price dataset. (a) LSTM. (b) GWO-LSTM. (c) PSO-LSTM. (d) GOOSE-LSTM. (e) DE-LSTM. (f) GA-LSTM. (g) SSA-LSTM. (h) SSA1-LSTM. (i) ISSA-LSTM.
Symmetry 17 01085 g004aSymmetry 17 01085 g004b
Table 1. Benchmark test functions.
Table 1. Benchmark test functions.
FunctionFunction NameTypeRangeOptimal Value
F1Sphere Unimodal[−100, 100]0
F2Schwefel 1.2Unimodal[−100, 100]0
F3Schwefel 2.21Unimodal[−100, 100]0
F4Generalized RosenbrockMultimodal[−30, 30]0
F5Quartic with NoiseMultimodal[−1.28, 1.28]0
F6Schwefel 2.26Multimodal[−500, 500]−12,569.5
F7Shekel’s Foxholes Multimodal[−65.536, 65.536]1
F8Shekel’s Family(A)Composite[0, 10]−10
F9Shekel’s Family(B)Composite[0, 10]−10
F10Shekel’s Family(C)Composite[0, 10]−10
Table 2. Simulation results of algorithm testing (d = 30).
Table 2. Simulation results of algorithm testing (d = 30).
FunctionAlgorithmBestMeanStd
F1GWO2.14 × 10−292.35 × 10−27 5.84 × 10−27
PSO−4.79 × 10−0−3.02 × 10−12.67 × 100
DE3.98 × 10−41.54 × 1015.72 × 101
GOOSE 5.22 × 10−31.83 × 1015.37 × 101
GA1.85 × 1042.61 × 1043.32 × 103
SSA0 1.54 × 10−678.38 × 10−67
SSA101.42 × 10−707.79 × 10−70
ISSA000
F2GWO2.14 × 10−292.35 × 10−275.84 × 10−27
PSO−4.79 × 100−3.02 × 10−12.67 × 100
DE3.98 × 10−41.54 × 1015.72 × 101
GOOSE 5.22 × 10−31.83 × 1015.37 × 101
GA1.85 × 1042.61 × 1043.32 × 103
SSA−1.05 × 101−9.45 × 1002.20 × 100
SSA101.42 × 10−707.79 × 10−70
ISSA000
F3GWO8.94 × 10−87.95 × 10−71.02 × 10−6
PSO−2.04 × 1011.55 × 1011.29 × 101
DE1.40 × 1012.49 × 1015.89 × 100
GOOSE 1.18 × 10−12.28 × 1012.16 × 101
GA6.22 × 1017.04 × 1013.91 × 100
SSA04.85 × 10−332.66 × 10−32
SSA105.93 × 10−343.24 × 10−33
ISSA07.65 × 10−2440
F4GWO2.58 × 1012.70 × 1016.61 × 10-1
PSO−1.47 × 1011.47 × 1009.04 × 100
DE1.13 × 1022.77 × 1037.00 × 103
GOOSE 2.55 × 1011.90 × 1024.24 × 102
GA8.89 × 1062.13 × 1075.85 × 106
SSA4.84 × 10−94.88 × 10−58.02 × 10−5
SSA11.05 × 10−82.20 × 10−44.02 × 10−4
ISSA5.92 × 10−147.16 × 10−63.35 × 10−5
F5GWO3.31 × 10−41.93 × 10−39.97 × 10−4
PSO−1.48 × 10−12.90 × 10−21.13 × 10−1
DE3.33 × 10−29.47 × 10−24.41 × 10−2
GOOSE 4.90 × 10−21.24 × 10−15.70 × 10−2
GA3.39 × 1001.05 × 1013.95 × 100
SSA3.50 × 10−54.62 × 10−44.17 × 10−4
SSA12.24 × 10−53.91 × 10−42.74 × 10−4
ISSA2.21 × 10−52.45 × 10−42.09 × 10−4
F6GWO−7.45 × 103−5.96 × 1039.88 × 102
PSO−5.00 × 1022.04 × 1012.56 × 102
DE−9.76 × 103−7.36 × 1031.29 × 103
GOOSE −8.18 × 103−6.88 × 1035.73 × 102
GA−1.18 × 104−1.14 × 1042.07 × 102
SSA−9.15 × 103−7.90E × 1037.88 × 102
SSA1−1.00 × 104−8.59 × 1036.82 × 102
ISSA−1.26 × 104−1.22 × 1046.68 × 102
F7GWO9.98 × 10−14.23 × 1004.08 × 100
PSO−3.20 × 101−3.20 × 1016.88 × 100
DE9.98 × 10−19.98 × 10−10
GOOSE 1.99 × 1001.31 × 1017.30 × 100
GA9.98 × 10−11.00 × 1005.95 × 10−3
SSA9.98 × 10−18.20 × 1005.60 × 100
SSA19.98 × 10−19.82 × 1004.87 × 100
ISSA9.98 × 10−17.50 × 1005.43 × 100
F8GWO−1.02 × 101−9.06 × 1002.26 × 100
PSO4.00 × 1005.73 × 1001.39 × 100
DE−1.02 × 101−8.65 × 1002.83 × 100
GOOSE −1.02 × 101−5.47 × 1003.27 × 100
GA−4.18 × 100−2.08 × 1008.50 × 10−1
SSA−1.02 × 101−8.28 × 1002.50 × 100
SSA1−1.02 × 101−9.12 × 1002.07 × 100
ISSA−1.02 × 101−1.00 × 1014.29 × 10−1
F9GWO−1.04 × 101−1.02 × 1019.70 × 10−1
PSO4.00 × 1004.77 × 1001.25 × 100
DE−1.04 × 101−1.01 × 1011.39 × 100
GOOSE −1.04 × 101−5.68 × 1003.26 × 100
GA−3.94 × 100−1.93 × 1006.00 × 10−1
SSA−1.04 × 101−9.52 × 1002.01 × 100
SSA1−1.04 × 101−9.68 × 1001.83 × 100
ISSA−1.04 × 101−1.02 × 1018.80 × 10−1
F10GWO−1.05 × 101−1.00 × 1012.06 × 100
PSO2.01 × 1004.57 × 1001.58 × 100
DE−1.05 × 101−1.04 × 1019.79 × 10−1
GOOSE −1.05 × 101−4.61 × 1003.38 × 100
GA−3.79 × 100−2.04 × 1006.19 × 10−1
SSA−1.05 × 101−9.45 × 1002.20 × 100
SSA1−1.05 × 101−9.61 × 1002.04 × 100
ISSA−1.05 × 101−1.05 × 1014.49 × 10−2
Table 3. Basic parameter settings of optimization algorithm.
Table 3. Basic parameter settings of optimization algorithm.
Parameter NameParameter Value
Pop_size20
Dim3
Max_iter200
HiddenUnits[10, 200]
MaxEpochs[30, 100]
InitialLearnRate[0.001, 0.05]
Table 4. Prediction errors and hyperparameters of different models (electric load dataset).
Table 4. Prediction errors and hyperparameters of different models (electric load dataset).
ModelsMAERmseR2HiddenUnitsLearningRateMaxEpochs
LSTM83.062106.3080.2631000.00180
PSO-LSTM73.457101.4680.329760.0039540
GWO-LSTM81.440101.9730.322160.0046067
GOOSE-LSTM73.234100.4480.342810.00172
DE-LSTM71.07996.0590.399410.028776
GA-LSTM72.111100.1460.346630.046851
SSA-LSTM80.47498.5230.367410.049587
SSA1-LSTM67.81890.5630.4651450.022560
ISSA-LSTM67.03783.1020.5501240.038590
Table 5. Prediction errors and hyperparameters of different models (real estate housing price dataset).
Table 5. Prediction errors and hyperparameters of different models (real estate housing price dataset).
ModelsMAERmseR2HiddenUnitsLearningRateMaxEpochs
LSTM18.56624.8050.1691000.001100
PSO-LSTM16.74822.6310.41850.002550
GWO-LSTM14.29420.8010.4161000.0022170
GOOSE-LSTM16.03524.2450.206470.0584
DE-LSTM15.39423.1290.278990.027259
GA-LSTM15.11721.3950.382160.043076
SSA-LSTM15.01621.0880.400120.033467
SSA1-LSTM14.91120.7650.4181490.036173
ISSA-LSTM12.88618.8220.5221680.04297
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, Y.-M.; Chu, S.-C.; Yildiz, A.R.; Pan, J.-S. Optimized Multiple Regression Prediction Strategies with Applications. Symmetry 2025, 17, 1085. https://doi.org/10.3390/sym17071085

AMA Style

Zhao Y-M, Chu S-C, Yildiz AR, Pan J-S. Optimized Multiple Regression Prediction Strategies with Applications. Symmetry. 2025; 17(7):1085. https://doi.org/10.3390/sym17071085

Chicago/Turabian Style

Zhao, Yi-Ming, Shu-Chuan Chu, Ali Riza Yildiz, and Jeng-Shyang Pan. 2025. "Optimized Multiple Regression Prediction Strategies with Applications" Symmetry 17, no. 7: 1085. https://doi.org/10.3390/sym17071085

APA Style

Zhao, Y.-M., Chu, S.-C., Yildiz, A. R., & Pan, J.-S. (2025). Optimized Multiple Regression Prediction Strategies with Applications. Symmetry, 17(7), 1085. https://doi.org/10.3390/sym17071085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop