Next Article in Journal
Inspiring from Galaxies to Green AI in Earth: Benchmarking Energy-Efficient Models for Galaxy Morphology Classification
Previous Article in Journal
Algorithms for Simulation of Shunt Currents in a Vanadium Redox Flow Battery
Previous Article in Special Issue
S-EPSO: A Socio-Emotional Particle Swarm Optimization Algorithm for Multimodal Search in Low-Dimensional Engineering Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Auto-Tuning Memory-Based Adaptive Local Search Gaining–Sharing Knowledge-Based Algorithm for Solving Optimization Problems

by
Nawaf Mijbel Alfadli
1,*,
Eman Mostafa Oun
1 and
Ali Wagdy Mohamed
1,2
1
Operations Research Department, Faculty of Graduate Studies for Statistical Research, Cairo University, Giza 12613, Egypt
2
Applied Science Research Center, Applied Science Private University, Amman 11931, Jordan
*
Author to whom correspondence should be addressed.
Algorithms 2025, 18(7), 398; https://doi.org/10.3390/a18070398
Submission received: 14 May 2025 / Revised: 22 June 2025 / Accepted: 25 June 2025 / Published: 28 June 2025

Abstract

The Gaining–Sharing Knowledge-based (GSK) algorithm is a human-inspired metaheuristic that models how people learn and disseminate knowledge across their lifetime. It has shown promising results across a range of engineering optimization problems. However, one of its major limitations lies in the use of fixed parameters to guide the search process, which often causes the algorithm to get stuck in local optima. To address this challenge, we propose an Auto-Tuning Memory-based Adaptive Local Search (ATMALS) empowered GSK, that is, ATMALS-GSK. This enhanced version of GSK introduces two key improvements: adaptive local search and memory-driven automatic tuning of parameters. Rather than relying on fixed values, ATMALS-GSK continuously adjusts its parameters during the optimization process. This is achieved through a Gaussian distribution mechanism that iteratively updates the likelihood of selecting different parameter values based on their historical impact on the fitness function. This selection process is guided by a weighted moving average that tracks each parameter’s contribution to fitness improvement over time. To further reduce the risk of premature convergence, an adaptive local search strategy is embedded, facilitating the algorithm’s escape from local traps and guiding it toward more optimal regions within the search domain. To validate the effectiveness of the ATMALS-GSK algorithm, it is evaluated on the CEC 2011 and CEC 2017 benchmarks. The results indicate that the ATMALS-GSK algorithm outperforms the original GSK, its variants, and other metaheuristics by delivering greater robustness, quicker convergence, and superior solution quality.

1. Introduction

Optimization seeks to identify the most effective solutions while adhering to specific constraints and conditions. It plays a vital role in both theoretical research and real-world engineering applications, where achieving peak performance, efficiency, or minimal cost is often the goal. As technology advances and we take on more complex challenges, the nature of optimization problems has evolved significantly. These problems have grown not only in scale but also in complexity, making the quest for optimal solutions increasingly difficult. While many optimization algorithms exist, reaching the global optimum, especially in large-scale or highly dynamic scenarios, remains a daunting task [1].
In general, numerical optimization algorithms are typically classified into two main groups: deterministic algorithms (also known as exact methods) and stochastic algorithms (also referred to as approximate methods) [2]. Traditional optimization methods are often grounded in exact mathematical formulations and rely on gradient-based information to navigate the solution space. These deterministic methods are well-suited to problems where all parameters are known, the objective function is smooth, and the search space is manageable. However, when faced with non-deterministic polynomial-time hardness (i.e., NP-hard problems), which are common in engineering and logistics, these traditional approaches struggle. High computational costs, long execution times, and a tendency to get trapped in local optima make them impractical for many real-world applications. This has opened the door for approximate optimization algorithms that favor speed, adaptability, and robustness over theoretical precision [3].
Metaheuristic algorithms have gained prominence as promising alternatives to conventional exact methods, especially for solving NP-hard problems. These algorithms are inherently flexible, employing probabilistic search mechanisms that do not require derivative information [4]. They can be applied across a wide array of applications, including those with noisy, nonlinear, or multi-modal objective functions. Their effectiveness mainly comes from their ability to strike a balance between exploring the search space widely (exploration) and focusing on the most promising areas (exploitation), which helps them avoid getting stuck too soon and discover high-quality solutions more quickly. This balance is key to the effectiveness of any metaheuristic approach [5].
Generally, metaheuristics can be grouped into four categories considering their underlying inspiration: evolutionary-based, swarm intelligence-based, physics-inspired, and human behavior-based techniques [6]. These encompass evolutionary approaches inspired by natural selection [7], swarm intelligence based on the collective behavior of animals [8], and physics-inspired models that simulate natural processes like gravity and thermodynamics [9]. A newer branch includes human-inspired algorithms, which model decision-making, negotiation, learning, or social behavior [10]. These human-based algorithms have shown considerable promise in solving complex optimization problems due to their intuitive mechanisms and adaptability, qualities that are especially relevant in today’s dynamic problem spaces [11,12,13,14].
Despite their versatility, existing metaheuristics face several limitations that hinder their widespread adoption and performance in highly demanding environments. One major challenge lies in their reliance on static control parameters, which often fail to adapt as the search progresses. This can result in premature convergence or inefficient exploration, especially in dynamic or high-dimensional problems [15]. Additionally, most current algorithms are not designed to fully utilize memory-based learning or historical performance, which limits their capacity to make intelligent, context-aware decisions [6]. Another critical gap is the lack of universal frameworks that can systematically tune the trade-off between exploitation and exploration based on problem complexity or runtime feedback. As real-world problems become more interconnected and volatile, there is a growing need for metaheuristics that can self-adapt, learn over time, and operate efficiently in uncertain, dynamic, and large-scale environments. Addressing these gaps is key to pushing the boundaries of what modern optimization algorithms can achieve.
The Gaining–Sharing Knowledge-based (GSK) algorithm is a relatively new member of the human-inspired metaheuristic family, originally proposed by Ali Wagdy et al. in 2020 [16]. It models the natural process of human learning and knowledge transfer, structured into two main stages: the junior phase, where knowledge is accumulated, and the senior phase, where it is shared. This intuitive and behaviorally grounded design has allowed GSK to tackle various real-world problems with promising results. However, like many algorithms, it has its shortcomings. GSK tends to suffer from premature convergence, often getting stuck in local optima, which limits its exploration capabilities. Moreover, its performance is highly sensitive to a set of manually tuned parameters. These parameters are usually determined through trial and error, a process that can be time-consuming, inconsistent, and difficult to generalize across different types of optimization challenges.
To overcome the parameter sensitivity issue in the original GSK algorithm, this paper introduces an Auto-Tuning Memory-based Adaptive Local Search (ATMALS) empowered GSK, that is, ATMALS-GSK. This improved version enhances GSK through two main innovations: memory-based automatic parameter tuning and an adaptive local search strategy. Unlike the traditional GSK, ATMALS-GSK dynamically adjusts its key parameters during the optimization process using a Gaussian-based mechanism. This mechanism updates the probability of selecting different parameter values based on their historical contribution to improving fitness. A memory-driven roulette wheel selection (RWS) method, powered by a weighted moving average, guides this process, ensuring the algorithm intelligently favors more effective parameter settings over time. In addition, an adaptive local search strategy is applied to the senior solutions, helping the algorithm better balance exploration and exploitation, escape local optima, and improve overall solution quality. These enhancements work to improve robustness and reduce premature convergence. The performance of ATMALS-GSK is evaluated through experiments on the CEC 2017 benchmark functions. Specifically, the key contributions of this paper can be mentioned as follows:
  • Proposing an adaptive Auto-Tuning Memory-based GSK (ATM-GSK) algorithm, which enhances the original GSK algorithm with dynamic parameter adaptation.
  • Integration of historical memory-based adaptiveness using a Gaussian-based weighted moving average mechanism that evaluates and adjusts parameter values based on their impact on the fitness function over time.
  • Development of a novel selection mechanism using a memory-driven RWS strategy that leverages exponential memory-based knowledge and weighted moving averages to prioritize more effective parameter values. This adaptive tuning strategy improves the ability of the algorithm in escaping local optima and enhancing the exploration–exploitation balance.
  • Incorporating an adaptive local search strategy to improve the exploitation capabilities of the algorithm. This approach enhances solution quality, preserves population diversity, and increases the ability of the algorithm to avoid trapping in local optima.
  • Validation of the ATM-GSK algorithm through experiments on CEC 2011 and CEC 2017 benchmarks. Results show that ATM-GSK outperforms the original GSK and other metaheuristic algorithms in terms of convergence accuracy and computational robustness.
The remainder of this study is structured as follows. Section 2 presents a review of the related literature and highlights the existing approaches in the field. Section 3 introduces the ATMALS-GSK algorithm, outlining its key mechanisms and innovations. Section 4 provides a comprehensive performance evaluation through testing on CEC 2011 and CEC 2017 benchmark functions. Finally, Section 5 concludes the study and suggests potential directions for future research.

2. Literature Review

2.1. Metaheuristic Optimization Algorithms

Numerical optimization algorithms are generally classified into two broad categories: deterministic and stochastic approaches [2], as illustrated in Figure 1. Deterministic algorithms operate under the assumption that the problem model is completely defined, with a clear and explicit link between potential solutions and their associated outcomes. Prominent examples of such algorithms include gradient descent, conjugate gradient methods, mixed-integer nonlinear programming, and the branch and bound technique. These methods typically leverage gradient information to navigate the solution space and are known for their robustness. Exact algorithms can guarantee optimal solutions, provided the problem is not NP-hard, in which case they may become computationally impractical. For complex, NP-hard problems commonly encountered in the real world, approximate methods (especially metaheuristic algorithms) have become increasingly popular. These approaches introduce randomness and probabilistic reasoning to efficiently explore the solution space. Rather than seeking the perfect solution, their goal is to find a sufficiently good one within a reasonable timeframe.
As depicted in Figure 1, metaheuristic algorithms are commonly grouped into four categories: evolutionary, swarm intelligence, physics-based, and human-based methods. Among evolutionary algorithms, traditional techniques such as Genetic Algorithms (GA) and Differential Evolution (DE) remain widely used. More recent innovations include the Covariance Matrix Adaptation Evolution Strategy with Adaptive Population (CMAES-APOP) [17], the Love Evolution Algorithm (LEA) [18], and the Multi-Adaptation DE (MadDE) [19].
Swarm intelligence-based algorithms have also found widespread use in engineering and optimization domains. Notable methods in this group include the Particle Swarm Optimization (PSO) [20], Whale Optimization Algorithm (WOA) [21], Quantum-based Avian Navigation Optimizer (QANA) [22], Starling Murmuration Optimizer (SMO) [23], and a modified version of Artificial Rabbits Optimization called MNEARO [24].
Physics-inspired algorithms draw their strategies from physical phenomena. These include Nuclear Reaction Optimization (NRO), modeled after nuclear interactions [25]; the Flow Regime Algorithm (FRA), inspired by principles of fluid dynamics [26]; Plasma Generation Optimization (PGO), based on plasma generation processes [27]; and Colliding Bodies Optimization (CBO), which mimics the physics of object collisions [28]. Another example is the Doppler Effect–Mean Euclidean Distance Threshold (DE-MEDT), which draws from the Doppler shift phenomenon [29].
Human-based metaheuristics are another fascinating class that has shown promise across various disciplines, including engineering, healthcare, and data analysis. These methods are inspired by social behaviors and cognitive processes. Some widely recognized human-based algorithms are Brain Storm Optimization (BSO) [30], Poor and Rich Optimization (PRO) [31], Teamwork Optimization Algorithm (TOA) [32], Skill Optimization Algorithm (SOA) [33], and the Gaining–Sharing Knowledge-Based Algorithm (GSK) [16].

2.2. GSK and Its Variants

GSK stands as a notable human-inspired metaheuristic algorithm that has successfully tackled numerous complex real-world optimization problems. However, it does have limitations. It often struggles to escape local optimal points and shows high sensitivity to parameter settings. In response, researchers have proposed various enhancements, which can be grouped into four categories: population improvement, hybridization techniques, strategy enhancement, and composite improvements. Each of these directions has brought meaningful contributions, as detailed below.

2.2.1. Population Enhancement Techniques

To boost GSK’s convergence and diversity, Jalali et al. [34] integrated opposition-based learning and introduced Cauchy mutation to diversify the search. Ma et al. [35] developed an enhanced version of the GSK algorithm, named mGSK-DPMO, which incorporates a dual-population framework and multiple operators to adaptively improve the algorithm’s performance. Bakir et al. [36] introduced FDBAGSK—an adaptive variant that balances fitness and distance from the best solution to enhance search capabilities. In another development, Pan et al. [37] proposed the Parallel Opposition-based GSK (POGSK), segmenting the population into multiple groups and employing the Taguchi method for cross-group knowledge sharing. Similarly, Kamarposhti et al. [38] proposed the Adaptive Population GSK-Improved Multi-Operator DE (APGSK-IMODE) algorithm to identify the optimal locations and operational states of switches for these enhancements. Furthermore, Liang and Wang introduced the EPD-GSK framework [39], an enhanced version of the GSK algorithm designed to prevent premature convergence due to loss of population diversity. EPD-GSK improves diversity through three key strategies: using Sobol sequences for diverse initialization, applying Cauchy mutation to the junior stage to broaden the search space, and employing reverse learning in the senior phase to help escape local optima.

2.2.2. Hybrid Algorithms

A growing trend in GSK improvement involves hybridization with other algorithms. Nabahat et al. [40] proposed integrating the GSK and WOA algorithms for noise reduction filter design. Nunes et al. [41] introduced a fusion of GSK and PSO, named GSKPSO, for the parameter estimation of lithium-ion batteries based on impedance data. Zhong et al. [42] hybridized GSK with the Harris Hawks Optimization (HHO), incorporating mutation operators to balance exploration and exploitation. In a similar vein, Krishnan and Thiyagarajan [43] combined GSK with DE, using DE to escape local optima after the senior sharing phase. Liang et al. [44] blended the WOA with GSK to enhance exploration and minimize early convergence.

2.2.3. Strategy Enhancement Techniques

To make GSK more effective in binary search spaces, several strategy-based improvements have been introduced. For example, Wagdy Mohamed et al. [45] proposed an enhanced GSK with adaptive parameters (APGSK) by introducing an adaptive parameter control mechanism. This modified version dynamically adjusts the knowledge ratio and knowledge factor based on the population’s diversity and progression, thereby simulating knowledge acquisition over time. Additionally, the algorithm employs a nonlinear population size reduction and starts with a larger initial population to strengthen exploration.

2.2.4. Composite Enhancement Techniques

The Adaptive GSK (AGSK) algorithm [46] utilizes adaptive controls for knowledge factor (kf) and knowledge ratio (kr), and adopted a population reduction mechanism for performance tuning. This variant enhances GSK performance in many ways. However, the strategy improvement variants are designed to solve the binary space problems and do not significantly improve the single-objective optimization performance of GSK. Numerous studies have focused on improving search strategies and achieved excellent performance [47]. Furthermore, Xie et al. 2025 [6] have introduced an adaptive variant named HPE-AGSK, which integrates historical probability expansion to enhance both local and global search capabilities. The key points include the implementation of an expansion sharing strategy in the junior phase, a historical probability expansion strategy in the senior phase, and a reverse gaining strategy in the early iterations to promote population diversity and prevent premature convergence.

2.3. Our Contributions Against the Existing Literature

While the existing literature presents numerous advances in metaheuristic algorithms, most continue to suffer from three critical limitations: the static nature of their control parameters, the lack of historical memory-based adaptability, and the lack of local search mechanisms in continuous-based search-space metaheuristics. In particular, the original GSK algorithm, although promising in concept and application, relies on fixed parameter settings that limit its performance to balance exploration and exploitation in a dynamic manner. This static design of the controllable parameters in GSK typically leads to premature convergence and suboptimal performance in complex or high-dimensional optimization problems. Furthermore, the absence of mechanisms to utilize historical information prevents GSK from automatic tuning during the search process.
To address these gaps, the proposed ATMALS-GSK algorithm introduces a novel framework that integrates auto-tuning, memory-based learning, and adaptive local search within the GSK paradigm. Unlike prior efforts that focused on trial-and-error tuning or simplistic parameter adaptation, ATMALS-GSK employs a Gaussian-driven, memory-aware tuning mechanism that dynamically updates parameter probabilities based on their historical impact on the objective function. Additionally, a RWS strategy, powered by exponential memory and a weighted moving average, intelligently prioritizes parameter configurations that have proven effective over time. Complementing these enhancements is an adaptive local search mechanism applied during the senior knowledge-sharing phase, enabling deeper exploration of promising regions and escape from local optima. Together, these contributions not only overcome the limitations of traditional GSK but also advance the field of human-inspired metaheuristics by introducing a robust, self-adaptive model that significantly outperforms existing methods on both benchmark functions and real-world engineering problems.

3. Proposed ATMALS-GSK Algorithm

In this section, at first, the original GSK algorithm is described (Section 3.1), and then, the proposed ATMALS-GSK algorithm is presented in detail (Section 3.2). The proposed ATMALS-GSK algorithm has three innovations compared to the GSK for enabling it with auto-tuning of the controllable parameters, memory-based knowledge enhancement, and adaptive local search-empowered search strategies. These enhancements are proposed in Section 3.2.1, Section 3.2.2, Section 3.2.3, Section 3.2.4 and Section 3.2.5.

3.1. GSK

The GSK algorithm begins with a randomly generated population where individuals, seen as “juniors,” lack prior experience. As the optimization evolves, these individuals interact with their environment. Through these interactions, they gain knowledge and eventually reach a “senior” stage, where they can share what they have learned with others, enhancing the overall intelligence of the population. This learning process is divided into the junior phase (beginner-intermediate) and the senior phase (intermediate-expert).
During the search process, individuals (i.e., solutions) continuously interact and influence one another through cooperation and competition. In the early stage of life, knowledge is mostly acquired from close, familiar networks, while the ability to judge others is still underdeveloped. As individuals engage with broader social and professional networks, they gain diverse experiences and develop the capacity to assess, categorize, and share knowledge with others. This transition enhances both personal learning and the collective performance of the population. The mathematical formulation of this process is detailed below. The overall pseudocode of GSK can be seen in Algorithm 1.
Let x j , j = 1,2 , 3 , . N represent feasible solutions (i.e., individuals), where N is the population size of the algorithm. Each solution x j is characterized using a vector x i j = ( x j 1 ,   x j 2 , , x j D ), where D denotes the number of knowledge domains or dimensions assigned to that person, i.e., decision variables. The corresponding fitness value for each individual j is given by f j   j = 1,2 , 3 , . N . The vector representation of an individual j during the junior phase (early-middle stage) is illustrated in Figure 2a, while its representation during the senior phase (middle-later stage) is depicted in Figure 2b.
Figure 2 illustrates that in the junior gaining and sharing phase (early-middle stage), more dimensions of each vector are updated compared to the senior phase. In contrast, during the senior phase (middle-later stage), the senior scheme updates more dimensions than the junior scheme. The number of dimensions updated in both phases depends on the knowledge rate, which controls how much knowledge is transferred between generations. The specific number of updated dimensions in each phase is determined using a nonlinear formula that adjusts over time, which can be formulated as follows:
D ( J u n i o r ) = P r o b S i z e × 1 G M a x G e n K
D ( S e n i o r ) = P r o b S i z e D ( J u n i o r ) = P r o b S i z e 1 1 G M a x G e n K
where K (K > 0) is a real number denoting the knowledge rate, ProbSize is the problem size (dimensions), G is the iteration number, and MaxGen is the maximum number of iterations.
Algorithm 1. Overall operation of original GSK Algorithm
Inputs:
- N: Population size
- D: Number of dimensions (decision variables)
- K: Knowledge rate (controls the volume of knowledge exchange)
- Kf: Knowledge factor (controls intensity of learning)
- Kr: Knowledge ratio (probability of applying knowledge update)
- MaxGen: Maximum number of generations
- P: Proportion for grouping in senior phase
Output: Global best solution (Xbest)
1   Initialize generation counter G ← 0
2   Generate an initial population of N individuals:
3      for 1 each i = 1:N
4         Initialize individual xi randomly in D-dimensional space
5         Evaluate fitness f (xi) for each individual i
6         for 2 each generation G = 1 to MaxGen:
7            Compute knowledge-sharing dimensions for both phases
8            Determine D(Junior) and D(Senior) using Equations (1) and (2).
9               Apply Junior Gaining–Sharing Knowledge Phase:
10              - For each individual i:
11              - Select nearest better (X_better) and worse (X_worse) individuals
12              - Choose a random peer (xr) for exploration
13              - Update D(Junior) dimensions based on the rules in Algorithm 2
14             Apply Senior Gaining–Sharing Knowledge Phase:
15              - Sort the population by fitness into Top P%, Middle N − 2P%, and Bottom P%
16              - For each individual i:
17             Select one peer from each group: X_best, X_mid, and X_worst
18             Update D(Senior) dimensions based on the rules in Algorithm 3
19             Evaluate new fitness values for all individuals
20             Update individuals based on new fitness:
21             Update the global best solution if a better one is found:
22       end for 2
23    end for 1
24  Return the global best solution found over all generations (X_best)

3.1.1. Junior Phase

In the junior phase, each solution primarily acquires knowledge from nearby individuals within small, familiar groups while also sharing with an outsider—someone not part of any group—driven by curiosity and a desire to explore. To update an individual’s position, the entire population is first sorted considering their objective function values in ascending order: x b e s t , …, x w o r s t . For a given individual xi, the nearest better person ( x i 1 ) and the nearest worse person ( x i + 1 ) are selected as knowledge sources, and a random individual xr is chosen for sharing. Special handling is applied for the best and worst solutions: the best individual uses the next two best as references, while the worst uses the two just above it. This process is formalized in the pseudo-code of Algorithm 2, which outlines the junior knowledge update mechanism.
Algorithm 2. Junior phase of GSK
1for 1 each solution i = 1:PopSize
2   for 2 each dimension j = 1:D
3      Generate a random number RND ∈ [0, 1]
4      if 1 RNDKr (Knowledge Ratio)
5         if 2  f ( x i ) > f ( x r )
6            Update   x i j n e w   =   x i   +   K f   ×   [ ( x r       x i   )   +   ( x i 1     x i + 1 ) ];
7         else if 2
8            Update   x i j n e w   =   x i   +   K f   ×   [ ( x i     x r )   +   ( x i 1     x i + 1 ) ];
9         end if 2
10      else if 1
11          Keep   current   value :   x i j n e w = x i j o l d ;
12      end if 1
13   end for 2
14end for 1

3.1.2. Senior Phase

In the senior phase, individuals utilize knowledge from various groups within the population, specifically, the best, middle, and worst individuals. This phase emphasizes how individuals are influenced by both high- and low-performing peers, simulating real-world judgment and selective knowledge exchange. First, the solutions are sorted based on their objective function values. Then, they are divided into three categories: the top P% as the best, the bottom P% as the worst, and the remaining N − 2 × P% as middle-performing individuals. The senior phase is illustrated in the pseudo-code of Algorithm 3.
For each individual xi, the knowledge update involves three sources: two vectors randomly selected from the best and worst groups (to simulate learning from extremes) and one from the middle group (to share or refine knowledge). If a random value RND is less than or equal to the knowledge ratio Kr, and the individual’s fitness is worse than that of the selected middle peer, its position is updated based on both extremes and the difference with the middle peer. Otherwise, the sharing component is reversed. If the random value RND exceeds Kr, no update occurs, and the individual’s value remains unchanged. The knowledge factor Kf regulates the amount of knowledge transferred, while the knowledge ratio Kr determines the likelihood of an update at each generation. Furthermore, P [0, 1] represents the percentage of the individuals.
Algorithm 3. Senior phase of GSK
1for 1 each individual i = 1:PopSize
2  for 2 each dimension j = 1:D
3    Generate a random number RND ∈ [0, 1]
4    if 1  RND     K r (Knowledge Ratio)
5      if 2  f ( x i ) > f ( x r )
6         Update   x i j n e w   =   x i   +   K f   ×   [ ( x m     x i )   +   ( x p b e s t     x p w o r s t ) ];
7      else if 2
8         Update   x i j n e w   =   x i   +   K f   ×   [ ( x i     x m )   +   ( x p b e s t     x p w o r s t ) ];
9      end if 2
10    else if 1
11         Keep   current   value :   x i j n e w = x i j o l d ;
12    end if 1
13  end for 2
14end for 1

3.2. ATMALS-GSK

As previously discussed, the original GSK algorithm is highly sensitive to a set of manually defined controllable parameters: Kf, Kr, K, and P. These parameters were originally set to fixed values based on empirical tuning as K = 10, Kf = 0.5, Kr = 0.9, and P = 0.1. However, this trial-and-error approach is not only time-consuming and tedious but also inconsistent and difficult to generalize across different problems. One of the key limitations of the original GSK is its reliance on these static parameter settings, which inherently restrict the ability of the algorithm to adaptively balance exploitation and exploration throughout the search process. This rigid configuration can result in suboptimal performance, particularly when tackling complex problems. Several underlying issues contribute to this limitation:
  • Lack of generalizability across problem domains: Fixed parameter values might yield strong results for a particular problem but often fail to perform well across different problem settings or objective landscapes. What works for a scheduling problem, for instance, may underperform in a continuous optimization task.
  • Inflexibility across search stages: The same parameter values are used throughout the entire execution of the algorithm, from the initial exploration phase to the final convergence phase. However, different phases of the search process typically benefit from different behaviors—strong exploration in early generations and refined exploitation in later ones. Static parameters fail to accommodate this shift, potentially leading to premature convergence or excessive wandering.
  • No feedback from historical performance: The original GSK lacks a learning mechanism to leverage past search experience. Without any form of memory or adaptive control, the algorithm is unable to assess the effectiveness of different parameter configurations over time. This absence of self-tuning prevents the algorithm from evolving and improving its own strategy dynamically.
To overcome the mentioned challenges, there is a need to design an adaptive historical memory-based version of GSK that can iteratively adjust and update its parameters in response to the current state of the search process, based on the problem specifications. This kind of smart tuning would make the algorithm more flexible, robust, and capable of handling a wider range of problems more effectively. To fill this gap, the proposed ATMALS-GSK algorithm introduces a new framework that combines automatic parameter tuning, memory-based learning, and adaptive local search into the GSK structure. The key idea is to boost GSK’s performance by learning from past experiences and adjusting its behavior accordingly.
Unlike the original GSK algorithm, ATMALS-GSK does not rely on fixed parameter settings. Instead, it uses a Gaussian-based mechanism to dynamically update the selection probabilities of different parameter values, based on how well they have performed in the past. A memory-guided RWS method, using a weighted moving average, helps the algorithm gradually favor more effective parameter settings over time. It is important to note that the ATMALS-GSK algorithm uses a memory-guided RWS strategy to adapt parameter values—not to select individuals, as is typically performed in traditional algorithms like GA. Specifically, this RWS utilizes a Gaussian-weighted moving average to intelligently guide the selection of optimal values for five key parameters: Kf, Kr, K, P, and PLS. By integrating memory-driven adaptation into the parameter control process, the algorithm enhances its ability to fine-tune performance in a data-driven and responsive manner.
In addition, ATMALS-GSK applies an adaptive local search to the top-performing (senior) solutions. This helps avoid getting stuck in local optima and ultimately leads to better results. The overall operation of the proposed ATMALS-GSK algorithm can be seen in Figure 3.
The initial value, range of variation, and discretized level for each parameter in the ATMALS-GSK algorithm are provided in Table 1. In this table, the parameter PLS represents the probability of applying a local search change for each decision variable of the NLS randomly selected solutions at every iteration of the ATMALS-GSK algorithm.

3.2.1. Adaptive K

The knowledge rate K serves as a key parameter in the GSK algorithm, controlling how much experience an individual accumulates throughout generations in the junior and senior phases. It directly affects the percentage of dimensions updated by each phase over time. As shown in Figure 4 and detailed in Table 2, different values of K result in distinct learning dynamics. When K = 1, the transition between junior and senior phases is smooth and linear. The number of junior dimensions decreases steadily, while the number of senior dimensions increases at the same rate, perfectly balancing exploration early on and gradually shifting to exploitation.
However, this linear behavior changes when K ≠ 1: For K < 1 (e.g., K = 0.5), the decline in junior updates is slow and non-linear, meaning the algorithm relies more heavily on exploration for a longer time. For instance, at 50% of the total generations, 71% of the dimensions are still updated by the junior scheme, with only 29% by the senior scheme. This results in gradual knowledge acquisition, allowing broader search of the solution space. Furthermore, for K > 1 (e.g., K = 2 or 5), the shift from junior to senior updates is faster and more aggressive. At K = 2, by halfway through the run, only 25% of dimensions are updated by the junior phase, and 75% by the senior. With K = 5, this shift is even more dramatic—only 3% of dimensions are updated via the junior scheme at the halfway point. This leads to faster convergence but may risk premature exploitation.
These patterns reveal that K functions as a learning rate controller within the GSK algorithm. Lower values of K promote slower, more cautious learning by extending the influence of the junior phase and encouraging prolonged exploration of the solution space. In contrast, higher K values accelerate the learning process by shifting focus to the senior phase earlier, thus enabling faster and more targeted exploitation. The selected K value directly influences how updated dimensions are allocated between the junior and senior schemes in each generation. These distributions are initialized at the start and adapt automatically as the algorithm evolves, underscoring the critical role of K in striking the right balance between exploration and exploitation based on the specific characteristics of the problem at hand.
In the proposed ATMALS-GSK algorithm, the range of K is considered within [5, 30] with a discretized level of 1. A historical weighted moving average function is used to adaptively tune the value of K based on an exponential function of the improvement rates of the different values of K in previous iterations. Weighted moving average [48] is a trend-following index used for smoothing the short-term variation of the parameter, which can be obtained on the basis of the past effectiveness of the different values of the parameter. The exponential weighted moving average function considers the weights that are assigned exponentially reducing for previous values, in which the closest values are of more importance.
After fitness evaluation for all individuals at every generation G, a Gaussian enhancement rate is applied to the different values of K as follows:
E R ( K V A L U E S , G ) = G F ( K s e l ) × exp F n e w F o l d F o l d                                   i f   F n e w < F o l d 0                                                                                                                                           i f   F n e w = F o l d 1 G F ( K s e l ) × exp F n e w F o l d F o l d             i f   F n e w > F o l d
where Fnew and Fold are the average fitness (cost) values of the top 50% of the population in the current generation G and previous generation G − 1, respectively. Furthermore, G is the current generation (iteration), and GF(Ksel) is a Gaussian function with a central point at the current selected value of K (i.e., Ksel) and standard deviation (STD) of δ, which can be calculated as follows:
G F ( K s e l ) = exp K V A L U E S K s e l 2 2 δ × m a x ( K V A L U E S ) m i n ( K V A L U E S ) 2
where Ksel is the current selected value of K (peak of the Gaussian function), and KVALUES are the discretized values of K within [5, 30], i.e., KVALUES = {5, 6, 7, …, 30}. In our experiments, we specified the STD of the Gaussian distribution as δ = 0.05 × (max(KVALUES) − min(KVALUES)). As an example for Ksel = 5, GF(Ksel) with different STDs can be illustrated as Figure 5.
Accordingly, the selection probability for each value v for K at generation G is calculated according to the weighted moving average of the Gaussian enhancement rates at the previous generations, which can be formulated as follows:
P r o b ( K V A L U E S , G ) = t = 1 G α t E R ( K V A L U E S , G t ) t = 1 G α t                                                             = α E R ( K V A L U E S , G 1 ) + α 2 E R ( K V A L U E S , G 2 ) + + α G E R ( K V A L U E S , 0 ) α + α 2 + + α G
In Equation (5), α is a constant (0 < α ≤ 1), which tunes the exponential type in the moving average method. The maximum value of α = 1 means simple averaging is applied to the previous enhancement rates. The higher the value of α, the more previous generations will be accounted for in the calculations. Typically, this parameter is set close to 1 [48]. An example considering α = 0.97 is illustrated in Figure 6.
At every generation during the optimization process, once the probability of the different values of K is computed according to their historical contributions to fitness improvement, a powered RWS mechanism is applied to guide the choice of K for the next update. According to this selection strategy, the probability of selecting the value of v for K ( V a l v K ) is calculated as follows:
P r o b R P ( V a l v K , G ) = P r o b ( V a l v K , G ) j K V A L U E S P r o b ( V a l j K , G ) R P ( G )
where RPG (RPG > 1) is the RWS power at generation G. If RPG = 1, the selection strategy is a simple RWS. The higher the RP, the more chance of elitism selection. To refine this mechanism over time, a gradually increasing RP is introduced according to Equation (7). In the early stages, RPG starts low, resulting in a flatter probability distribution that allows for broader exploration across a wider range of K values, even those with lower past performance. As the algorithm progresses and gains more insight into what works best for the problem at hand, RPG is gradually increased, sharpening the selection focus toward more historically successful values. This dynamic adjustment shifts the balance toward exploitation in later stages, helping the algorithm converge more effectively while still avoiding premature stagnation. It makes the proposed ATMALS-GSK algorithm more intelligent throughout the search process.
R P G = R P min + R P max R P min × 1 G M a x G e n

3.2.2. Adaptive Kr

The knowledge ratio (Kr) is a key parameter in the GSK algorithm that determines the proportion of individuals participating in the senior (exploitative) versus junior (explorative) phases. Its value typically lies within the range (0, 1], and it’s often set around 0.9 in the original GSK. When Kr is low (e.g., 0.3–0.6), more individuals are assigned to the junior phase, leading to broader exploration. This supports preserving variety within the population and is particularly useful for complex or multimodal problems. However, it may also slow down convergence. As Kr increases (e.g., 0.7–0.9), the algorithm shifts focus toward the senior phase, improving exploitation and accelerating convergence. This is effective for problems where refining known good solutions is more valuable than broad search. With very high values of Kr (close to 1), nearly the entire population engages in the senior scheme. Although this enhances exploitation and can lead to quicker convergence in simpler or well-known problems, it significantly limits diversity and raises the chances of converging too early to suboptimal solutions.
Finding the right balance for the knowledge ratio Kr is essential, as it directly influences how much of an individual’s knowledge is retained versus how much is gained from others during the learning process. To address the limitations of setting a fixed value for the knowledge ratio, the ATMALS-GSK algorithm adaptively adjusts Kr during the optimization run, allowing it to dynamically respond to the algorithm’s progress and the problem landscape. This adaptive adjustment is driven by historical improvement trends mirroring the mechanisms used for tuning K in Equations (3)–(7), which ensures that Kr evolves based on the success rate of previous generations. By allowing Kr to adaptively vary within [0.85, 1] with a discretized level of 0.01, the ATMALS-GSK algorithm intelligently balances knowledge retention and acquisition, improving both convergence speed and solution quality across diverse optimization problems.

3.2.3. Adaptive Kf

The knowledge factor (Kf) in the GSK algorithm governs the intensity with which individuals move toward their knowledge sources during the learning process. Serving as a scaling factor, Kf directly influences the magnitude of updates in both the junior and senior phases. With its range defined over (0, ∞), different values of Kf induce diverse search behaviors. Smaller values (e.g., 0.1–0.4) encourage fine-grained, cautious movements, favoring local refinement and stability—particularly beneficial in complex or noisy problem spaces. Moderate settings (around 0.5–1) offer a balanced trade-off, promoting steady convergence without excessive risk, which is often ideal for general-purpose optimization tasks. On the other hand, higher values of Kf (e.g., 2, 5, or even 10+) induce more aggressive movements toward knowledge sources, enabling faster exploration and improving the algorithm’s chances of escaping local optima early in the search. However, such large steps can lead to instability or overshooting if not regulated.
In the proposed ATMALS-GSK framework, Kf is adaptively tuned within a practical range of [0.2, 0.8] with a discretized level of 0.1, allowing the algorithm to adjust its aggressiveness over time. To support this adaptivity, the same Gaussian-based probability function used in Equations (3)–(7) is applied to Kf as well, enabling intelligent learning from historical performance and dynamically steering the search behavior. This adaptive mechanism ensures that Kf evolves with the problem landscape, contributing to a robust balance between effective exploration and reliable convergence.

3.2.4. Adaptive P

The parameter P, ranging within [0, 1], determines the probability of selecting individuals to participate in knowledge sharing during both the junior and senior learning phases. In practice, P is typically set to values less than 0.5, as lower values promote more selective and targeted interactions. This cautious approach helps the algorithm avoid excessive imitation, maintain diversity, and support broader exploration of the search space. In the proposed ATMALS-GSK algorithm, P is adaptively adjusted during the execution of the algorithm, typically within a refined range of 0.01 to 0.5, depending on the evolving needs of the search. Keeping P relatively small ensures that only a selective subset of individuals engages in knowledge exchange at each generation, preventing premature convergence and preserving healthy exploratory behavior, especially in early iterations.
To adaptively control P, the same Gaussian-based probability function used for tuning K in Equations (3)–(7) is applied, enabling the algorithm to learn from historical success and gradually fine-tune P as optimization proceeds, considering a range of variation for P as [0.05, 0.15] with a discretized level of 0.01. By restricting knowledge diffusion to high-quality individuals early on and allowing more sharing in later stages, the algorithm achieves a smooth transition from diverse exploration to focused exploitation, resulting in improved convergence reliability and solution quality across diverse problem domains.

3.2.5. Adaptive PLS

In the proposed ATMALS-GSK algorithm, an adaptive local search mechanism is incorporated to improve solution quality through controlled mutation. This local refinement is designed to balance exploitation and exploration dynamically as the algorithm progresses. At each iteration G, NLS solutions are randomly chosen from the entire population, and their variables are considered for local changing. The probability of mutating each decision variable in a selected individual (i.e., PLS/ProbSize) is governed by the same Gaussian-based probability function used in Equations (3)–(7), considering a variation range of [1, 2] with a discretized value of 0.1.
Simultaneously, the range of local search, denoted RangeLS, shrinks linearly over generations, as shown in Equation (8), which means that mutations at the beginning are more exploratory, enabling the algorithm to explore broader areas of the search space. Over time, the range contracts, guiding the search toward local regions around promising solutions, which enhances the exploitation capabilities of the algorithm. This dynamic adjustment promotes intensive search and fine-tuning near convergence.
R a n g e L S ( G ) = R a n g e L S , max R a n g e L S , max R a n g e L S , min × G M a x G e n
Together, these two adaptive mechanisms (i.e., auto-tuning adaptiveness of the local search probability and shrinking the local search range) help the proposed ATMALS-GSK algorithm evolve from automatic adaptive broad global search to focused local refinement, improving its ability to find high-quality, well-converged solutions.

3.2.6. Overall Operation of ATMALS-GSK

The overall pseudo-code of the ATMALS-GSK algorithm is outlined in Algorithm 4. It integrates adaptive parameter tuning, memory-guided learning, and a probabilistic local search mechanism to enhance solution quality over generations. Key parameters (K, Kf, Kr, P, and PLS) are dynamically selected using a Gaussian-enhanced probability scheme and powered RW S. The algorithm iteratively partitions the population into Junior and Senior groups, applies corresponding update rules, and performs local refinements to exploit promising areas of the solution space.
Algorithm 4. Proposed ATMALS-GSK Algorithm
Inputs:
  - MaxGen: Maximum generations
  - PopSize: Population size
  - ProbSize: Problem dimension
  - Fitness(): Objective function to minimize
  - KVALUES ← [1, 2, 3, …, 30]
  - KfVALUES ← [0.2, 0.3, …, 0.8]
  - KrVALUES ← [0.8, 0.81, …, 1.0]
  - PVALUES ← [0.05, 0.06, …, 0.2]
  - PLSVALUES ← [0.1, 0.11, …, 0.2]
  Initialize Probabilities for K, Kf, Kr, P, and PSL, using uniform distribution
  Initialize
    RPG ← 1,
    NLS ← 0.05 × PopSize,
    α ← 0.97,
    δ ← 0.05 × (max(VALUES) − min(VALUES))
Main ATMALS-GSK Algorithm:
1  Initialize population of PopSize individuals randomly
2  for 1 generation G = 1 to MaxGen do:
3    Evaluate Fitness of all individuals
4    for 2 each parameter in {K, Kf, Kr, P, PLS} do:
5      Compute Gaussian enhancement rate based on Equations (3) and (4)
6      Update selection probabilities using Equation (5)
7      Apply powered RWS using Equations (6) and (7) to choose parameter values
8    end for 2
9    RPG ← Adjust RPG based on generation G (gradual increase)
10     Partition individuals into Junior and Senior based on adaptive Kr value
11     for 3 each individual i in population do:
12        if i is in Junior group then:
13           Update D(junior) using junior learning rule (based on K)
14        else if i is in Senior group then:
15           Update D(senior) using senior learning rule (based on K)
16        end if
17      end for 3
18      Update RangeLS(G) using Equation (8)
19      Select NLS individuals randomly
20      for 4 each selected individual in NLS do:
21        for 5 each dimension d do:
22           Apply adaptive local search with probability of PLS/ProbSize
23        end for 5
24      end for 4
25      Update historical memory of parameters’ fitness improvements
26      Update the global best solution found so far
27    end for 1
Output: Best solution found and its fitness value

4. Performance Evaluation

In this section, the detailed results of the ATMALS-GSK algorithm are provided and compared against the original GSK [16] and advanced GSK versions, including AGSK [46], APGSK [45], FDBAGSK [36], and eGSK [49], as well as recent metaheuristic algorithms, including IADE [50], GJO-JOS [51], QCSCA [52], cSM [53], and FOWFO [54].

4.1. Settings

In the original GSK algorithm, the parameters K, Kf, Kr, and P were tuned empirically using a trial-and-error process. This involved running multiple experiments on benchmark functions and selecting the parameter values that yielded the best average performance. However, this method is problem-specific and lacks generalizability. Therefore, the values tuned for one problem may not transfer well to different optimization problems, especially technical or real-world ones. In contrast, the ATMALS-GSK algorithm adaptively tunes these parameters during runtime based on historical performance, making it more robust and broadly applicable across different problem domains.
The effectiveness of the proposed ATMALS-GSK algorithm is assessed using the 22 functions from the CEC 2011 and 30 functions from the CEC 2017 test suites (including unimodal functions, multimodal functions, hybrid functions, and composition functions). The full details of these benchmark problems can be found in the technical reports of CEC 2011 [55] and CEC 2017 [56].
In all simulations on the CEC 2011 benchmark suite, we follow the official recommendation by setting the maximum number of function evaluations (FEs) to 150,000 [55] to ensure a fair comparison with other methods. With a fixed population size of 100, this results in 1500 iterations for each function (since FEs = population size × iterations). Each algorithm is run independently 25 times on each function to ensure statistical reliability. All settings for simulations on CEC 2011 have been considered based on the technical report of the CEC 2011 benchmark in [55].
To assess the algorithm’s performance on the CEC 2017 benchmarks, solution error is used as the evaluation metric, defined as f(x) − f(x*), where x is the best solution found in a single run and x* is the known global optimum. As recommended in [56], errors smaller than 10−8 are considered zero. Experiments are conducted across four problem dimensions: D = 10, 30, 50, and 100, with FEs set to 10,000 × D. Each algorithm is run independently 51 times per function for every dimension. These settings are based on the technical report in [56].
In the following, to provide a thorough evaluation, the experimental results are organized into four main subsections. Section 4.2 focuses on analyzing the performance of ATMALS-GSK on the CEC 2011 and CEC 2017 benchmark functions. Section 4.3 presents a comparison between ATMALS-GSK, the original GSK algorithm, and its other modified variants. In Section 4.4, ATMALS-GSK is benchmarked against several recent metaheuristics. Furthermore, Section 4.5 includes statistical tests using the Wilcoxon signed-rank and Friedman methods to verify the significance of the results.

4.2. Results of ATMALS-GSK

4.2.1. CEC 2011

The comparative analysis between the original GSK algorithm [16] and the proposed ATMALS-GSK algorithm on the CEC 2011 benchmark suite reveals several noteworthy improvements brought about by the proposed enhancements. Table 3 reports the performance results in terms of Best, Median, Worst, Mean, and Standard Deviation (STD) across 25 independent runs, while Table 4 summarizes the statistical significance through the Wilcoxon signed-rank test.
According to the results obtained in Table 3, ATMALS-GSK demonstrates superior robustness and accuracy on a majority of the functions. For instance, in F2, the improvement is particularly noticeable; ATMALS-GSK reached a better minimum of −2.40 × 101 versus −1.35 × 101 by GSK, with a wider performance range indicating better exploration capacity. For multimodal functions like F6 and F7, which are known for their local optima traps, ATMALS-GSK consistently outperformed GSK in both mean and worst-case scenarios, showcasing its ability to avoid premature convergence. The mean value for F6 improved significantly from −6.94 to −1.79 × 101, and for F7, both mean and worst-case values decreased, reflecting more consistent convergence behavior. Notably, in some functions such as F3, F4, and F8, both techniques achieved identical or near-identical performances, indicating that these functions were relatively easy for both methods, leaving little room for improvement. The superiority of ATMALS-GSK becomes even more evident in complex hybrid and composition functions like F17 through F20. These functions typically combine the challenges of high dimensionality, non-separability, and rugged landscapes. On F17, the mean objective value decreased from 2.09 × 106 to 1.90 × 106, and the standard deviation dropped sharply, highlighting more reliable convergence. Similar improvements can be seen across F18 to F20, where ATMALS-GSK consistently provided better mean performance and significantly lower standard deviations.
The Wilcoxon signed-rank test in Table 4 confirms the overall effectiveness of ATMALS-GSK. With a P-value of 0.0168, the improvement is statistically significant at the 5% level. The test outcomes include 12 functions where ATMALS-GSK outperformed GSK, 5 where the two algorithms tied, and only 5 cases where GSK was better. The positive ranking sum (R+ = 127) far exceeds the negative sum (R− = 26), clearly supporting the conclusion that ATMALS-GSK provides a significant and consistent performance gain over the original GSK across diverse benchmark functions. The detailed analysis confirms that ATMALS-GSK is a promising enhancement over the baseline GSK.
As mentioned earlier, the proposed ATMALS-GSK algorithm incorporates adaptive parameters. Figure 7 shows the adaptive value of K across iterations for functions 1–6 as examples, along with the average value over all functions. Figure 8 illustrates the variations of the parameters K, KF, KR, P, and PLS during the algorithm’s execution. Additionally, Figure 9 presents the average convergence curves (over 25 runs) for the GSK and ATMALS-GSK algorithms on various functions from the CEC 2011 benchmark.

4.2.2. CEC 2017

The experimental results of the original GSK algorithm [16] and the proposed ATMALS-GSK algorithm over 51 independent runs are provided in Table 5, Table 6, Table 7 and Table 8 for D = 10, 30, 50, and 100, respectively. The results obtained highlight the improvements brought by ATMALS-GSK over GSK, particularly in terms of robustness, accuracy, and stability across most benchmark functions in the 10-dimensional test setting. It can be seen in Table 5 that both algorithms perform identically on some simpler or unimodal functions such as F1, F3, F4, F6, and F9. In these cases, all metrics are consistently zero, indicating that both algorithms successfully find the global optimum with perfect reliability. This suggests that the difficulty level of these functions is relatively low for both approaches.
However, when it comes to more complex and multimodal functions, ATMALS-GSK demonstrates a significant edge. For instance, in F5, the mean fitness value achieved by ATMALS-GSK is 7.57, substantially lower than the 20.3 reported for GSK, with a reduced standard deviation. This trend is consistently observed in F7 and F8, where ATMALS-GSK yields much better best and mean results, indicating not only better convergence but also enhanced robustness against local optima. The improvements are even more evident in challenging composite and hybrid functions. Take F10 as an example: while GSK results in a mean error of 1060, ATMALS-GSK drastically reduces this to 455, cutting the error by more than half. A similar pattern is seen in F12, F13, and F14, where ATMALS-GSK not only achieves better average fitness values but also shows much lower variability (i.e., lower standard deviation), signifying higher consistency across multiple runs.
One striking result is observed in F21, where GSK shows a mean of 193 with a high deviation of 50.7, while ATMALS-GSK consistently scores 100 across all runs with zero variance, reflecting a perfectly stable and optimal outcome. Likewise, in Function F25, ATMALS-GSK closely matches the best performance of GSK but with remarkably improved stability (STD = 0.132 compared to GSK’s 20.5), indicating better reliability in difficult scenarios. In some functions such as F17 and F22, while ATMALS-GSK does not always outperform GSK in all statistics, it still offers highly competitive or comparable results with a favorable trade-off between accuracy and stability. This suggests that the adaptive and learning-based mechanisms integrated into ATMALS-GSK enhance its generalization ability across a broad spectrum of optimization landscapes.
Figure 10 shows the adaptive value of K across iterations for functions F1 (as a unimodal function), F4 (as a multimodal function), F11 (as a hybrid function), and F21 (as a composition function), along with the average value over all functions. Figure 11 illustrates the variations of all parameters of the ATMALS-GSK algorithm over iterations during the execution of the algorithm. Additionally, Figure 12 illustrates the average convergence behavior of the original GSK algorithm and the proposed ATMALS-GSK algorithm across various CEC 2017 functions, with a problem dimensionality of D = 10. The results, averaged over 51 independent runs, demonstrate the optimization efficiency of both algorithms in terms of their function error value over iterations. As shown, ATMALS-GSK consistently outperforms the original GSK algorithm by achieving faster convergence rates and lower final error values across most test functions.

4.3. Comparison with GSK and Modified GSKs

The comparisons of the ATMALS-GSK algorithm with the original GSK and its variants (AGSK, APGSK, FDBAGSK, and eGSK) in terms of the average and standard deviation results (Mean ± STD) over 51 runs for different dimensions D = 10, 30, 50, and 100 on CEC 2017 are provided in Table 9, Table 10, Table 11 and Table 12, respectively.
As seen in Table 9, for functions 1 to 4 and 6, which demonstrate perfect or near-perfect convergence across all algorithms, the basic GSK algorithm alone is sufficient. For functions 5, 7, and 8, we begin to see notable differences. For example, function 5 showcases that ATMALS-GSK (7.57 ± 2.71) falls somewhere between AGSK (6.33 ± 1.89) and eGSK (4.82 ± 1.90). However, for more rugged terrains like functions 7 and 8, ATMALS-GSK performs more competitively, especially when compared to the original GSK, which suffers from high error means (30.7 and 20.2, respectively). Here, ATMALS-GSK manages to cut down the error significantly (to 16.3 and 7.57), underscoring its adaptive capacity. Functions like 10 and 12 really illustrate the leap ATMALS-GSK offers. In function 10, the original GSK’s performance (1060 ± 133) is substantially improved by AGSK and APGSK, but ATMALS-GSK achieves a more favorable trade-off (455 ± 193), suggesting it maintains robustness even in highly multimodal scenarios. Function 12 reveals an even more dramatic advantage—ATMALS-GSK posts a mean of 11.2 with very low deviation, compared to AGSK’s 44.6 and the original’s 89.3.
Although some modifications, particularly AGSK (0.039 ± 0.195), deliver astonishing performance, ATMALS-GSK lags slightly behind (2.78 ± 0.93). ATMALS-GSK continues to impress in functions 13, 15, and 16. Especially in function 13, its low mean (2.65) coupled with a reasonable standard deviation suggests more reliable convergence than many competitors. In function 17, however, ATMALS-GSK shows a significant spike in variance (mean = 12.5, SD = 90.2), possibly revealing sensitivity to initial population or parameter randomness in highly stochastic landscapes. This outlier behavior points to a possible need for parameter fine-tuning or incorporating stability controls. Looking at complex composite functions like 21–28, ATMALS-GSK either performs on par with the best (e.g., functions 26 and 28, both achieving perfect or near-perfect solutions like eGSK) or slightly underperforms in terms of variance but retains competitive means. Notably, in function 24, despite not having the best mean, ATMALS-GSK is very close to eGSK and performs better than all other variants, suggesting good generalization.
The results in Table 9, Table 10, Table 11 and Table 12 show that across most benchmark problems, the STDs of ATMALS-GSK are relatively controlled, signaling dependable convergence behavior. Its hybridization strategy, by embedding adaptive tuning and local search layers, appears to provide resilience against premature convergence—a known weakness in some simpler modifications. The results show that ATMALS-GSK proves itself to be a strong contender among the GSK family. While it does not dominate in every single function, it consistently offers balanced, reliable, and often superior performance, especially for complex or multimodal problems.

4.4. Comparison with Other Metaheuristic Algorithms

In this section, the performance of ATMALS-GSK is compared with recent metaheuristics (i.e., IADE, GJO-JOS, QCSCA, cSM, and FOWFO) in terms of the average and standard deviation results (Mean ± STD) over 51 runs for different dimensions D = 10, 30, 50, and 100 on CEC 2017, as reported in Table 13, Table 14, Table 15 and Table 16, respectively.
Table 13 highlights the strong performance of the ATMALS-GSK algorithm against other metaheuristics across a range of 10-dimensional benchmark functions. It consistently outperforms or matches competing methods, showing exceptional accuracy and robustness—often achieving exact solutions (e.g., 0.00 with zero variance) for simpler functions like F1, F3, and F4, while remaining highly competitive on more complex or multimodal ones like F5 and F10. Even on the most challenging hybrid and composite functions (F14–F29), ATMALS-GSK frequently delivers near-optimal results with low variability, outperforming methods like IADE, GJO-JOS, and QCSCA in many cases. Its ability to maintain low standard deviations throughout confirms its reliability and stable search behavior, making it a dependable choice for diverse and complex optimization tasks.
The results in Table 13, Table 14, Table 15 and Table 16 for different dimensions demonstrate the superior performance of ATMALS-GSK for solving high-dimensional problems. Its hybrid mechanism—blending adaptive learning and guided search via GSK—allows it to effectively balance exploration and exploitation. This advantage is evident not only in reaching optimal or near-optimal solutions but also in doing so reliably across a broad set of benchmark problems, outperforming recent metaheuristic algorithms.

4.5. Statistical Analysis

The statistical analysis of the CEC 2017 benchmark functions using the Wilcoxon and Friedman tests (shown in Table 17, Table 18, Table 19 and Table 20 and Figure 13 and Figure 14) supports the strong performance of ATMALS-GSK.
The Wilcoxon test results in Table 17 demonstrate that the proposed ATMALS-GSK algorithm surpasses GSK, AGSK, APGSK, FDBAGSK, and eGSK algorithms in 87, 86, 91, 77, and 76 cases out of all 29 × 4 = 116 cases (29 functions over four dimensions). This results in success ratios of ATMALS-GSK against GSK, AGSK, APGSK, FDBAGSK, and eGSK of 75%, 74.1%, 78.4%, 66.4%, and 65.5%, respectively. Furthermore, the results in Table 19 indicate that ATMALS-GSK exceeds IADE, GJO-JOS, QCSCA, cSM, and FOWFO, recording success ratios of 96.6%, 96.6%, 87.1%, 91.4%, and 54.3%, respectively, across all functions in overall dimensions.
The Friedman test results in Table 18 and Table 20 further showcase the effectiveness of the proposed ATMALS-GSK algorithm. In these tables, “Ranks” provide the average Friedman test rankings of the different algorithms over 51 runs for each dimension (D = 10, 30, 50, and 100). Furthermore, “Mean Rank” calculates the average rankings for all dimensions. Finally, “Overall Rank” provides the sorted rank of the different algorithms based on their Mean Rank.
Furthermore, Figure 13 and Figure 14 visually reinforce the outcomes derived from the Friedman test shown in Table 18 and Table 20, in terms of the “Mean Rank” and “Overall Rank” of the different algorithms. In Figure 13, based on the “Overall Rank” results, ATMALS-GSK secures the top rank when compared with its GSK-based variants across all dimensions. Similarly, Figure 14 presents a comparative ranking against other metaheuristics, where ATMALS-GSK again outperforms existing metaheuristics by a significant margin. The consistent first-place ranking across all dimensions highlights its competitiveness and dominance in achieving optimal solutions.
When compared to its GSK-based variants (Table 17 and Table 18 and Figure 13) and other well-known metaheuristics (Table 19 and Table 20 and Figure 14), ATMALS-GSK consistently performs better. The Wilcoxon test demonstrates that it significantly outperforms most of the competing methods across all problem dimensions (D = 10 to 100), with p-values often much lower than 0.05, especially when compared to the original GSK and its variants. Although ATMALS-GSK performs similarly to improved variants like AGSK and APGSK at lower dimensions (e.g., for D = 10), it becomes noticeably superior as the problem size increases. It also clearly outperforms popular metaheuristics like IADE, GJO-JOS, QCSCA, and cSM in nearly all cases. These results show that ATMALS-GSK is not only effective but also consistently reliable across a wide range of benchmark problems and dimensions.

4.6. Parametric Study of ATMALS-GSK

This section provides a parametric study of the proposed ATMALS-GSK algorithm to evaluate the individual effect of each contribution to the overall performance. To achieve this purpose, we compare ATMALS-GSK with its five components (ATMALS-GSK-1 to ATMALS-GSK-5). For each algorithm, one change among the five adaptive parameters (K, Kr, Kf, P, and PLS) is enabled, while keeping the other four parameters fixed as the basic GSK algorithm. The aim is to show that our algorithm ATMALS-GSK (with five adaptive parameters) is superior to ATMALS-GSK-1, ATMALS-GSK-2, ATMALS-GSK-3, ATMALS-GSK-4, and ATMALS-GSK-5 (each considering just one parameter as adaptive and fixing the other parameters).
In the following, we test the individual components of the ATMALS-GSK algorithm to assess the impact and efficiency of the modifications made to the GSK algorithm and to verify that ATMALS-GSK performs as expected. The ATMALS-GSK algorithm includes five main modifications compared to the original GSK algorithm: GSK with adaptive K (ATMALS-GSK-1), GSK with adaptive Kr (ATMALS-GSK-2), GSK with adaptive Kf (ATMALS-GSK-3), GSK with adaptive P (ATMALS-GSK-4), and GSK with adaptive local search mechanism PLS (ATMALS-GSK-5). These modifications influence the convergence and balance the exploration and exploitation aspects of the ATMALS-GSK algorithm.
To assess the effectiveness of the five ATMALS-GSK components, these versions are compared with ATMALS-GSK and GSK on CEC 2017 with D = 30. The average results of ATMALS-GSK, ATMALS-GSK-1, ATMALS-GSK-2, ATMALS-GSK-3, ATMALS-GSK-4, ATMALS-GSK-5, and GSK over 51 runs on different problems are reported in Table 21. The results show that all components impact the efficiency of the combined modifications and confirm that the ATMALS-GSK algorithm functions as intended. It is evident that the ATMALS-GSK algorithm significantly outperforms others in F1, F3, F4, F7, F8, F10, F11, F12, F13, F14, F16, F17, F18, F20, F21, F22, F23, F24, F25, F28, F29, and F30.
The statistical analysis results are provided in Table 22 and Figure 15. The Wilcoxon test results in Table 22 indicate that the proposed ATMALS-GSK algorithm surpasses its five versions in 22, 24, 23, 22, and 19 functions, which results in success ratios of ATMALS-GSK against ATMALS-GSK-1 to ATMALS-GSK-5 by 75.9%, 82.7%, 79.3%, 82.7%, and 65.5%, respectively. The rank analysis using the Friedman test in Figure 15 demonstrates the effectiveness of applying all modifications of the ATMALS-GSK algorithm together. According to the mean and overall ranks in Figure 15, ATMALS-GSK secures the top rank when compared with GSK and the ATMALS-GSK components. The Friedman test yields a p-value of 4.11 × 10−10, indicating a significant difference in ATMALS-GSK performance compared to GSK and the components of the ATMALS-GSK algorithm.

5. Conclusions

This paper introduced ATMALS-GSK, an enhanced version of the GSK algorithm, empowering auto-tuning memory-based historical information and adaptive local search mechanisms. Unlike the original GSK algorithm, which relies on fixed control parameters and often struggles with premature convergence, the proposed ATMALS-GSK algorithm integrates memory-based auto-tuning and an adaptive local search strategy. The key advantage of the proposed ATMALS-GSK algorithm lies in the adaptive Gaussian-driven adjustment of its controllable parameters using a historical memory mechanism that tracks their effectiveness over time. A historical memory-powered RWS, informed by a weighted moving average, ensures that the algorithm dynamically prioritizes more successful configurations. Furthermore, the adaptive local search component focuses on fine-tuning senior solutions, enabling better exploration of the search space and avoiding getting stuck in local optima.
Extensive experiments conducted on the CEC 2011 and CEC 2017 benchmark suites have demonstrated the superiority of the proposed ATMALS-GSK algorithm over the original GSK and its modified variants, as well as other recent metaheuristic algorithms. The Wilcoxon test results indicated that ATMALS-GSK surpasses the original GSK and its variants by obtaining success ratios of 75%, 74.1%, 78.4%, 66.4%, and 65.5% against GSK, AGSK, APGSK, FDBAGSK, and eGSK, respectively. Furthermore, ATMALS-GSK outperforms other metaheuristic algorithms—IADE, GJO-JOS, QCSCA, cSM, and FOWFO—achieving success ratios of 96.6%, 96.6%, 87.1%, 91.4%, and 54.3%, respectively, across all functions in overall dimensions. The Friedman test results also indicate that ATMALS-GSK ranks first across all dimensions in terms of average performance. The p-values from both tests confirm that these improvements are statistically significant. These results consistently showed improved robustness and higher solution quality across a variety of complex optimization problems.
Despite these achievements, the proposed algorithm is not without limitations. The introduction of memory-based mechanisms and local search strategies inevitably adds computational overhead, which might become significant in high-dimensional or real-time applications. Additionally, while the Gaussian-based parameter tuning works well in the current setup, it might need recalibration or refinement for highly irregular or domain-specific fitness landscapes. Furthermore, the performance of ATMALS-GSK may vary based on the characteristics of the problem being solved, suggesting that further customization or hybridization with problem-specific heuristics might be beneficial.
Therefore, there are several promising directions for future research. One main area involves exploring alternative memory management techniques, such as adaptive forgetting factors or reinforcement learning-inspired models, to further improve the parameter selection strategy. Integrating multi-objective capabilities into ATMALS-GSK would also broaden its applicability to problems with competing criteria. Moreover, the use of parallel and distributed computing could help mitigate the added computational complexity and make the algorithm more suitable for large-scale real-time problems. Finally, applying the ATMALS-GSK algorithm to real-world engineering optimization problems (especially those with discrete and integer variables) and exploring hybridization with problem-specific heuristics would be a promising direction for future work.

Author Contributions

Conceptualization, A.W.M.; Methodology, N.M.A.; Software, N.M.A.; Validation, N.M.A.; Formal analysis, N.M.A. and E.M.O.; Investigation, E.M.O. and A.W.M.; Resources, N.M.A.; Writing—original draft, N.M.A.; Writing—review & editing, N.M.A.; Supervision, E.M.O. and A.W.M.; Project administration, A.W.M.; Funding acquisition, N.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Dataset available on request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Selvarajan, S. A comprehensive study on modern optimization techniques for engineering applications. Artif. Intell. Rev. 2024, 57, 194. [Google Scholar] [CrossRef]
  2. Deng, J.; Zhou, C.; Wang, J. Approaches and application of heat and water network integration in chemical process system engineering: A review. Chem. Eng. Process.-Process Intensif. 2023, 183, 109263. [Google Scholar] [CrossRef]
  3. Fanian, F.; Rafsanjani, M.K.; Shokouhifar, M. Combined fuzzy-metaheuristic framework for bridge health monitoring using UAV-enabled rechargeable wireless sensor networks. Appl. Soft Comput. 2024, 167, 112429. [Google Scholar] [CrossRef]
  4. Shokouhifar, M.; Fanian, F.; Rafsanjani, M.K.; Hosseinzadeh, M.; Mirjalili, S. AI-driven cluster-based routing protocols in WSNs: A survey of fuzzy heuristics, metaheuristics, and machine learning models. Comput. Sci. Rev. 2024, 54, 100684. [Google Scholar] [CrossRef]
  5. Memarian, S.; Behmanesh-Fard, N.; Aryai, P.; Shokouhifar, M.; Mirjalili, S.; del Carmen Romero-Ternero, M. TSFIS-GWO: Metaheuristic-driven takagi-sugeno fuzzy system for adaptive real-time routing in WBANs. Appl. Soft Comput. 2024, 155, 111427. [Google Scholar] [CrossRef]
  6. Xie, L.; Wang, Y.; Tang, S.; Li, Y.; Zhang, Z.; Huang, C. Adaptive gaining-sharing knowledge-based variant algorithm with historical probability expansion and its application in escape maneuver decision making. Artif. Intell. Rev. 2025, 58, 1–43. [Google Scholar] [CrossRef]
  7. Janga Reddy, M.; Nagesh Kumar, D. Evolutionary algorithms, swarm intelligence methods, and their applications in water resources engineering: A state-of-the-art review. H2Open J. 2020, 3, 135–188. [Google Scholar] [CrossRef]
  8. Shokouhifar, M. FH-ACO: Fuzzy heuristic-based ant colony optimization for joint virtual network function placement and routing. Appl. Soft Comput. 2021, 107, 107401. [Google Scholar] [CrossRef]
  9. Siddique, N.; Adeli, H. Physics-based search and optimization: Inspirations from nature. Expert Syst. 2016, 33, 607–623. [Google Scholar] [CrossRef]
  10. Emami, H. Stock exchange trading optimization algorithm: A human-inspired method for global optimization. J. Supercomput. 2022, 78, 2125–2174. [Google Scholar] [CrossRef]
  11. Tanweer, M.R.; Sundaram, S. Human cognition inspired particle swarm optimization algorithm. In Proceedings of the 2014 IEEE Ninth International Conference on Intelligent Sensors, Sensor Networks and Information Processing, Singapore, 21–24 April 2014; pp. 1–6. [Google Scholar]
  12. Debnath, S.; Arif, W.; Baishya, S. Buyer inspired meta-heuristic optimization algorithm. Open Comput. Sci. 2020, 10, 194–219. [Google Scholar] [CrossRef]
  13. Liu, C.; Ma, Y.; Yin, H.; Yu, L. Human resource allocation for multiple scientific research projects via improved pigeon-inspired optimization algorithm. Sci. China Technol. Sci. 2021, 64, 139–147. [Google Scholar] [CrossRef]
  14. Veysari, E.F. A new optimization algorithm inspired by the quest for the evolution of human society: Human felicity algorithm. Expert Syst. Appl. 2022, 193, 116468. [Google Scholar]
  15. Liu, Z.; Jian, X.; Sadiq, T.; Shaikh, Z.A.; Alfarraj, O.; Alblehai, F.; Tolba, A. Efficient control of spider-like medical robots with capsule neural networks and modified spring search algorithm. Sci. Rep. 2025, 15, 13828. [Google Scholar] [CrossRef] [PubMed]
  16. Mohamed, A.W.; Hadi, A.A.; Mohamed, A.K. Gaining-sharing knowledge based algorithm for solving optimization problems: A novel nature-inspired algorithm. Int. J. Mach. Learn. Cybern. 2020, 11, 1501–1529. [Google Scholar] [CrossRef]
  17. Nguyen, D.M. The effect of mirrored sampling with active CMA and sample reuse in the CMAES-APOP algorithm. In Proceedings of the Genetic and Evolutionary Computation Conference Companion, Boston, MA, USA, 9–13 July 2022; pp. 403–406. [Google Scholar]
  18. Gao, Y.; Zhang, J.; Wang, Y.; Wang, J.; Qin, L. Love evolution algorithm: A stimulus–value–role theory-inspired evolutionary algorithm for global optimization. J. Supercomput. 2024, 80, 12346–12407. [Google Scholar] [CrossRef]
  19. Biswas, S.; Saha, D.; De, S.; Cobb, A.D.; Das, S.; Jalaian, B.A. Improving differential evolution through Bayesian hyperparameter optimization. In Proceedings of the 2021 IEEE Congress on Evolutionary Computation, Kraków, Poland, 28 June–1 July 2021; pp. 832–840. [Google Scholar]
  20. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  21. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  22. Zamani, H.; Nadimi-Shahraki, M.H.; Gandomi, A.H. QANA: Quantum-based avian navigation optimizer algorithm. Eng. Appl. Artif. Intell. 2021, 104, 104314. [Google Scholar] [CrossRef]
  23. Zamani, H.; Nadimi-Shahraki, M.H.; Gandomi, A.H. Starling murmuration optimizer: A novel bio-inspired algorithm for global and engineering optimization. Comput. Methods Appl. Mech. Eng. 2022, 392, 114616. [Google Scholar] [CrossRef]
  24. Jangir, P.; Ezugwu, A.E.; Saleem, K.; Arpita Agrawal, S.P.; Pandya, S.B.; Parmar, A.; Gulothungan, G.; Abualigah, L. A hybrid mutational Northern Goshawk and elite opposition learning artificial rabbits optimizer for PEMFC parameter estimation. Sci. Rep. 2024, 14, 28657. [Google Scholar] [CrossRef]
  25. Wei, Z.; Huang, C.; Wang, X.; Han, T.; Li, Y. Nuclear reaction optimization: A novel and powerful physics-based algorithm for global optimization. IEEE Access 2019, 7, 66084–66109. [Google Scholar] [CrossRef]
  26. Tahani, M.; Babayan, N. Flow Regime Algorithm (FRA): A physics-based meta-heuristics algorithm. Knowl. Inf. Syst. 2019, 60, 1001–1038. [Google Scholar] [CrossRef]
  27. Kaveh, A.; Akbari, H.; Hosseini, S.M. Plasma generation optimization: A new physically-based metaheuristic algorithm for solving constrained optimization problems. Eng. Comput. 2021, 38, 1554–1606. [Google Scholar] [CrossRef]
  28. Chen, D.; Lu, R.; Li, S.; Zou, F.; Liu, Y. An enhanced colliding bodies optimization and its application. Artif. Intell. Rev. 2020, 53, 1127–1186. [Google Scholar] [CrossRef]
  29. Kaveh, A.; Hosseini, S.M.; Zaerreza, A. A physics-based metaheuristic algorithm based on doppler effect phenomenon and mean euclidian distance threshold. Period. Polytech. Civ. Eng. 2022, 66, 820–842. [Google Scholar] [CrossRef]
  30. Shi, Y. Brain storm optimization algorithm. In Advances in Swarm Intelligence: Second International Conference, ICSI 2011, Chongqing, China, June 12–15, 2011, Proceedings, Part I; Springer: Berlin/Heidelberg, Germany, 2011; pp. 303–309. [Google Scholar]
  31. Moosavi, S.H.S.; Bardsiri, V.K. Poor and rich optimization algorithm: A new human-based and multi populations algorithm. Eng. Appl. Artif. Intell. 2019, 86, 165–181. [Google Scholar] [CrossRef]
  32. Dehghani, M.; Trojovský, P. Teamwork optimization algorithm: A new optimization approach for function minimization/maximization. Sensors 2021, 21, 4567. [Google Scholar] [CrossRef]
  33. Givi, H.; Hubalovska, M. Skill Optimization Algorithm: A New Human-Based Metaheuristic Technique. Comput. Mater. Contin. 2023, 74, 179–202. [Google Scholar] [CrossRef]
  34. Jalali, S.M.J.; Ahmadian, M.; Ahmadian, S.; Khosravi, A.; Alazab, M.; Nahavandi, S. An oppositional-Cauchy based GSK evolutionary algorithm with a novel deep ensemble reinforcement learning strategy for COVID-19 diagnosis. Appl. Soft Comput. 2021, 111, 107675. [Google Scholar] [CrossRef]
  35. Ma, H.; Zhang, J.; Wei, W.; Cheng, W.; Liu, Q. A Modified Gaining-Sharing Knowledge Algorithm Based on Dual-Population and Multi-operators for Unconstrained Optimization. In Advances in Swarm Intelligence: Second International Conference, ICSI 2011, Chongqing, China, June 12–15, 2011, Proceedings, Part I; Springer Nature: Cham, Switzerland, 2023; pp. 309–319. [Google Scholar]
  36. Bakır, H.; Duman, S.; Guvenc, U.; Kahraman, H.T. Improved adaptive gaining-sharing knowledge algorithm with FDB-based guiding mechanism for optimization of optimal reactive power flow problem. Electr. Eng. 2023, 105, 3121–3160. [Google Scholar] [CrossRef]
  37. Pan, J.S.; Liu, L.F.; Chu, S.C.; Song, P.C.; Liu, G.G. A new gaining-sharing knowledge based algorithm with parallel opposition-based learning for Internet of Vehicles. Mathematics 2023, 11, 2953. [Google Scholar] [CrossRef]
  38. Kamarposhti, M.A.; Shokouhandeh, H.; Lee, Y.; Kang, S.K.; Colak, I.; Barhoumi, E.M. Optimizing switch placement and status for sustainable grid operation and effective distributed generation integration using the APGSK-IMODE algorithm. Int. J. Low-Carbon Technol. 2024, 19, 2676–2686. [Google Scholar] [CrossRef]
  39. Liang, Z.; Wang, Z. Enhancing population diversity based gaining-sharing knowledge based algorithm for global optimization and engineering design problems. Expert Syst. Appl. 2024, 252, 123958. [Google Scholar] [CrossRef]
  40. Nabahat, M.; Khiyabani, F.M.; Navmipour, N.J. Hybrid Noise Reduction Filter Using the Gaining–Sharing Knowledge-Based Optimization and the Whale Optimization Algorithms. SN Comput. Sci. 2024, 5, 417. [Google Scholar] [CrossRef]
  41. Nunes, H.; Pombo, J.; Mariano, S.; do Rosário Calado, M. Parameter Estimation of Lithium-Ion Batteries Using GSKPSO Hybrid Method Based on Impedance Data. In Proceedings of the 2024 IEEE International Conference on Environment and Electrical Engineering and 2024 IEEE Industrial and Commercial Power Systems Europe (EEEIC/I&CPS Europe), Rome, Italy, 17–20 June 2024; pp. 1–6. [Google Scholar]
  42. Zhong, X.; Duan, M.; Zhang, X.; Cheng, P. A hybrid differential evolution based on gaining-sharing knowledge algorithm and harris hawks optimization. PLoS ONE 2021, 16, e0250951. [Google Scholar] [CrossRef] [PubMed]
  43. Navaneetha Krishnan, M.; Thiyagarajan, R. Multi-objective task scheduling in fog computing using improved gaining sharing knowledge based algorithm. Concurr. Comput. Pract. Exp. 2022, 34, e7227. [Google Scholar] [CrossRef]
  44. Liang, Z.; Shu, T.; Ding, Z. A novel improved whale optimization algorithm for global optimization and engineering applications. Mathematics 2024, 12, 636. [Google Scholar] [CrossRef]
  45. Mohamed, A.W.; Abutarboush, H.F.; Hadi, A.A.; Mohamed, A.K. Gaining-sharing knowledge based algorithm with adaptive parameters for engineering optimization. IEEE Access 2021, 9, 65934–65946. [Google Scholar] [CrossRef]
  46. Mohamed, A.W.; Hadi, A.A.; Mohamed, A.K.; Awad, N.H. Evaluating the performance of adaptive gainingsharing knowledge based algorithm on CEC 2020 benchmark problems. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation, Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  47. Hodashinsky, I.A. Methods for improving the efficiency of swarm optimization algorithms. A survey. Autom. Remote Control 2020, 82, 935–967. [Google Scholar] [CrossRef]
  48. Sadeghi, A.; Daneshvar, A.; Zaj, M.M. Combined ensemble multi-class SVM and fuzzy NSGA-II for trend forecasting and trading in Forex markets. Expert Syst. Appl. 2021, 185, 115566. [Google Scholar] [CrossRef]
  49. Jawad, M.A.; Roshdy, H.S.M.; Mohamed, A.W. Enhanced Gaining-Sharing Knowledge-based algorithm. Results Control Optim. 2025, 19, 100542. [Google Scholar] [CrossRef]
  50. Zhou, Z.; Shojafar, M.; Abawajy, J.; Bashir, A.K. IADE: An improved differential evolution algorithm to preserve sustainability in a 6G network. IEEE Trans. Green Commun. Netw. 2021, 5, 1747–1760. [Google Scholar] [CrossRef]
  51. Arini, F.Y.; Sunat, K.; Soomlek, C. Golden jackal optimization with joint opposite selection: An enhanced nature-inspired optimization algorithm for solving optimization problems. IEEE Access 2022, 10, 128800–128823. [Google Scholar] [CrossRef]
  52. Hu, H.; Shan, W.; Tang, Y.; Heidari, A.A.; Chen, H.; Liu, H.; Wang, M.; Escorcia-Gutierrez, J.; Mansour, R.F.; Chen, J. Horizontal and vertical crossover of sine cosine algorithm with quick moves for optimization and feature selection. J. Comput. Des. Eng. 2022, 9, 2524–2555. [Google Scholar] [CrossRef]
  53. Khalfi, S.; Iacca, G.; Draa, A. A single-solution–compact hybrid algorithm for continuous optimization. Memetic Comput. 2023, 15, 155–204. [Google Scholar] [CrossRef]
  54. Tang, Z.; Wang, K.; Zang, Y.; Zhu, Q.; Todo, Y.; Gao, S. Fractional-order water flow optimizer. Int. J. Comput. Intell. Syst. 2024, 17, 84. [Google Scholar] [CrossRef]
  55. Das, S.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for CEC 2011 Competition on Testing Evolutionary Algorithms on Real World Optimization Problems; Technical Report; Jadavpur University: Kolkata, India; Nanyang Technological University: Singapore, 2010. [Google Scholar]
  56. Awad, N.H.; Ali, M.Z.; Liang, J.J.; Qu, B.Y.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2017 Special Session and Competition on Single Objective Bound Constrained Real-Parameter Numerical Optimization; Technical Report; Nanyang Technological University: Singapore, 2016; pp. 1–34. [Google Scholar]
Figure 1. Classification of numerical optimization algorithms: exact and metaheuristic techniques.
Figure 1. Classification of numerical optimization algorithms: exact and metaheuristic techniques.
Algorithms 18 00398 g001
Figure 2. Solution representation of the GSK algorithm during junior and senior phases.
Figure 2. Solution representation of the GSK algorithm during junior and senior phases.
Algorithms 18 00398 g002
Figure 3. Overall flowchart of the ATMALS-GSK algorithm.
Figure 3. Overall flowchart of the ATMALS-GSK algorithm.
Algorithms 18 00398 g003
Figure 4. Effect of K on the population of junior and senior individuals versus generation.
Figure 4. Effect of K on the population of junior and senior individuals versus generation.
Algorithms 18 00398 g004
Figure 5. Gaussian function multiplication factor on the current value of K, with different STDs.
Figure 5. Gaussian function multiplication factor on the current value of K, with different STDs.
Algorithms 18 00398 g005
Figure 6. Weight of the enhancement rates at previous generations in current generation G = 100.
Figure 6. Weight of the enhancement rates at previous generations in current generation G = 100.
Algorithms 18 00398 g006
Figure 7. Adaptive behavior of K in the ATMALS-GSK algorithm vs. iterations in CEC 2011.
Figure 7. Adaptive behavior of K in the ATMALS-GSK algorithm vs. iterations in CEC 2011.
Algorithms 18 00398 g007
Figure 8. Adaptive behavior of K, KF, KR, P, and PLS in the ATMALS-GSK algorithm vs. iterations, on average for all benchmark functions in CEC 2011.
Figure 8. Adaptive behavior of K, KF, KR, P, and PLS in the ATMALS-GSK algorithm vs. iterations, on average for all benchmark functions in CEC 2011.
Algorithms 18 00398 g008
Figure 9. Average convergence curves (over 25 runs) of the GSK and proposed ATMALS-GSK algorithms for different functions of CEC 2011.
Figure 9. Average convergence curves (over 25 runs) of the GSK and proposed ATMALS-GSK algorithms for different functions of CEC 2011.
Algorithms 18 00398 g009aAlgorithms 18 00398 g009b
Figure 10. Adaptive behavior of K in the ATMALS-GSK algorithm vs. iterations in CEC 2017.
Figure 10. Adaptive behavior of K in the ATMALS-GSK algorithm vs. iterations in CEC 2017.
Algorithms 18 00398 g010
Figure 11. Adaptive behavior of K, KF, KR, P, and PLS in the ATMALS-GSK algorithm vs. iterations, on average for all benchmark functions in CEC 2017.
Figure 11. Adaptive behavior of K, KF, KR, P, and PLS in the ATMALS-GSK algorithm vs. iterations, on average for all benchmark functions in CEC 2017.
Algorithms 18 00398 g011
Figure 12. Average convergence curves (over 51 runs) of the original GSK and proposed ATMALS-GSK algorithms for different functions of CEC 2017, considering D = 10.
Figure 12. Average convergence curves (over 51 runs) of the original GSK and proposed ATMALS-GSK algorithms for different functions of CEC 2017, considering D = 10.
Algorithms 18 00398 g012aAlgorithms 18 00398 g012bAlgorithms 18 00398 g012c
Figure 13. Rank analysis of the proposed ATMALS-GSK algorithm against the original GSK and its modifications, based on the Friedman test.
Figure 13. Rank analysis of the proposed ATMALS-GSK algorithm against the original GSK and its modifications, based on the Friedman test.
Algorithms 18 00398 g013
Figure 14. Rank analysis of the proposed ATMALS-GSK algorithm against the existing metaheuristic algorithms, based on the Friedman test.
Figure 14. Rank analysis of the proposed ATMALS-GSK algorithm against the existing metaheuristic algorithms, based on the Friedman test.
Algorithms 18 00398 g014
Figure 15. Rank analysis of the ATMALS-GSK algorithm against original GSK and five ATMALS-GSK versions, based on the Friedman test.
Figure 15. Rank analysis of the ATMALS-GSK algorithm against original GSK and five ATMALS-GSK versions, based on the Friedman test.
Algorithms 18 00398 g015
Table 1. Comparison of fixed and adaptive parameters in GSK and ATMALS-GSK algorithms.
Table 1. Comparison of fixed and adaptive parameters in GSK and ATMALS-GSK algorithms.
ParameterGSKATMALS-GSKRange of VariationDiscretized Levels
Fixed Value Initial Value
K1010[5, 30]1
Kf0.50.5[0.2, 0.8]0.1
Kr0.90.9[0.85, 1]0.01
P0.10.1[0.05, 0.15]0.01
PLSN/A1[1, 2]/ProbSize0.1
NLSN/A0.05 × PopSizeN/AN/A
RPGN/A5[3, 5]N/A
Table 2. Comparison of the number of junior and senior dimensions for different values of K (problem size = 100).
Table 2. Comparison of the number of junior and senior dimensions for different values of K (problem size = 100).
KGD (Junior)D (Senior)
0.501000
0.25 MaxGen8713
0.5 MaxGen7129
0.75 MaxGen5050
MaxGen0100
101000
0.25 MaxGen7525
0.5 MaxGen5050
0.75 MaxGen2575
MaxGen0100
201000
0.25 MaxGen5644
0.5 MaxGen2575
0.75 MaxGen694
MaxGen0100
501000
0.25 MaxGen2476
0.5 MaxGen397
0.75 MaxGen199
MaxGen0100
Table 3. Results of GSK and ATMALS-GSK for CEC 2011 benchmark functions.
Table 3. Results of GSK and ATMALS-GSK for CEC 2011 benchmark functions.
FunctionGSK [16]ATMALS-GSK (Proposed)
BestMedianWorstMeanSTDBestMedianWorstMeanSTD
10.001.85 × 10−211.49 × 1013.285.210.14 × 10−31.01 × 1011.54 × 1016.636.10
2−1.35 × 101−1.14 × 101−9.25−1.13 × 1011.03−2.40 × 101−1.69 × 101−1.19 × 101−1.76 × 1013.13
31.15 × 10−51.15 × 10−51.15 × 10−51.15 × 10−59.12 × 10−131.15 × 10−51.15 × 10−51.15 × 10−51.15 × 10−51.13× 10−20
40.000.000.000.000.000.000.000.000.000.00
5−2.39 × 101−2.08 × 101−1.87 × 101−2.06 × 1011.21−3.59 × 101−2.89 × 101−2.42 × 101−2.97 × 1013.31
6−1.16 × 101−6.850.00−6.942.48−2.91 × 101−1.85 × 101−7.69−1.79 × 1014.85
71.591.781.971.781.08 × 10−19.99 × 10−11.21.491.291.25 × 10−1
82.20 × 1022.20 × 1022.20 × 1022.20 × 1020.002.20 × 1022.20 × 1022.20 × 1022.20 × 1020.00
91.29 × 1032.07 × 1033.09 × 1032.11 × 1035.02 × 1021.52 × 1032.43 × 1035.31 × 1032.67 × 1039.43 × 102
10−2.18 × 101−2.16 × 101−2.14 × 101−2.16 × 1011.19 × 10−1−2.18 × 101−1.75 × 101−1.30 × 101−1.81 × 1013.07
115.09 × 1045.24 × 1045.39 × 1045.24 × 1046.88 × 1025.15 × 1045.23 × 1045.31 × 1045.23 × 1044.29 × 102
121.07 × 1061.07 × 1061.08 × 1061.07 × 1061.73 × 1031.06 × 1061.07 × 1061.07 × 1061.07 × 1062.10 × 103
131.54 × 1041.54 × 1041.55 × 1041.54 × 1042.441.54 × 1041.54 × 1041.54 × 1041.54 × 1041.29 × 10−5
141.82 × 1041.84 × 1041.86 × 1041.84 × 1041.22 × 1021.80 × 1041.81 × 1041.84 × 1041.82 × 1041.02 × 102
153.28 × 1043.28 × 1043.28 × 1043.28 × 1041.55 × 1013.27 × 1043.27 × 1043.28 × 1043.27 × 1043.13 × 101
161.32 × 1051.35 × 1051.39 × 1051.35 × 1052.22 × 1031.25 × 1051.27 × 1051.29 × 1051.27 × 1059.35 × 102
171.97 × 1062.03 × 1062.36 × 1062.09 × 1061.20 × 1051.87 × 1061.89 × 1061.93 × 1061.90 × 1061.61 × 104
181.17 × 1061.27 × 1061.57 × 1061.27 × 1067.56 × 1049.36 × 1059.42 × 1059.46 × 1059.42 × 1052476.651
191.75 × 1062.03 × 1062.25 × 1062.00 × 1061.36 × 1059.37 × 1059.47 × 1059.76 × 1059.47 × 1057.47 × 103
201.12 × 1061.29 × 1061.46 × 1061.29 × 1069.20 × 1049.35 × 1059.41 × 1059.47 × 1059.41 × 1053.28 × 103
211.34 × 1011.61 × 1012.49 × 1011.70 × 1013.111.52 × 1011.95 × 1012.40 × 1011.92 × 1012.11
228.611.28 × 1012.09 × 1011.29 × 1012.938.611.63 × 1012.39 × 1011.61 × 1013.97
Table 4. Wilcoxon signed-rank test of ATMALS-GSK along with the original GSK for CEC 2011.
Table 4. Wilcoxon signed-rank test of ATMALS-GSK along with the original GSK for CEC 2011.
AlgorithmsR+R−p Value+=Dec.
ATMALS−GSK vs. GSK127260.01681255+
Table 5. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 10.
Table 5. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 10.
FunctionGSK [16]ATMALS-GSK (Proposed)
BestMedianWorstMeanSTDBestMedianWorstMeanSTD
10.000.000.000.000.000.000.000.000.000.00
30.000.000.000.000.000.000.000.000.000.00
40.000.000.000.000.000.000.000.000.000.00
51.53 × 1012.01 × 1012.58 × 1012.03 × 1012.792.998.471.00 × 1017.572.71
60.000.000.000.000.000.000.000.000.000.00
72.46 × 1013.10 × 1013.74 × 1013.07 × 1013.081.27 × 1011.69 × 1011.85 × 1011.63 × 1012.21
81.45 × 1011.99 × 1012.58 × 1012.02 × 1012.923.987.971.09 × 1017.572.11
90.000.000.000.000.000.000.000.000.000.00
107.54 × 1021.10 × 1031.29 × 1031.06 × 1031.33 × 1021.40 × 1025.07 × 1026.57 × 1024.55 × 1021.93 × 102
110.000.001.02× 10−80.000.006.40 × 10−11.772.671.677.10 × 10−1
121.42 × 1017.57 × 1012.69 × 1028.93 × 1017.26 × 1012.12 × 10−11.05 × 1013.45 × 1011.12 × 1019.45
139.19 × 10−16.569.736.561.413.32 × 10−11.805.702.652.18
145.19 × 10−35.421.19 × 1015.883.063.57 × 10−12.943.562.789.25 × 10−1
152.96 × 10−31.75 × 10−15.00 × 10−12.22 × 10−12.13 × 10−11.45 × 10−21.30 × 10−12.63 × 10−11.47 × 10−18.60 × 10−2
163.29 × 10−19.48 × 10−11.18 × 1014.274.954.89 × 10−11.111.801.124.70 × 10−1
179.76 × 10−19.942.52 × 1011.17 × 1017.092.261.26 × 1012.27 × 1011.25 × 1019.02 × 101
182.05 × 10−24.30 × 10−15.01 × 10−13.20 × 10−11.92 × 10−13.94 × 10−23.32 × 10−14.95 × 10−13.04 × 10−11.85 × 10−1
192.81 × 10−26.95 × 10−21.801.55 × 10−13.51 × 10−15.36 × 10−21.27 × 10−12.20 × 10−11.39 × 10−15.66 × 10−2
203.12 × 10−13.12 × 10−12.03 × 1011.193.991.161.974.552.511.14
211.00 × 1022.20 × 1022.31 × 1021.93 × 1025.07 × 1011.00 × 1021.00 × 1021.00 × 1021.00 × 1020.00
221.00 × 1021.00 × 1021.02 × 1021.00 × 1024.87 × 10−10.001.00 × 1021.00 × 1026.38 × 1014.73 × 101
233.11 × 1023.18 × 1023.29 × 1023.18 × 1023.992.24 × 10−13.06 × 1023.10 × 1022.76 × 1029.72 × 101
242.78 × 1023.50 × 1023.57 × 1023.44 × 1021.87 × 1011.00 × 1023.35 × 1023.39 × 1023.11 × 1027.44 × 101
253.98 × 1024.34 × 1024.46 × 1024.27 × 1022.05 × 1013.97 × 1023.97 × 1023.98 × 1023.97 × 1021.32 × 10−1
263.00 × 1023.00 × 1023.00 × 1023.00 × 1020.003.00 × 1023.00 × 1023.00 × 1023.00 × 1020.00
273.89 × 1023.90 × 1023.90 × 1023.89 × 1022.17 × 10−13.89 × 1023.89 × 1023.89 × 1023.89 × 1021.98 × 10−1
283.00 × 1023.00 × 1023.97 × 1023.12 × 1023.20 × 1013.00 × 1023.00 × 1023.00 × 1023.00 × 1020.00
292.39 × 1022.48 × 1022.57 × 1022.48 × 1024.762.32 × 1022.39 × 1022.43 × 1022.38 × 1023.33
303.96 × 1024.65 × 1025.02 × 1024.56 × 1023.44 × 1013.94 × 1024.42 × 1024.42 × 1024.29 × 1022.09 × 101
Table 6. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 30.
Table 6. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 30.
FunctionGSK [16]ATMALS-GSK (Proposed)
BestMedianWorstMeanSTDBestMedianWorstMeanSTD
10.000.000.000.000.000.000.000.000.000.00
30.005.66 × 10−88.54 × 10−67.07 × 10−71.82 × 10−60.000.000.000.000.00
49.28 × 10−34.007.25 × 1011.11 × 1012.17 × 1010.001.993.981.992.10
51.36 × 1021.62 × 1021.73 × 1021.60 × 1028.871.08 × 1012.67 × 1019.72 × 1011.53 × 1011.90 × 101
60.005.47 × 10−78.42 × 10−61.52 × 10−62.36 × 10−62.26 × 10−55.71 × 10−51.39 × 10−46.82 × 10−53.90 × 10−5
71.64 × 1021.87 × 1021.99 × 1021.87 × 1028.402.52 × 1015.60 × 1019.83 × 1014.33 × 1011.15 × 101
81.24 × 1021.56 × 1021.73 × 1021.55 × 1021.11 × 1011.07 × 1012.34 × 1015.06 × 1011.35 × 1011.29 × 101
90.000.000.000.000.000.000.008.95 × 10−21.79 × 10−23.77 × 10−2
105.84 × 1036.68 × 1037.16 × 1036.69 × 1033.54 × 1021.63 × 1032.96 × 1033.49 × 1032.80 × 1035.90 × 102
116.00 × 10−18.579.85 × 1013.30 × 1013.83 × 1019.942.04 × 1017.19 × 1011.46 × 1011.73 × 101
121.03 × 1035.32 × 1031.91 × 1046.62 × 1034.59 × 1039.02 × 1023.12 × 1038.19 × 1033.84 × 1032.62 × 103
134.61 × 1019.11 × 1011.82 × 1029.83 × 1013.42 × 1012.06 × 1013.91 × 1016.34 × 1013.01 × 1011.12 × 101
144.72 × 1015.73 × 1016.65 × 1015.69 × 1015.491.04 × 1011.79 × 1014.57 × 1011.78 × 1014.26
152.139.177.48 × 1011.45 × 1011.49 × 1015.881.75 × 1011.92 × 1011.43 × 1015.56
164.35 × 1027.76 × 1021.15 × 1037.96 × 1021.94 × 1021.97 × 1015.47 × 1027.50 × 1022.53 × 1022.81 × 102
176.28 × 1012.04 × 1023.77 × 1021.89 × 1029.44 × 1012.92 × 1016.37 × 1011.41 × 1024.71 × 1012.30 × 101
182.25 × 1013.69 × 1014.80 × 1013.68 × 1015.422.25 × 1012.74 × 1014.02 × 1013.02 × 1016.67
193.421.13 × 1012.22 × 1011.29 × 1016.028.031.37 × 1011.65 × 1011.30 × 1013.34
201.345.51 × 1014.51 × 1021.08 × 1021.14 × 1023.73 × 1015.58 × 1011.62 × 1024.74 × 1013.35 × 101
213.20 × 1023.49 × 1023.55 × 1023.46 × 1028.302.51 × 1022.65 × 1022.79 × 1022.66 × 1029.12
221.00 × 1021.00 × 1021.00 × 1021.00 × 1020.001.00 × 1021.00 × 1021.00 × 1021.00 × 1020.00
233.50 × 1024.86 × 1025.07 × 1024.70 × 1024.49 × 1013.97 × 1024.28 × 1024.43 × 1023.44 × 1021.70 × 101
245.29 × 1025.70 × 1025.86 × 1025.68 × 1021.45 × 1014.52 × 1024.77 × 1024.91 × 1024.75 × 1021.48 × 101
253.87 × 1023.87 × 1023.87 × 1023.87 × 1022.11 × 10−13.83 × 1023.86 × 1023.87 × 1023.86 × 1021.07
266.93 × 1029.56 × 1022.12 × 1039.87 × 1022.49 × 1021.45 × 1031.87 × 1031.97 × 1031.79 × 1031.93 × 102
274.80 × 1024.92 × 1025.14 × 1024.93 × 1028.024.87 × 1024.97 × 1025.06 × 1024.96 × 1026.31
283.00 × 1023.00 × 1024.04 × 1023.21 × 1024.22 × 1013.00 × 1023.00 × 1024.03 × 1023.10 × 1023.26 × 101
294.23 × 1025.92 × 1027.69 × 1025.77 × 1021.01 × 1024.09 × 1025.13 × 1025.68 × 1024.32 × 1025.95 × 101
301.94 × 1032.09 × 1032.36 × 1032.08 × 1039.27 × 1011.97 × 1032.06 × 1032.12 × 1032.06 × 1035.55 × 101
Table 7. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 50.
Table 7. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 50.
FunctionGSK [16]ATMALS-GSK (Proposed)
BestMedianWorstMeanSTDBestMedianWorstMeanSTD
11.71 × 1014.73 × 1024.61 × 1031.09 × 1031.24 × 1039.10 × 10−41.25 × 10−13.41 × 10−11.25 × 10−12.15 × 10−1
31.61 × 1033.79 × 1036.73 × 1033.85 × 1031.51 × 1031.12 × 1027.85 × 1021.46 × 1037.85 × 1026.73 × 102
41.33 × 10−27.21 × 1011.46 × 1028.33 × 1015.00 × 1018.209.802.70 × 1019.801.82 × 101
52.65 × 1023.25 × 1023.45 × 1023.20 × 1021.79 × 1012.98 × 1014.21 × 1019.48 × 1013.21 × 1011.27 × 101
63.11 × 10−72.80 × 10−61.64 × 10−53.78 × 10−63.52 × 10−61.17 × 10−77.14 × 10−61.55 × 10−57.14 × 10−68.31 × 10−6
73.29 × 1023.74 × 1023.86 × 1023.70 × 1021.41 × 1019.68 × 1011.08 × 1021.19 × 1021.08 × 1021.12 × 101
82.90 × 1023.28 × 1023.38 × 1023.24 × 1021.36 × 1015.16 × 1016.15 × 1017.14 × 1013.85 × 1019.92
90.000.008.95 × 10−21.07 × 10−22.79 × 10−21.50 × 10−21.91 × 10−13.97 × 10−11.91 × 10−12.06 × 10−1
101.21 × 1041.28 × 1041.37 × 1041.30 × 1044.50 × 1027.88 × 1038.80 × 1039.73 × 1038.80 × 1039.25 × 102
112.34 × 1012.96 × 1011.45 × 1023.45 × 1012.32 × 1012.59 × 1013.11 × 1013.63 × 1012.91 × 1015.17
122.16 × 1037.71 × 1033.17 × 1049.46 × 1037.01 × 1031.53 × 1043.65 × 1045.77 × 1043.65 × 1042.12 × 104
137.41 × 1016.24 × 1028.98 × 1031.49 × 1032.16 × 1036.72 × 1011.19 × 1021.71 × 1021.19 × 1025.17 × 101
145.76 × 1011.28 × 1021.42 × 1021.24 × 1021.87 × 1012.77 × 1013.69 × 1014.60 × 1013.69 × 1019.02
152.52 × 1013.62 × 1011.00 × 1024.20 × 1011.68 × 1013.31 × 1014.45 × 1015.59 × 1014.45 × 1011.14 × 101
161.30 × 1022.01 × 1032.70 × 1031.83 × 1036.59 × 1024.56 × 1027.91 × 1021.13 × 1034.91 × 1023.35 × 102
177.75 × 1021.39 × 1031.63 × 1031.35 × 1031.90 × 1024.77 × 1027.08 × 1029.39 × 1027.08 × 1022.31 × 102
181.78 × 1025.01 × 1021.45 × 1035.98 × 1023.37 × 1024.40 × 1011.17 × 1021.90 × 1021.17 × 1027.30 × 101
191.84 × 1012.92 × 1015.13 × 1013.05 × 1019.592.05 × 1012.83 × 1013.62 × 1012.83 × 1017.93
201.17 × 1031.40 × 1031.62 × 1031.37 × 1031.28 × 1025.44 × 1026.91 × 1028.38 × 1026.91 × 1021.48 × 102
214.90 × 1025.25 × 1025.46 × 1025.21 × 1021.31 × 1011.91 × 1021.98 × 1022.05 × 1021.98 × 1026.81
221.00 × 1021.31 × 1041.35 × 1041.10 × 1044.85 × 1036.29 × 1038.71 × 1031.11 × 1048.71 × 1032.41 × 103
234.20 × 1024.46 × 1027.49 × 1025.42 × 1021.39 × 1024.66 × 1024.79 × 1024.92 × 1024.39 × 1021.27 × 101
245.01 × 1025.24 × 1028.11 × 1026.34 × 1021.39 × 1025.41 × 1025.54 × 1025.68 × 1025.14 × 1021.35 × 101
254.60 × 1025.64 × 1026.11 × 1025.56 × 1024.62 × 1015.18 × 1025.45 × 1025.72 × 1025.45 × 1022.74 × 101
261.06 × 1031.29 × 1031.42 × 1031.27 × 1039.18 × 1019.95 × 1021.35 × 1031.71 × 1031.35 × 1033.55 × 102
275.18 × 1025.64 × 1029.16 × 1025.92 × 1028.29 × 1015.43 × 1025.71 × 1025.99 × 1025.71 × 1022.78 × 101
284.59 × 1024.97 × 1025.70 × 1024.94 × 1022.24 × 1014.64 × 1024.88 × 1025.12 × 1024.88 × 1022.39 × 101
293.28 × 1023.55 × 1024.09 × 1023.60 × 1022.23 × 1014.59 × 1025.01 × 1025.43 × 1025.01 × 1024.19 × 101
305.79 × 1055.81 × 1056.52 × 1055.96 × 1052.24 × 1045.74 × 1055.92 × 1056.11 × 1055.92 × 1051.85 × 104
Table 8. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 100.
Table 8. Results of GSK and ATMALS-GSK for CEC 2017 benchmark functions in D = 100.
FunctionGSK [16]ATMALS-GSK (Proposed)
BestMedianWorstMeanSTDBestMedianWorstMeanSTD
13.96 × 10−15.38 × 1032.11 × 1045.80 × 1034.63 × 1031.96 × 10−11.987.982.554.79
36.01 × 1041.15 × 1051.48 × 1051.15 × 1052.15 × 1044.87 × 1044.91 × 1045.30 × 1044.91 × 1041.26 × 104
48.46 × 1012.17 × 1022.89 × 1022.05 × 1024.57 × 1011.81 × 1021.97 × 1022.03 × 1021.98 × 1021.45 × 101
57.26 × 1017.70 × 1028.14 × 1025.31 × 1023.39 × 1023.70 × 1025.30 × 1025.90 × 1025.40 × 1022.17 × 101
61.25 × 10−56.83 × 10−53.14 × 10−21.72 × 10−36.46 × 10−31.80 × 10−66.20 × 10−33.90 × 10−21.24 × 10−22.73 × 10−2
78.33 × 1028.76 × 1029.16 × 1028.75 × 1021.83 × 1016.99 × 1028.21 × 1028.99 × 1028.20 × 1021.91 × 101
85.90 × 1017.36 × 1028.18 × 1024.92 × 1023.41 × 1023.95 × 1024.91 × 1025.88 × 1024.92 × 1021.12 × 101
95.457.691.72 × 1018.463.414.758.901.60 × 1018.803.14
102.87 × 1042.94 × 1043.04 × 1042.95 × 1044.44 × 1022.28 × 1042.45 × 1042.85 × 1042.44 × 1041.01 × 103
111.61 × 1022.65 × 1024.64 × 1022.78 × 1026.85 × 1011.30 × 1021.70 × 1022.80 × 1021.76 × 1026.28 × 101
122.25 × 1046.53 × 1043.23 × 1058.34 × 1047.52 × 1041.20 × 1042.92 × 1055.88 × 1053.78 × 1051.69 × 105
135.17 × 1012.85 × 1039.31 × 1033.20 × 1032.64 × 1031.06 × 1021.30 × 1022.00 × 1021.31 × 1025.43 × 101
143.30 × 1022.53 × 1031.69 × 1044.64 × 1034.47 × 1036.20 × 1022.49 × 1033.50 × 1032.57 × 1032.31 × 103
153.20 × 1014.23 × 1025.20 × 1037.33 × 1021.09 × 1035.90 × 1019.90 × 1011.40 × 1021.04 × 1024.05 × 101
162.87 × 1028.25 × 1027.04 × 1032.27 × 1032.61 × 1038.60 × 1021.80 × 1032.30 × 1031.89 × 1034.24 × 102
172.14 × 1034.12 × 1034.59 × 1033.91 × 1036.68 × 1028.00 × 1021.30 × 1032.80 × 1031.57 × 1036.92 × 102
181.93 × 1044.43 × 1041.87 × 1055.73 × 1043.63 × 1049.50 × 1031.13 × 1041.30 × 1041.14 × 1044.25 × 103
195.04 × 1018.41 × 1023.25 × 1031.00 × 1038.30 × 1029.00 × 1011.55 × 1022.30 × 1021.61 × 1027.63 × 101
203.97 × 1034.52 × 1034.80 × 1034.46 × 1032.22 × 1023.10 × 1032.80 × 1032.90 × 1032.84 × 1035.21 × 102
212.83 × 1023.29 × 1021.02 × 1036.07 × 1023.37 × 1022.10 × 1022.85 × 1023.10 × 1022.89 × 1022.11 × 101
222.86 × 1043.01 × 1043.09 × 1043.00 × 1044.57 × 1022.10 × 1042.35 × 1043.80 × 1042.43 × 1041.92 × 103
235.86 × 1026.11 × 1026.49 × 1026.11 × 1021.58 × 1015.25 × 1025.85 × 1026.45 × 1025.87 × 1021.45 × 101
248.89 × 1029.33 × 1029.69 × 1029.32 × 1021.70 × 1018.40 × 1029.10 × 1029.50 × 1029.11 × 1021.77 × 101
257.61 × 1028.20 × 1028.99 × 1028.21 × 1024.34 × 1017.20 × 1027.50 × 1027.80 × 1027.53 × 1024.18 × 101
263.33 × 1033.63 × 1033.96 × 1033.66 × 1031.78 × 1023.12 × 1033.35 × 1033.65 × 1033.42 × 1031.28 × 103
276.20 × 1026.46 × 1027.30 × 1026.57 × 1023.07 × 1016.10 × 1026.78 × 1027.90 × 1026.87 × 1023.47 × 101
284.99 × 1025.57 × 1026.09 × 1025.53 × 1023.25 × 1013.10 × 1023.30 × 1023.85 × 1023.51 × 1021.10 × 102
299.12 × 1021.21 × 1031.56 × 1031.21 × 1031.74 × 1021.01 × 1031.37 × 1031.78 × 1031.39 × 1032.14 × 102
302.41 × 1033.02 × 1033.72 × 1032.99 × 1032.73 × 1022.40 × 1033.21 × 1033.40 × 1033.23 × 1038.70 × 102
Table 9. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 10 on CEC 2017.
Table 9. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 10 on CEC 2017.
Fun.GSK [16]AGSK [46]APGSK [45]FDBAGSK [36]eGSK [49]ATMALS-GSK
10.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
30.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.000.00 ± 0.00
40.00 ± 0.000.00 ± 0.000.00 ± 0.00 × 1010.00 ± 0.000.00 ± 0.000.00 ± 0.00
52.03 × 101 ± 2.796.33 ± 1.898.04 ± 2.415.83 ± 1.644.82 ± 1.907.57 ± 2.71
60.00 ± 0.009.52 × 10−6 ± 2.42 × 10−54.45 × 10−5 ± 1.33 × 10−45.59 × 10−6 ± 1.35 × 10−50.00 ± 0.000.00 ± 0.00
73.07 × 101 ± 3.081.80 × 101 ± 2.391.90 × 101 ± 3.501.81 × 101 ± 3.571.54 × 101 ± 2.431.63 × 101 ± 2.21
82.02 × 101 ± 2.928.12 ± 2.299.16 ± 3.077.54 ± 2.305.29 ± 2.087.57 ± 2.11
90.00 ± 0.001.42 × 10−2 ± 6.64 × 10−21.76 × 10−3 ± 1.25 × 10−20.00 ± 0.000.00 ± 0.000.00 ± 0.00
101.06 × 103 ± 1.33 × 1023.20 × 102 ± 8.64 × 1013.05 × 102 ± 1.05 × 1023.02 × 102 ± 1.15 × 1027.59 × 102 ± 2.35 × 1024.55 × 102 ± 1.93 × 102
110.00 ± 0.003.99 × 10−1 ± 8.06 × 10−11.16 ± 9.76 × 10−19.30 × 10−1 ± 9.07 × 10−10.00 ± 0.001.67 ± 7.10 × 10−1
128.93 × 101 ± 7.26 × 1014.46 × 101 ± 5.34 × 1019.50 × 101 ± 8.64 × 1017.06 × 101 ± 5.51 × 1015.58 × 101 ± 6.17 × 1011.12 × 101 ± 9.45
136.56 ± 1.413.35 ± 2.164.29 ± 2.704.15 ± 1.965.50 ± 6.56 × 10−12.65 ± 2.18
145.88 ± 3.063.91 × 10−2 ± 1.95 × 10−15.73 × 10−1 ± 6.10 × 10−11.41 × 10−1 ± 3.70 × 10−12.20 × 10−1 ± 4.61 × 10−12.78 ± 9.25 × 10−1
152.22 × 10−1 ± 2.13 × 10−11.97 × 10−2 ± 2.88 × 10−21.48 × 10−1 ± 2.02 × 10−13.33 × 10−2 ± 4.30 × 10−23.06 × 10−1 ± 2.18 × 10−11.47 × 10−1 ± 8.60 × 10−2
164.27 ± 4.957.01 × 10−1 ± 3.37 × 10−17.44 × 10−1 ± 7.35 × 10−11.52 ± 1.014.37 ± 5.171.12 ± 4.70 × 10−1
171.17 × 101 ± 7.091.48 ± 9.38 × 10−11.08 ± 8.74 × 10−11.41 ± 1.116.01 ± 6.221.25 × 101 ± 9.02 × 101
183.20 × 10−1 ± 1.92 × 10−19.94 × 10−2 ± 1.56 × 10−13.97 × 10−1 ± 3.95 × 10−11.25 × 10−1 ± 1.09 × 10−16.95 × 10−1 ± 2.843.04 × 10−1 ± 1.85 × 10−1
191.55 × 10−1 ± 3.51 × 10−16.96 × 10−2 ± 4.46 × 10−28.32 × 10−2 ± 9.37 × 10−28.91 × 10−2 ± 4.13 × 10−23.75 × 10−2 ± 4.31 × 10−21.39 × 10−1 ± 5.66 × 10−2
201.19 ± 3.994.28 × 10−2 ± 1.08 × 10−19.79 × 10−2 ± 2.84 × 10−14.28 × 10−2 ± 1.08 × 10−18.16 × 10−1 ± 2.792.51 ± 1.14
211.93 × 102 ± 5.07 × 1011.02 × 102 ± 1.53 × 1011.03 × 102 ± 3.34 × 1019.80 × 101 ± 1.40 × 1011.77 × 102 ± 4.80 × 1011.00 × 102 ± 0.00
221.00 × 102 ± 4.87 × 10−18.00 × 101 ± 3.74 × 1015.24 × 101 ± 4.54 × 1015.87 × 101 ± 4.55 × 1019.82 × 101 ± 1.40 × 1016.38 × 101 ± 4.73 × 101
233.18 × 102 ± 3.992.97 × 102 ± 5.75 × 1013.00 × 102 ± 5.77 × 1012.98 × 102 ± 4.42 × 1013.04 × 102 ± 2.172.76 × 102 ± 9.72 × 101
243.44 × 102 ± 1.87 × 1011.04 × 102 ± 1.99 × 1019.33 × 101 ± 2.36 × 1011.00 × 102 ± 3.45 × 1013.05 × 102 ± 7.57 × 1013.11 × 102 ± 7.44 × 101
254.27 × 102 ± 2.05 × 1013.86 × 102 ± 5.84 × 1013.22 × 102 ± 1.31 × 1023.57 × 102 ± 1.03 × 1024.17 × 102 ± 2.30 × 1013.97 × 102 ± 1.32 × 10−1
263.00 × 102 ± 0.002.53 × 102 ± 1.10 × 1021.17 × 102 ± 1.10 × 1022.33 × 102 ± 1.09 × 1023.00 × 102 ± 6.43 × 10−143.00 × 102 ± 0.00
273.89 × 102 ± 2.17 × 10−13.89 × 102 ± 5.59 × 10−13.89 × 102 ± 1.733.89 × 102 ± 6.64 × 10−13.89 × 102 ± 1.70 × 10−13.89 × 102 ± 1.98 × 10−1
283.12 × 102 ± 3.20 × 1012.29 × 102 ± 1.29 × 1022.05 × 102 ± 1.40 × 1022.18 × 102 ± 1.35 × 1023.00 × 102 ± 1.37 × 10−133.00 × 102 ± 0.00
292.48 × 102 ± 4.762.43 × 102 ± 5.662.44 × 102 ± 1.03 × 1012.44 × 102 ± 6.142.32 × 102 ± 2.882.38 × 102 ± 3.33
304.56 × 102 ± 3.44 × 1014.70 × 102 ± 4.92 × 1015.15 × 102 ± 8.71 × 1014.97 × 102 ± 8.37 × 1014.45 × 102 ± 3.58 × 1014.29 × 102 ± 2.09 × 101
Table 10. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 30 on CEC 2017.
Table 10. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 30 on CEC 2017.
Fun.GSK [16]AGSK [46]APGSK [45]FDBAGSK [36]eGSK [49]ATMALS-GSK
10.00 ± 0.000.00 ± 0.001.65 × 10−5 ± 8.63 × 10−50.00 ± 0.001.75 × 10−1 ± 1.240.00 ± 0.00
37.07 × 10−7 ± 1.82 × 10−68.62 × 10−3 ± 3.06 × 10−24.37 × 10−1 ± 8.49 × 10−11.75 × 10−1 ± 1.080.00 ± 0.000.00 ± 0.00
41.11 × 101 ± 2.17 × 1018.87 ± 1.81 × 1016.98 ± 1.68 × 1013.07 × 101 ± 3.18 × 1012.58 ± 1.921.99 ± 2.10
51.60 × 102 ± 8.879.21 × 101 ± 1.27 × 1011.15 × 102 ± 1.33 × 1017.12 × 101 ± 1.07 × 1012.31 × 101 ± 8.241.53 × 101 ± 1.90 × 101
61.52 × 10−6 ± 2.36 × 10−66.98 × 10−6 ± 1.80 × 10−51.11 × 10−4 ± 3.68 × 10−44.71 × 10−6 ± 8.00 × 10−63.78 × 10−6 ± 4.11 × 10−66.82 × 10−5 ± 3.90 × 10−5
71.87 × 102 ± 8.401.17 × 102 ± 9.961.53 × 102 ± 1.78 × 1011.06 × 102 ± 9.575.31 × 101 ± 8.294.33 × 101 ± 1.15 × 101
81.55 × 102 ± 1.11 × 1019.74 × 101 ± 1.35 × 1011.13 × 102 ± 1.66 × 1017.87 × 101 ± 1.19 × 1012.45 × 101 ± 8.471.35 × 101 ± 1.29 × 101
90.00 ± 0.007.50 ± 7.083.65 × 101 ± 3.79 × 1012.22 ± 1.848.91 × 10−3 ± 6.36 × 10−21.79 × 10−2 ± 3.77 × 10−2
106.69 × 103 ± 3.54 × 1023.31 × 103 ± 1.86 × 1022.80 × 103 ± 2.40 × 1023.24 × 103 ± 2.22 × 1024.92 × 103 ± 9.42 × 1022.80 × 103 ± 5.90 × 102
113.30 × 101 ± 3.83 × 1013.65 × 101 ± 1.47 × 1013.19 × 101 ± 1.32 × 1011.96 × 101 ± 5.741.33 × 101 ± 2.06 × 1011.46 × 101 ± 1.73 × 101
126.62 × 103 ± 4.59 × 1038.86 × 103 ± 3.97 × 1031.31 × 104 ± 9.01 × 1031.04 × 104 ± 5.90 × 1034.24 × 103 ± 4.97 × 1033.84 × 103 ± 2.62 × 103
139.83 × 101 ± 3.42 × 1016.17 × 101 ± 4.51 × 1015.00 × 101 ± 2.35 × 1016.03 × 101 ± 2.78 × 1013.17 × 101 ± 2.31 × 1013.01 × 101 ± 1.12 × 101
145.69 × 101 ± 5.494.19 × 101 ± 5.984.38 × 101 ± 1.02 × 1013.76 × 101 ± 4.712.32 × 101 ± 3.711.78 × 101 ± 4.26
151.45 × 101 ± 1.49 × 1011.90 × 101 ± 9.212.37 × 101 ± 1.98 × 1011.42 × 101 ± 4.355.28 ± 3.641.43 × 101 ± 5.56
167.96 × 102 ± 1.94 × 1027.40 × 102 ± 1.32 × 1025.45 × 102 ± 1.39 × 1026.42 × 102 ± 1.30 × 1022.96 × 102 ± 2.22 × 1022.53 × 102 ± 2.81 × 102
171.89 × 102 ± 9.44 × 1011.82 × 102 ± 8.02 × 1011.92 × 102 ± 7.73 × 1011.65 × 102 ± 7.82 × 1015.92 × 101 ± 3.89 × 1014.71 × 101 ± 2.30 × 101
183.68 × 101 ± 5.425.28 × 102 ± 6.26 × 1021.31 × 103 ± 1.14 × 1032.80 × 102 ± 2.52 × 1023.27 × 101 ± 2.09 × 1013.02 × 101 ± 6.67
191.29 × 101 ± 6.021.52 × 101 ± 2.511.48 × 101 ± 3.351.37 × 101 ± 1.576.52 ± 1.881.30 × 101 ± 3.34
201.08 × 102 ± 1.14 × 1022.42 × 102 ± 8.46 × 1012.65 × 102 ± 7.96 × 1012.04 × 102 ± 7.54 × 1015.24 × 101 ± 4.31 × 1014.74 × 101 ± 3.35 × 101
213.46 × 102 ± 8.302.91 × 102 ± 2.46 × 1012.75 × 102 ± 7.85 × 1012.47 × 102 ± 6.00 × 1012.22 × 102 ± 9.582.66 × 102 ± 9.12
221.00 × 102 ± 0.003.51 × 102 ± 8.70 × 1021.65 × 102 ± 4.42 × 1021.00 × 102 ± 1.061.00 × 102 ± 7.33 × 10−131.00 × 102 ± 0.00
234.70 × 102 ± 4.49 × 1014.39 × 102 ± 1.38 × 1014.69 × 102 ± 2.84 × 1014.18 × 102 ± 9.603.59 × 102 ± 8.303.44 × 102 ± 1.70 × 101
245.68 × 102 ± 1.45 × 1015.06 × 102 ± 1.57 × 1015.65 × 102 ± 5.90 × 1014.96 × 102 ± 1.32 × 1014.33 × 102 ± 7.374.75 × 102 ± 1.48 × 101
253.87 × 102 ± 2.11 × 10−13.87 × 102 ± 1.923.87 × 102 ± 1.343.87 × 102 ± 6.73 × 10−13.87 × 102 ± 1.863.86 × 102 ± 1.07
269.87 × 102 ± 2.49 × 1025.85 × 102 ± 6.62 × 1023.74 × 102 ± 4.90 × 1027.18 × 102 ± 6.73 × 1028.55 × 102 ± 3.21 × 1021.79 × 103 ± 1.93 × 102
274.93 × 102 ± 8.024.98 × 102 ± 1.04 × 1015.10 × 102 ± 1.35 × 1015.01 × 102 ± 1.05 × 1014.97 × 102 ± 8.594.96 × 102 ± 6.31
283.21 × 102 ± 4.22 × 1013.13 × 102 ± 4.15 × 1013.16 × 102 ± 3.46 × 1013.32 × 102 ± 5.33 × 1013.10 × 102 ± 3.10 × 1013.10 × 102 ± 3.26 × 101
295.77 × 102 ± 1.01 × 1026.41 × 102 ± 8.28 × 1016.30 × 102 ± 5.35 × 1015.88 × 102 ± 5.92 × 1014.35 × 102 ± 4.07 × 1014.32 × 102 ± 5.95 × 101
302.08 × 103 ± 9.27 × 1012.14 × 103 ± 8.88 × 1012.25 × 103 ± 1.79 × 1022.18 × 103 ± 1.95 × 1022.08 × 103 ± 1.35 × 1022.06 × 103 ± 5.55 × 101
Table 11. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 50 on CEC 2017.
Table 11. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 50 on CEC 2017.
Fun.GSK [16]AGSK [46]APGSK [45]FDBAGSK [36]eGSK [49]ATMALS-GSK
11.09 × 103 ± 1.24 × 1031.59 ± 3.549.59 × 102 ± 1.25 × 1031.86 × 10−1 ± 6.16 × 10−12.82 × 10−1 ± 8.75 × 10−11.25 × 10−1 ± 2.15 × 10−1
33.85 × 103 ± 1.51 × 1031.87 × 103 ± 2.10 × 1033.60 × 103 ± 1.66 × 1036.83 × 102 ± 7.91 × 1029.42 × 10−7 ± 5.65 × 10−87.85 × 102 ± 6.73 × 102
48.33 × 101 ± 5.00 × 1015.80 × 101 ± 4.07 × 1018.13 × 101 ± 3.96 × 1015.73 × 101 ± 4.13 × 1011.02 × 101 ± 2.78 × 1019.80 ± 1.82 × 101
53.20 × 102 ± 1.79 × 1012.57 × 102 ± 2.06 × 1012.71 × 102 ± 2.56 × 1011.98 × 102 ± 2.09 × 1013.79 × 101 ± 9.723.21 × 101 ± 1.27 × 101
63.78 × 10−6 ± 3.52 × 10−62.48 × 10−4 ± 3.74 × 10−41.17 × 10−3 ± 2.24 × 10−31.73 × 10−4 ± 2.59 × 10−41.18 × 10−5 ± 9.13 × 10−67.14 × 10−6 ± 8.31 × 10−6
73.70 × 102 ± 1.41 × 1012.99 × 102 ± 2.85 × 1013.53 × 102 ± 3.40 × 1012.54 × 102 ± 2.41 × 1019.07 × 101 ± 9.861.08 × 102 ± 1.12 × 101
83.24 × 102 ± 1.36 × 1012.56 × 102 ± 2.81 × 1012.75 × 102 ± 2.44 × 1011.95 × 102 ± 1.91 × 1014.09 × 101 ± 1.01 × 1013.85 × 101 ± 9.92
91.07 × 10−2 ± 2.79 × 10−25.13 × 101 ± 4.26 × 1011.46 × 103 ± 1.16 × 1032.34 × 101 ± 2.49 × 1012.02 × 10−1 ± 4.51 × 10−11.91 × 10−1 ± 2.06 × 10−1
101.30 × 104 ± 4.50 × 1027.60 × 103 ± 3.54 × 1026.06 × 103 ± 2.61 × 1027.57 × 103 ± 3.37 × 1021.06 × 104 ± 1.22 × 1038.80 × 103 ± 9.25 × 102
113.45 × 101 ± 2.32 × 1011.25 × 102 ± 2.74 × 1019.18 × 101 ± 2.22 × 1017.83 × 101 ± 2.09 × 1012.97 × 101 ± 6.592.91 × 101 ± 5.17
129.46 × 103 ± 7.01 × 1039.71 × 104 ± 8.84 × 1041.08 × 105 ± 5.96 × 1041.22 × 105 ± 8.58 × 1043.98 × 104 ± 2.56 × 1043.65 × 104 ± 2.12 × 104
131.49 × 103 ± 2.16 × 1033.50 × 102 ± 1.74 × 1023.31 × 102 ± 1.81 × 1023.17 × 102 ± 1.76 × 1028.40 × 101 ± 7.01 × 1011.19 × 102 ± 5.17 × 101
141.24 × 102 ± 1.87 × 1011.11 × 102 ± 3.55 × 1011.58 × 102 ± 1.09 × 1028.35 × 101 ± 1.15 × 1014.03 × 101 ± 8.343.69 × 101 ± 9.02
154.20 × 101 ± 1.68 × 1018.43 × 101 ± 3.82 × 1011.06 × 102 ± 3.85 × 1016.12 × 101 ± 9.465.08 × 101 ± 2.03 × 1014.45 × 101 ± 1.14 × 101
161.83 × 103 ± 6.59 × 1021.47 × 103 ± 1.39 × 1021.02 × 103 ± 1.83 × 1021.19 × 103 ± 1.84 × 1025.01 × 102 ± 4.01 × 1024.91 × 102 ± 3.35 × 102
171.35 × 103 ± 1.90 × 1021.10 × 103 ± 1.51 × 1028.70 × 102 ± 1.14 × 1028.45 × 102 ± 1.71 × 1027.48 × 102 ± 2.95 × 1027.08 × 102 ± 2.31 × 102
185.98 × 102 ± 3.37 × 1021.07 × 104 ± 1.23 × 1049.11 × 103 ± 5.51 × 1033.72 × 103 ± 2.25 × 1031.21 × 102 ± 3.42 × 1021.17 × 102 ± 7.30 × 101
193.05 × 101 ± 9.594.23 × 101 ± 8.074.03 × 101 ± 6.803.86 × 101 ± 6.893.00 × 101 ± 9.422.83 × 101 ± 7.93
201.37 × 103 ± 1.28 × 1028.68 × 102 ± 1.25 × 1027.39 × 102 ± 1.35 × 1026.04 × 102 ± 1.68 × 1027.57 × 102 ± 2.08 × 1026.91 × 102 ± 1.48 × 102
215.21 × 102 ± 1.31 × 1014.47 × 102 ± 3.33 × 1014.55 × 102 ± 2.34 × 1013.89 × 102 ± 2.14 × 1012.36 × 102 ± 8.861.98 × 102 ± 6.81
221.10 × 104 ± 4.85 × 1037.90 × 103 ± 1.64 × 1036.18 × 103 ± 1.76 × 1036.79 × 103 ± 2.97 × 1039.57 × 103 ± 3.65 × 1038.71 × 103 ± 2.41 × 103
235.42 × 102 ± 1.39 × 1026.65 × 102 ± 2.22 × 1017.69 × 102 ± 3.06 × 1016.19 × 102 ± 2.23 × 1014.51 × 102 ± 1.14 × 1014.39 × 102 ± 1.27 × 101
246.34 × 102 ± 1.39 × 1027.26 × 102 ± 2.94 × 1018.40 × 102 ± 4.04 × 1016.90 × 102 ± 2.31 × 1015.20 × 102 ± 8.565.14 × 102 ± 1.35 × 101
255.56 × 102 ± 4.62 × 1015.52 × 102 ± 3.40 × 1015.63 × 102 ± 3.10 × 1015.22 × 102 ± 4.06 × 1015.63 × 102 ± 3.89 × 1015.45 × 102 ± 2.74 × 101
261.27 × 103 ± 9.18 × 1013.59 × 103 ± 7.34 × 1022.80 × 103 ± 1.80 × 1033.00 × 103 ± 4.95 × 1021.16 × 103 ± 4.43 × 1021.35 × 103 ± 3.55 × 102
275.92 × 102 ± 8.29 × 1015.28 × 102 ± 1.63 × 1015.62 × 102 ± 3.51 × 1015.24 × 102 ± 1.94 × 1016.09 × 102 ± 1.26 × 1025.71 × 102 ± 2.78 × 101
284.94 × 102 ± 2.24 × 1014.99 × 102 ± 3.50 × 1014.93 × 102 ± 1.91 × 1014.83 × 102 ± 2.86 × 1014.93 × 102 ± 3.43 × 1014.88 × 102 ± 2.39 × 101
293.60 × 102 ± 2.23 × 1011.00 × 103 ± 1.36 × 1021.01 × 103 ± 1.16 × 1028.32 × 102 ± 1.26 × 1024.37 × 102 ± 5.82 × 1015.01 × 102 ± 4.19 × 101
305.96 × 105 ± 2.24 × 1045.83 × 105 ± 6.20 × 1035.86 × 105 ± 5.35 × 1035.86 × 105 ± 1.60 × 1046.09 × 105 ± 3.27 × 1045.92 × 105 ± 1.85 × 104
Table 12. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 100 on CEC 2017.
Table 12. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with GSK and its modifications over 51 runs for D = 100 on CEC 2017.
Fun.GSK [16]AGSK [46]APGSK [45]FDBAGSK [36]eGSK [49]ATMALS-GSK
15.80 × 103 ± 4.63 × 1035.08 × 101 ± 1.19 × 1024.38 × 103 ± 2.89 × 1035.99 ± 9.799.58 × 103 ± 4.05 × 1042.55 ± 4.79
31.15 × 105 ± 2.15 × 1041.08 × 105 ± 2.11 × 1047.83 × 104 ± 9.64 × 1035.55 × 104 ± 1.42 × 1042.00 × 10−3 ± 7.54 × 10−34.91 × 104 ± 1.26 × 104
42.05 × 102 ± 4.57 × 1011.68 × 102 ± 4.47 × 1012.10 × 102 ± 5.52 × 1012.07 × 102 ± 4.53 × 1016.30 ± 1.87 × 1014.90 ± 1.45 × 101
55.31 × 102 ± 3.39 × 1027.28 × 102 ± 5.28 × 1017.72 × 102 ± 4.46 × 1016.66 × 102 ± 5.37 × 1011.12 × 102 ± 1.72 × 1019.70 × 101 ± 2.17 × 101
61.72 × 10−3 ± 6.46 × 10−32.01 × 10−2 ± 1.20 × 10−23.30 × 10−1 ± 2.97 × 10−11.63 × 10−2 ± 1.27 × 10−24.01 × 10−2 ± 5.90 × 10−21.24 × 10−2 ± 2.73 × 10−2
78.75 × 102 ± 1.83 × 1018.73 × 102 ± 5.20 × 1011.17 × 103 ± 6.49 × 1018.03 × 102 ± 5.29 × 1012.13 × 102 ± 2.45 × 1011.68 × 102 ± 1.91 × 101
84.92 × 102 ± 3.41 × 1027.15 × 102 ± 4.81 × 1017.82 × 102 ± 4.48 × 1016.47 × 102 ± 5.60 × 1011.12 × 102 ± 1.66 × 1019.65 × 101 ± 1.12 × 101
98.46 ± 3.415.50 × 102 ± 2.39 × 1021.34 × 104 ± 4.65 × 1037.14 × 102 ± 3.16 × 1021.98 × 101 ± 1.11 × 1011.05 × 101 ± 3.14
102.95 × 104 ± 4.44 × 1022.33 × 104 ± 6.30 × 1021.82 × 104 ± 4.05 × 1022.34 × 104 ± 6.43 × 1022.52 × 104 ± 3.03 × 1032.44 × 104 ± 1.01 × 103
112.78 × 102 ± 6.85 × 1018.81 × 102 ± 1.30 × 1029.17 × 102 ± 1.39 × 1029.43 × 102 ± 1.41 × 1021.86 × 102 ± 7.58 × 1011.76 × 102 ± 6.28 × 101
128.34 × 104 ± 7.52 × 1042.36 × 105 ± 9.14 × 1043.86 × 105 ± 1.42 × 1054.37 × 105 ± 1.86 × 1054.57 × 105 ± 2.05 × 1053.78 × 105 ± 1.69 × 105
133.20 × 103 ± 2.64 × 1033.61 × 103 ± 2.30 × 1033.66 × 103 ± 2.39 × 1034.10 × 103 ± 3.05 × 1039.96 × 101 ± 6.48 × 1011.31 × 102 ± 5.43 × 101
144.64 × 103 ± 4.47 × 1031.57 × 104 ± 1.14 × 1042.47 × 104 ± 1.82 × 1042.17 × 104 ± 1.30 × 1043.12 × 103 ± 2.51 × 1032.57 × 103 ± 2.31 × 103
157.33 × 102 ± 1.09 × 1034.34 × 102 ± 1.65 × 1025.57 × 102 ± 2.25 × 1024.64 × 102 ± 9.10 × 1018.68 × 101 ± 3.39 × 1011.04 × 102 ± 4.05 × 101
162.27 × 103 ± 2.61 × 1034.11 × 103 ± 1.62 × 1022.89 × 103 ± 3.19 × 1023.68 × 103 ± 2.09 × 1021.34 × 103 ± 4.90 × 1021.89 × 103 ± 4.24 × 102
173.91 × 103 ± 6.68 × 1022.87 × 103 ± 1.17 × 1022.32 × 103 ± 1.77 × 1022.51 × 103 ± 2.85 × 1021.68 × 103 ± 7.56 × 1021.57 × 103 ± 6.92 × 102
185.73 × 104 ± 3.63 × 1046.22 × 104 ± 2.83 × 1041.12 × 105 ± 4.34 × 1047.75 × 104 ± 3.64 × 1049.77 × 103 ± 4.86 × 1031.14 × 104 ± 4.25 × 103
191.00 × 103 ± 8.30 × 1021.82 × 102 ± 4.94 × 1012.39 × 102 ± 1.06 × 1021.93 × 102 ± 5.25 × 1011.96 × 102 ± 3.37 × 1021.61 × 102 ± 7.63 × 101
204.46 × 103 ± 2.22 × 1022.76 × 103 ± 1.53 × 1022.45 × 103 ± 2.07 × 1022.41 × 103 ± 2.63 × 1023.02 × 103 ± 6.43 × 1022.84 × 103 ± 5.21 × 102
216.07 × 102 ± 3.37 × 1029.36 × 102 ± 3.92 × 1019.75 × 102 ± 5.17 × 1018.77 × 102 ± 5.08 × 1013.27 × 102 ± 1.75 × 1012.89 × 102 ± 2.11 × 101
223.00 × 104 ± 4.57 × 1022.38 × 104 ± 3.51 × 1031.99 × 104 ± 3.84 × 1022.41 × 104 ± 7.56 × 1022.64 × 104 ± 2.41 × 1032.43 × 104 ± 1.92 × 103
236.11 × 102 ± 1.58 × 1011.19 × 103 ± 2.41 × 1011.37 × 103 ± 3.88 × 1011.04 × 103 ± 5.64 × 1016.26 × 102 ± 1.94 × 1015.87 × 102 ± 1.45 × 101
249.32 × 102 ± 1.70 × 1011.62 × 103 ± 5.79 × 1011.65 × 103 ± 5.48 × 1011.53 × 103 ± 5.13 × 1019.57 × 102 ± 2.07 × 1019.11 × 102 ± 1.77 × 101
258.21 × 102 ± 4.34 × 1018.07 × 102 ± 5.07 × 1018.19 × 102 ± 3.76 × 1017.71 × 102 ± 5.78 × 1017.85 × 102 ± 6.35 × 1017.53 × 102 ± 4.18 × 101
263.66 × 103 ± 1.78 × 1021.09 × 104 ± 1.11 × 1031.22 × 104 ± 7.50 × 1021.01 × 104 ± 7.08 × 1023.06 × 103 ± 1.64 × 1033.42 × 103 ± 1.28 × 103
276.57 × 102 ± 3.07 × 1016.41 × 102 ± 5.25 × 1019.63 × 102 ± 5.64 × 1016.37 × 102 ± 3.41 × 1017.08 × 102 ± 4.65 × 1016.87 × 102 ± 3.47 × 101
285.53 × 102 ± 3.25 × 1015.63 × 102 ± 3.15 × 1015.74 × 102 ± 2.83 × 1015.72 × 102 ± 2.75 × 1013.84 × 102 ± 1.17 × 1023.51 × 102 ± 1.10 × 102
291.21 × 103 ± 1.74 × 1023.69 × 103 ± 1.61 × 1023.71 × 103 ± 2.41 × 1023.17 × 103 ± 2.80 × 1021.55 × 103 ± 2.86 × 1021.39 × 103 ± 2.14 × 102
302.99 × 103 ± 2.73 × 1025.01 × 103 ± 2.41 × 1037.42 × 103 ± 2.35 × 1034.71 × 103 ± 2.90 × 1033.68 × 103 ± 7.86 × 1023.23 × 103 ± 8.70 × 102
Table 13. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 10 on CEC 2017.
Table 13. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 10 on CEC 2017.
Fun.IADE [50]GJO-JOS [51]QCSCA [52]cSM [53]FOWFO [54]ATMALS-GSK
18.54 × 102 ± 1.51 × 1032.30 × 104 ± 1.40 × 1043.09 × 104 ± 1.37 × 1052.19 × 103 ± 2.39 × 1033.91 × 10−4 ± 3.18 × 10−40.00 ± 0.00
33.00 × 102 ± 8.90 × 10−103.00 × 102 ± 3.70 × 10−13.62 × 102 ± 1.59 × 1021.38 × 10−12 ± 9.10 × 10−134.58 × 10−7 ± 4.09 × 10−70.00 ± 0.00
43.51 × 103 ± 4.55 × 10−134.10 × 102 ± 1.704.11 × 102 ± 1.79 × 1016.32 × 10−1 ± 3.23 × 10−15.60 × 10−6 ± 1.71 × 10−50.00 ± 0.00
56.18 × 102 ± 3.885.20 × 102 ± 1.10 × 1015.22 × 102 ± 7.131.46 × 101 ± 4.845.70 ± 1.697.57 ± 2.71
66.43 × 102 ± 6.736.00 × 102 ± 9.60 × 10−16.01 × 102 ± 2.69 × 10−15.44 × 10−4 ± 1.93 × 10−31.91 × 10−2 ± 1.04 × 10−20.00 ± 0.00
77.38 × 102 ± 1.53 × 1017.30 × 102 ± 1.20 × 1016.20 × 102 ± 3.722.35 × 101 ± 7.141.73 × 101 ± 3.031.63 × 101 ± 2.21
88.13 × 102 ± 4.028.20 × 102 ± 8.708.20 × 102 ± 5.901.57 × 101 ± 7.466.65 ± 2.217.57 ± 2.11
91.54 × 103 ± 5.51 × 1019.00 × 102 ± 2.90 × 10−19.00 × 102 ± 1.31 × 10−13.56 × 10−2 ± 2.42 × 10−13.68 × 10−4 ± 8.83 × 10−40.00 ± 0.00
101.77 × 103 ± 2.69 × 1021.70 × 103 ± 4.00 × 1021.69 × 103 ± 2.11 × 1025.92 × 102 ± 2.36 × 1022.56 × 102 ± 1.24 × 1024.55 × 102 ± 1.93 × 102
111.12 × 103 ± 6.181.10 × 103 ± 8.601.12 × 103 ± 6.321.32 × 101 ± 6.381.51 ± 9.92 × 10−11.67 ± 7.10 × 10−1
128.05 × 103 ± 1.29 × 1045.60 × 105 ± 6.30 × 1051.43 × 104 ± 1.39 × 1045.79 × 104 ± 4.39 × 1043.42 × 101 ± 4.12 × 1011.12 × 101 ± 9.45
134.96 × 103 ± 7.30 × 1031.10 × 104 ± 5.70 × 1036.08 × 103 ± 6.93 × 1034.92 × 103 ± 6.76 × 1035.09 ± 2.972.65 ± 2.18
141.42 × 103 ± 2.02 × 1011.50 × 103 ± 1.90 × 1011.43 × 103 ± 2.49 × 1018.09 × 101 ± 3.66 × 1012.49 ± 1.782.78 ± 9.25 × 10−1
151.51 × 103 ± 6.901.60 × 103 ± 7.40 × 1011.65 × 103 ± 4.16 × 1024.40 × 102 ± 3.34 × 1025.33 × 10−1 ± 3.54 × 10−11.47 × 10−1 ± 8.60 × 10−2
161.70 × 103 ± 1.17 × 1021.80 × 103 ± 1.40 × 1021.63 × 103 ± 4.81 × 1015.18 × 101 ± 6.35 × 1011.86 ± 6.26 × 10−11.12 ± 4.70 × 10−1
171.76 × 103 ± 6.04 × 1011.70 × 103 ± 1.20 × 1011.74 × 103 ± 2.17 × 1013.76 × 101 ± 2.94 × 1015.21 ± 2.301.25 × 101 ± 9.02 × 101
181.83 × 103 ± 9.113.00 × 104 ± 1.30 × 1047.20 × 103 ± 7.51 × 1031.02 × 104 ± 9.11 × 1031.88 ± 1.103.04 × 10−1 ± 1.85 × 10−1
193.45 × 103 ± 8.352.00 × 103 ± 5.60 × 1022.10 × 103 ± 7.50 × 1024.68 × 102 ± 9.55 × 1027.92 × 10−1 ± 3.88 × 10−11.39 × 10−1 ± 5.66 × 10−2
202.06 × 103 ± 5.02 × 1012.10 × 103 ± 6.00 × 1012.03 × 103 ± 3.05 × 1011.00 × 101 ± 1.07 × 1014.21 ± 2.442.51 ± 1.14
212.23 × 103 ± 6.78 × 1012.20 × 103 ± 4.00 × 1012.30 × 103 ± 4.95 × 1011.67 × 102 ± 5.80 × 1019.80 × 101 ± 1.40 × 1011.00 × 102 ± 0.00
222.30 × 103 ± 1.002.30 × 103 ± 3.002.30 × 103 ± 2.02 × 1018.66 × 101 ± 3.13 × 1011.42 × 101 ± 3.23 × 1016.38 × 101 ± 4.73 × 101
232.62 × 103 ± 1.43 × 1012.60 × 103 ± 1.00 × 1012.62 × 103 ± 5.073.15 × 102 ± 6.132.75 × 102 ± 9.27 × 1012.76 × 102 ± 9.72 × 101
242.75 × 103 ± 1.04 × 1012.60 × 103 ± 1.00 × 1022.75 × 103 ± 3.58 × 1013.22 × 102 ± 6.89 × 1018.88 × 101 ± 3.13 × 1013.11 × 102 ± 7.44 × 101
252.91 × 103 ± 2.35 × 1012.90 × 103 ± 2.30 × 1012.93 × 103 ± 5.10 × 1014.03 × 102 ± 4.73 × 1011.94 × 102 ± 1.12 × 1023.97 × 102 ± 1.32 × 10−1
262.94 × 103 ± 4.97 × 1012.90 × 103 ± 7.70 × 1012.98 × 103 ± 2.48 × 1023.20 × 102 ± 1.91 × 1028.73 × 101 ± 1.01 × 1023.00 × 102 ± 0.00
273.10 × 103 ± 5.623.10 × 103 ± 1.70 × 1013.10 × 103 ± 7.533.92 × 102 ± 5.243.82 × 102 ± 4.52 × 1013.89 × 102 ± 1.98 × 10−1
283.17 × 103 ± 6.95 × 10−133.30 × 103 ± 1.20 × 1023.29 × 103 ± 1.20 × 1023.70 × 102 ± 1.15 × 1021.76 × 102 ± 1.42 × 1023.00 × 102 ± 0.00
293.68 × 103 ± 3.40 × 1013.20 × 103 ± 2.50 × 1013.20 × 103 ± 2.53 × 1012.68 × 102 ± 2.57 × 1012.45 × 102 ± 1.27 × 1012.38 × 102 ± 3.33
305.08 × 107 ± 6.10 × 10−85.10 × 105 ± 1.60 × 1053.32 × 105 ± 3.68 × 1051.26 × 105 ± 3.51 × 1053.97 × 102 ± 6.674.29 × 102 ± 2.09 × 101
Table 14. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 30 on CEC 2017.
Table 14. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 30 on CEC 2017.
Fun.IADE [50]GJO-JOS [51]QCSCA [52]cSM [53]FOWFO [54]ATMALS-GSK
18.49 × 103 ± 6.38 × 1032.10 × 105 ± 1.10 × 1053.39 × 103 ± 3.66 × 1032.30 × 103 ± 3.01 × 1038.75 × 10−4 ± 3.89 × 10−40.00 ± 0.00
33.00 × 102 ± 7.95 × 10−16.40 × 102 ± 2.50 × 1023.09 × 102 ± 2.95 × 1019.89 × 10−10 ± 1.17× 10−94.53 × 10−6 ± 3.42 × 10−60.00 ± 0.00
41.94 × 104 ± 0.005.00 × 102 ± 1.70 × 1014.07 × 102 ± 9.728.07 × 101 ± 1.02 × 1015.43 ± 1.52 × 1011.99 ± 2.10
57.93 × 102 ± 1.47 × 1016.00 × 102 ± 5.60 × 1015.17 × 102 ± 4.788.61 × 101 ± 2.30 × 1015.49 × 101 ± 1.39 × 1011.53 × 101 ± 1.90 × 101
66.55 × 102 ± 4.636.10 × 102 ± 5.906.00 × 102 ± 3.52 × 10−24.13 × 10−4 ± 8.21 × 10−44.61 × 10−2 ± 3.07 × 10−26.82 × 10−5 ± 3.90 × 10−5
79.04 × 102 ± 7.38 × 1018.30 × 102 ± 6.20 × 1017.27 × 102 ± 6.061.33 × 102 ± 2.61 × 1017.91 × 101 ± 1.13 × 1014.33 × 101 ± 1.15 × 101
89.24 × 102 ± 3.24 × 1018.90 × 102 ± 4.60 × 1018.14 × 102 ± 4.321.01 × 102 ± 2.82 × 1015.60 × 101 ± 1.19 × 1011.35 × 101 ± 1.29 × 101
95.44 × 103 ± 4.41 × 1021.20 × 103 ± 4.40 × 1029.00 × 102 ± 4.28 × 10−39.32 × 102 ± 8.65 × 1021.90 × 101 ± 2.15 × 1011.79 × 10−2 ± 3.77 × 10−2
104.75 × 103 ± 1.12 × 1035.00 × 103 ± 1.70 × 1031.57 × 103 ± 2.19 × 1022.85 × 103 ± 4.80 × 1022.02 × 103 ± 3.11 × 1022.80 × 103 ± 5.90 × 102
111.22 × 103 ± 4.95 × 1011.20 × 103 ± 4.20 × 1011.11 × 103 ± 1.861.61 × 102 ± 5.33 × 1011.26 × 101 ± 4.651.46 × 101 ± 1.73 × 101
125.57 × 104 ± 4.87 × 1041.20 × 107 ± 1.10 × 1071.80 × 104 ± 1.44 × 1048.87 × 105 ± 8.57 × 1052.92 × 102 ± 1.78 × 1023.84 × 103 ± 2.62 × 103
131.57 × 104 ± 1.84 × 1041.00 × 105 ± 7.50 × 1044.94 × 103 ± 6.23 × 1032.17 × 104 ± 2.35 × 1044.43 × 101 ± 2.89 × 1013.01 × 101 ± 1.12 × 101
141.64 × 103 ± 8.18 × 1013.80 × 104 ± 3.00 × 1041.42 × 103 ± 2.20 × 1011.01 × 104 ± 7.92 × 1032.91 × 101 ± 7.911.78 × 101 ± 4.26
156.84 × 103 ± 1.01 × 1043.40 × 104 ± 2.20 × 1041.53 × 103 ± 3.46 × 1011.79 × 104 ± 1.39 × 1041.27 × 101 ± 4.031.43 × 101 ± 5.56
162.37 × 103 ± 2.31 × 1022.50 × 103 ± 3.60 × 1021.61 × 103 ± 3.896.87 × 102 ± 2.37 × 1024.08 × 102 ± 1.68 × 1022.53 × 102 ± 2.81 × 102
172.08 × 103 ± 1.11 × 1022.00 × 103 ± 1.40 × 1021.73 × 103 ± 1.99 × 1012.83 × 102 ± 1.70 × 1028.79 × 101 ± 4.18 × 1014.71 × 101 ± 2.30 × 101
181.47 × 104 ± 1.93 × 1043.20 × 105 ± 3.30 × 1056.49 × 103 ± 9.86 × 1032.02 × 105 ± 1.89 × 1052.67 × 101 ± 4.823.02 × 101 ± 6.67
197.20 × 103 ± 4.67 × 1033.40 × 105 ± 3.20 × 1051.91 × 103 ± 9.972.30 × 104 ± 2.07 × 1041.36 × 101 ± 3.531.30 × 101 ± 3.34
202.53 × 103 ± 2.27 × 1022.30 × 103 ± 1.10 × 1022.01 × 103 ± 9.033.75 × 102 ± 1.45 × 1021.12 × 102 ± 6.64 × 1014.74 × 101 ± 3.35 × 101
212.38 × 103 ± 2.88 × 1012.40 × 103 ± 5.70 × 1012.30 × 103 ± 4.74 × 1012.86 × 102 ± 2.26 × 1011.96 × 102 ± 7.84 × 1012.66 × 102 ± 9.12
224.08 × 103 ± 2.49 × 1032.30 × 103 ± 1.302.30 × 103 ± 1.89 × 1012.61 × 103 ± 1.27 × 1039.93 × 101 ± 5.191.00 × 102 ± 0.00
232.73 × 103 ± 2.00 × 1012.80 × 103 ± 5.30 × 1012.62 × 103 ± 2.934.35 × 102 ± 2.61 × 1012.64 × 102 ± 1.47 × 1023.44 × 102 ± 1.70 × 101
242.94 × 103 ± 5.02 × 1012.90 × 103 ± 5.10 × 1012.72 × 103 ± 8.70 × 1014.96 × 102 ± 6.26 × 1013.64 × 102 ± 1.77 × 1024.75 × 102 ± 1.48 × 101
252.90 × 103 ± 9.482.90 × 103 ± 1.20 × 1012.93 × 103 ± 2.24 × 1013.86 × 102 ± 1.213.85 × 102 ± 1.643.86 × 102 ± 1.07
265.56 × 103 ± 6.13 × 1023.40 × 103 ± 8.50 × 1022.90 × 103 ± 1.53 × 1021.95 × 103 ± 4.32 × 1022.55 × 102 ± 4.88 × 1011.79 × 103 ± 1.93 × 102
273.24 × 103 ± 1.34 × 1013.20 × 103 ± 1.60 × 1013.09 × 103 ± 7.505.17 × 102 ± 9.774.97 × 102 ± 1.22 × 1014.96 × 102 ± 6.31
283.54 × 103 ± 1.51 × 10−123.20 × 103 ± 2.70 × 1013.23 × 103 ± 1.26 × 1023.63 × 102 ± 5.49 × 1013.00 × 102 ± 2.34 × 10−23.10 × 102 ± 3.26 × 101
297.34 × 103 ± 2.95 × 1023.70 × 103 ± 1.60 × 1023.18 × 103 ± 1.95 × 1017.81 × 102 ± 1.72 × 1024.97 × 102 ± 5.92 × 1014.32 × 102 ± 5.95 × 101
302.65 × 109 ± 4.42 × 10−62.00 × 106 ± 9.60 × 1051.18 × 105 ± 2.66 × 1052.26 × 104 ± 1.53 × 1041.98 × 103 ± 5.31 × 1012.06 × 103 ± 5.55 × 101
Table 15. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 50 on CEC 2017.
Table 15. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 50 on CEC 2017.
Fun.IADE [50]GJO-JOS [51]QCSCA [52]cSM [53]FOWFO [54]ATMALS-GSK
13.25 × 103 ± 4.46 × 1037.80 × 105 ± 4.50 × 1053.02 × 103 ± 3.12 × 1035.99 × 103 ± 6.65 × 1031.78 × 10−2 ± 1.37 × 10−21.25 × 10−1 ± 2.15 × 10−1
32.46 × 103 ± 3.21 × 1033.90 × 103 ± 1.90 × 1033.02 × 102 ± 3.331.76 × 10−7 ± 1.72 × 10−72.24 ± 3.727.85 × 102 ± 6.73 × 102
43.12 × 104 ± 1.24 × 10−115.80 × 102 ± 5.10 × 1014.05 × 102 ± 2.281.13 × 102 ± 3.89 × 1012.15 × 101 ± 1.73 × 1019.80 ± 1.82 × 101
58.88 × 102 ± 3.44 × 1016.90 × 102 ± 8.60 × 1015.15 × 102 ± 3.681.89 × 102 ± 4.15 × 1011.17 × 102 ± 1.94 × 1013.21 × 101 ± 1.27 × 101
66.62 × 102 ± 3.926.20 × 102 ± 1.20 × 1016.00 × 102 ± 1.40 × 10−21.75 × 10−4 ± 4.74 × 10−41.01 × 10−1 ± 4.49 × 10−27.14 × 10−6 ± 8.31 × 10−6
71.23 × 103 ± 7.64 × 1019.80 × 102 ± 1.20 × 1027.25 × 102 ± 4.682.48 × 102 ± 4.20 × 1011.45 × 102 ± 1.80 × 1011.08 × 102 ± 1.12 × 101
81.09 × 103 ± 4.66 × 1011.00 × 103 ± 9.00 × 1018.12 × 102 ± 3.461.85 × 102 ± 3.01 × 1011.18 × 102 ± 2.37 × 1013.85 × 101 ± 9.92
91.31 × 104 ± 5.76 × 1029.00 × 103 ± 7.00 × 1039.00 × 102 ± 3.31 × 10−44.67 × 103 ± 2.57 × 1033.95 × 102 ± 2.83 × 1021.91 × 10−1 ± 2.06 × 10−1
107.98 × 103 ± 1.36 × 1038.20 × 103 ± 3.00 × 1031.46 × 103 ± 1.49 × 1025.05 × 103 ± 6.73 × 1023.84 × 103 ± 4.36 × 1028.80 × 103 ± 9.25 × 102
111.32 × 103 ± 6.22 × 1011.40 × 103 ± 6.80 × 1011.11 × 103 ± 1.312.26 × 102 ± 5.19 × 1015.21 × 101 ± 1.41 × 1012.91 × 101 ± 5.17
129.99 × 105 ± 5.62 × 1054.80 × 107 ± 2.50 × 1072.00 × 104 ± 1.69 × 1043.89 × 106 ± 2.38 × 1062.83 × 103 ± 2.42 × 1033.65 × 104 ± 2.12 × 104
138.85 × 103 ± 8.49 × 1031.30 × 105 ± 6.50 × 1043.48 × 103 ± 3.45 × 1031.62 × 104 ± 1.37 × 1041.04 × 102 ± 4.79 × 1011.19 × 102 ± 5.17 × 101
145.27 × 103 ± 2.38 × 1031.90 × 105 ± 1.00 × 1051.43 × 103 ± 2.19 × 1015.51 × 104 ± 4.25 × 1045.83 × 101 ± 1.03 × 1013.69 × 101 ± 9.02
151.00 × 104 ± 9.72 × 1035.00 × 104 ± 2.60 × 1041.52 × 103 ± 2.87 × 1011.44 × 104 ± 9.47 × 1034.33 × 101 ± 9.084.45 × 101 ± 1.14 × 101
163.07 × 103 ± 4.27 × 1022.90 × 103 ± 4.50 × 1021.61 × 103 ± 4.96 × 1011.44 × 104 ± 3.69 × 1029.73 × 102 ± 2.25 × 1024.91 × 102 ± 3.35 × 102
173.27 × 103 ± 3.10 × 1022.80 × 103 ± 4.30 × 1021.72 × 103 ± 1.87 × 1011.44 × 104 ± 2.87 × 1026.46 × 102 ± 1.76 × 1027.08 × 102 ± 2.31 × 102
181.12 × 105 ± 8.98 × 1041.50 × 106 ± 7.80 × 1053.60 × 103 ± 3.85 × 1031.44 × 104 ± 1.11 × 1054.18 × 101 ± 5.981.17 × 102 ± 7.30 × 101
191.67 × 108 ± 4.01 × 1015.00 × 105 ± 3.40 × 1051.90 × 103 ± 6.171.44 × 104 ± 1.71 × 1043.05 × 101 ± 7.382.83 × 101 ± 7.93
203.39 × 103 ± 2.65 × 1022.80 × 103 ± 3.60 × 1022.00 × 103 ± 3.251.44 × 104 ± 1.80 × 1025.07 × 102 ± 1.35 × 1026.91 × 102 ± 1.48 × 102
212.52 × 103 ± 4.60 × 1012.50 × 103 ± 8.70 × 1012.29 × 103 ± 4.82 × 1011.44 × 104 ± 4.00 × 1013.15 × 102 ± 3.84 × 1011.98 × 102 ± 6.81
221.01 × 104 ± 7.53 × 1025.80 × 103 ± 4.20 × 1032.30 × 103 ± 1.97 × 1011.44 × 104 ± 7.06 × 1023.82 × 103 ± 1.67 × 1038.71 × 103 ± 2.41 × 103
232.97 × 103 ± 3.90 × 1012.90 × 103 ± 5.40 × 1012.61 × 103 ± 2.881.44 × 104 ± 3.65 × 1015.40 × 102 ± 8.26 × 1014.39 × 102 ± 1.27 × 101
243.13 × 103 ± 3.48 × 1013.10 × 103 ± 6.90 × 1012.72 × 103 ± 8.02 × 1011.44 × 104 ± 4.28 × 1017.02 × 102 ± 1.12 × 1025.14 × 102 ± 1.35 × 101
253.09 × 103 ± 1.91 × 1013.10 × 103 ± 3.60 × 1012.92 × 103 ± 2.34 × 1011.44 × 104 ± 3.34 × 1014.89 × 102 ± 3.03 × 1015.45 × 102 ± 2.74 × 101
268.18 × 103 ± 1.53 × 1032.90 × 103 ± 3.30 × 1022.90 × 103 ± 1.41 × 1021.44 × 104 ± 3.58 × 1026.14 × 102 ± 8.00 × 1021.35 × 103 ± 3.55 × 102
273.62 × 103 ± 1.34 × 1023.40 × 103 ± 6.30 × 1013.09 × 103 ± 7.061.44 × 104 ± 4.77 × 1015.50 × 102 ± 3.24 × 1015.71 × 102 ± 2.78 × 101
283.88 × 103 ± 2.03 × 1013.30 × 103 ± 4.10 × 1013.23 × 103 ± 1.48 × 1021.44 × 104 ± 1.70 × 1014.60 × 102 ± 4.294.88 × 102 ± 2.39 × 101
291.55 × 105 ± 6.22 × 1024.20 × 103 ± 2.90 × 1023.17 × 103 ± 1.66 × 1011.44 × 104 ± 2.58 × 1026.76 × 102 ± 1.60 × 1025.01 × 102 ± 4.19 × 101
306.21 × 109 ± 7.78 × 10−52.00 × 107 ± 5.20 × 1061.79 × 105 ± 3.34 × 1051.44 × 104 ± 3.35 × 1065.97 × 105 ± 3.14 × 1045.92 × 105 ± 1.85 × 104
Table 16. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 100 on CEC 2017.
Table 16. Comparison of the average and standard deviation results (Mean ± STD) of ATMALS-GSK with existing metaheuristics over 51 runs for D = 100 on CEC 2017.
Fun.IADE [50]GJO-JOS [51]QCSCA [52]cSM [53]FOWFO [54]ATMALS-GSK
19.46 × 104 ± 1.85 × 1053.90 × 106 ± 1.10 × 1063.27 × 103 ± 3.72 × 1038.14 × 103 ± 1.28 × 1041.88 × 103 ± 2.03 × 1032.55 ± 4.79
33.68 × 104 ± 8.20 × 1035.20 × 104 ± 1.10 × 1043.00 × 102 ± 3.85 × 10−14.83 × 103 ± 2.26 × 1034.23 × 104 ± 9.62 × 1034.91 × 104 ± 1.26 × 104
47.98 × 104 ± 1.23 × 10−107.70 × 102 ± 5.20 × 1014.04 × 102 ± 2.272.31 × 102 ± 2.72 × 1011.52 × 102 ± 4.63 × 1014.90 ± 1.45 × 101
51.41 × 103 ± 3.44 × 1011.00 × 103 ± 1.70 × 1025.11 × 102 ± 2.325.00 × 102 ± 6.65 × 1013.11 × 102 ± 5.11 × 1019.70 × 101 ± 2.17 × 101
66.62 × 102 ± 1.186.40 × 102 ± 9.706.00 × 102 ± 3.58 × 10−36.53 × 10−5 ± 1.06 × 10−42.96 × 10−1 ± 8.85 × 10−21.24 × 10−2 ± 2.73 × 10−2
72.59 × 103 ± 2.92 × 1021.50 × 103 ± 2.30 × 1027.23 × 102 ± 3.627.07 × 102 ± 7.90 × 1013.43 × 102 ± 3.11 × 1011.68 × 102 ± 1.91 × 101
81.66 × 103 ± 1.28 × 1021.30 × 103 ± 1.80 × 1028.09 × 102 ± 2.705.19 × 102 ± 7.77 × 1013.28 × 102 ± 4.38 × 1019.65 × 101 ± 1.12 × 101
92.39 × 104 ± 7.15 × 1024.50 × 104 ± 1.60 × 1049.00 × 102 ± 1.88 × 10−51.84 × 104 ± 5.08 × 1037.71 × 103 ± 2.00 × 1031.05 × 101 ± 3.14
101.57 × 104 ± 1.19 × 1031.60 × 104 ± 5.70 × 1031.28 × 103 ± 1.05 × 1021.18 × 104 ± 1.00 × 1031.05 × 104 ± 7.56 × 1022.44 × 104 ± 1.01 × 103
113.40 × 103 ± 1.16 × 1032.80 × 103 ± 2.90 × 1021.10 × 103 ± 8.65 × 10−11.29 × 103 ± 2.17 × 1022.70 × 102 ± 5.01 × 1011.76 × 102 ± 6.28 × 101
121.07 × 107 ± 2.86 × 1062.70 × 108 ± 1.20 × 1081.55 × 104 ± 1.43 × 1041.27 × 107 ± 4.80 × 1061.36 × 105 ± 6.66 × 1043.78 × 105 ± 1.69 × 105
131.02 × 104 ± 3.42 × 1034.00 × 105 ± 8.70 × 1052.96 × 103 ± 2.86 × 1031.29 × 104 ± 1.32 × 1043.49 × 102 ± 1.10 × 1021.31 × 102 ± 5.43 × 101
142.31 × 105 ± 1.40 × 1051.00 × 106 ± 4.50 × 1051.42 × 103 ± 2.04 × 1011.67 × 105 ± 7.05 × 1042.10 × 102 ± 3.31 × 1012.57 × 103 ± 2.31 × 103
154.79 × 103 ± 2.66 × 1036.40 × 104 ± 2.90 × 1041.42 × 103 ± 2.04 × 1018.26 × 103 ± 8.01 × 1031.92 × 102 ± 6.45 × 1011.04 × 102 ± 4.05 × 101
165.59 × 103 ± 6.96 × 1025.40 × 103 ± 5.50 × 1021.61 × 103 ± 4.813.39 × 103 ± 6.43 × 1022.87 × 103 ± 3.97 × 1021.89 × 103 ± 4.24 × 102
174.86 × 103 ± 7.32 × 1024.60 × 103 ± 5.00 × 1021.72 × 103 ± 2.08 × 1012.81 × 103 ± 4.78 × 1022.07 × 103 ± 3.26 × 1021.57 × 103 ± 6.92 × 102
185.51 × 105 ± 2.28 × 1052.10 × 106 ± 9.40 × 1052.02 × 103 ± 8.45 × 1023.66 × 105 ± 1.29 × 1053.90 × 102 ± 2.46 × 1021.14 × 104 ± 4.25 × 103
192.82 × 109 ± 1.06 × 1021.70 × 106 ± 9.60 × 1051.90 × 103 ± 9.141.17 × 104 ± 1.16 × 1041.10 × 102 ± 2.40 × 1011.61 × 102 ± 7.63 × 101
205.27 × 103 ± 6.49 × 1024.60 × 103 ± 1.00 × 1032.00 × 103 ± 2.19 × 10−12.66 × 103 ± 4.39 × 1021.97 × 103 ± 2.92 × 1022.84 × 103 ± 5.21 × 102
212.95 × 103 ± 1.46 × 1022.80 × 103 ± 1.50 × 1022.29 × 103 ± 4.70 × 1017.33 × 102 ± 7.54 × 1015.46 × 102 ± 4.89 × 1012.89 × 102 ± 2.11 × 101
221.83 × 104 ± 1.13 × 1031.30 × 104 ± 9.10 × 1032.30 × 103 ± 8.75 × 10−31.32 × 104 ± 1.24 × 1031.18 × 104 ± 8.06 × 1022.43 × 104 ± 1.92 × 103
233.47 × 103 ± 9.11 × 1013.30 × 103 ± 1.40 × 1022.61 × 103 ± 2.659.18 × 102 ± 6.16 × 1017.72 × 102 ± 3.13 × 1015.87 × 102 ± 1.45 × 101
244.07 × 103 ± 1.09 × 1023.80 × 103 ± 1.90 × 1022.71 × 103 ± 7.38 × 1011.40 × 103 ± 7.62 × 1011.25 × 103 ± 5.48 × 1019.11 × 102 ± 1.77 × 101
253.38 × 103 ± 6.40 × 1013.50 × 103 ± 4.90 × 1012.92 × 103 ± 2.34 × 1017.09 × 102 ± 6.05 × 1017.30 × 102 ± 5.31 × 1017.53 × 102 ± 4.18 × 101
261.87 × 104 ± 3.28 × 1036.20 × 103 ± 4.20 × 1032.92 × 103 ± 2.34 × 1018.62 × 103 ± 8.60 × 1026.37 × 103 ± 1.61 × 1033.42 × 103 ± 1.28 × 103
273.92 × 103 ± 2.80 × 1023.50 × 103 ± 5.90 × 1013.10 × 103 ± 1.19 × 1017.47 × 102 ± 4.96 × 1016.85 × 102 ± 3.88 × 1016.87 × 102 ± 3.47 × 101
285.46 × 103 ± 1.54 × 1023.50 × 103 ± 5.10 × 1013.20 × 103 ± 1.26 × 1025.53 × 102 ± 3.30 × 1015.55 × 102 ± 2.73 × 1013.51 × 102 ± 1.10 × 102
295.20 × 104 ± 9.53 × 1026.90 × 103 ± 8.90 × 1023.16 × 103 ± 1.40 × 1013.95 × 103 ± 5.28 × 1022.67 × 103 ± 2.49 × 1021.39 × 103 ± 2.14 × 102
302.06 × 1010 ± 1.98 × 10−12.10 × 107 ± 6.10 × 1061.70 × 105 ± 3.26 × 1051.78 × 104 ± 6.67 × 1032.85 × 103 ± 2.31 × 1023.23 × 103 ± 8.70 × 102
Table 17. Wilcoxon signed-rank test of the ATMALS-GSK algorithm with the GSK variants.
Table 17. Wilcoxon signed-rank test of the ATMALS-GSK algorithm with the GSK variants.
DMethods R+R−p Value+=Dec.
10ATMALS-GSK vsGSK238150.00031973+
AGSK1341910.443211414
APGSK1381870.509813412
FDBAGSK972030.13008516
eGSK152790.204511810
30ATMALS-GSK vsGSK337410.00032225+
AGSK378280.00012612+
APGSK379270.00002711+
FDBAGSK340380.00022324+
eGSK2441070.08181838
50ATMALS-GSK vsGSK375600.00032306+
AGSK354810.00232504+
APGSK356790.00192504+
FDBAGSK3141210.03652108+
eGSK3301050.01382405+
100ATMALS-GSK vsGSK378.5560.00022306+
AGSK368670.00062405+
APGSK383520.00012603+
FDBAGSK394410.00002504+
eGSK323.5111.50.02152306+
Table 18. Friedman test of the ATMALS-GSK algorithm with the GSK variants.
Table 18. Friedman test of the ATMALS-GSK algorithm with the GSK variants.
MethodRanksMean RankOverall Rank
D = 10D = 30D = 50D = 100
ATMALS-GSK3.371.841.931.822.241
eGSK3.532.242.762.862.842
FDBAGSK2.813.663.093.823.343
AGSK2.874.364.413.933.894
GSK4.864.164.283.584.225
APGSK3.534.744.534.964.446
p-value0.00030.00010.00000.00000.0000
Table 19. Wilcoxon signed-rank test of the ATMALS-GSK algorithm with existing metaheuristics.
Table 19. Wilcoxon signed-rank test of the ATMALS-GSK algorithm with existing metaheuristics.
DMethods R+R−p Value+=Dec.
10ATMALS-GSK vs.IADE43500.00002900+
GJO-JOS43500.00002900+
QCSCA43500.00002900+
cSM43500.00002900+
FOWFO1432920.110414015
30ATMALS-GSK vs.IADE43500.00002900+
GJO-JOS43500.00002900+
QCSCA425100.00002801+
cSM40600.00002810+
FOWFO205.5229.50.798316013
50ATMALS-GSK vs.IADE43320.00002801+
GJO-JOS415200.00002702+
QCSCA3221130.02292405+
cSM392430.00002603+
FOWFO2172181.0015014
100ATMALS-GSK vs.IADE383520.00012603+
GJO-JOS396390.00002702+
QCSCA2761590.21322009
cSM360750.00132306+
FOWFO2361990.701318011
Table 20. Friedman test of the ATMALS-GSK algorithm with existing metaheuristics.
Table 20. Friedman test of the ATMALS-GSK algorithm with existing metaheuristics.
MethodRanksMean RankOverall Rank
D = 10D = 30D = 50D = 100
ATMALS-GSK1.521.501.862.031.721
FOWFO1.551.651.722.071.752
cSM3.103.474.483.413.613
QCSCA5.003.783.022.863.664
GJO-JOS4.8.35.314.845.215.045
IADE5.005.295.075.415.196
p-value0.00000.00000.00000.00000.0000
Table 21. Comparison of the average results of ATMALS-GSK with original GSK and five ATMALS-GSK versions over 51 runs for D = 30 on CEC 2017.
Table 21. Comparison of the average results of ATMALS-GSK with original GSK and five ATMALS-GSK versions over 51 runs for D = 30 on CEC 2017.
Fun.GSKATMALS-GSK-1ATMALS-GSK-2ATMALS-GSK-3ATMALS-GSK-4ATMALS-GSK-5ATMALS-GSK
10.000.000.000.000.000.000.00
37.07 × 10−70.005.03 × 10−74.28 × 10−71.75 × 10−70.000.00
41.11 × 1014.329.321.18 × 1017.933.271.99
51.60 × 1026.42 × 1014.29 × 1011.13 × 1016.11 × 1012.82 × 1011.53 × 101
61.52 × 10−61.37 × 10−69.86 × 10−54.80 × 10−57.12 × 10−51.12 × 10−66.82 × 10−5
71.87 × 1028.44 × 1016.93 × 1018.33 × 1011.63 × 1028.27 × 1014.33 × 101
81.55 × 1021.34 × 1024.28 × 1017.30 × 1015.25 × 1011.35 × 1011.35 × 101
90.006.52 × 10−10.002.58 × 10−35.22 × 10−30.001.79 × 10−2
106.69 × 1035.23 × 1034.22 × 1036.88 × 1036.48 × 1035.40 × 1032.80 × 103
113.30 × 1012.91 × 1013.55 × 1013.08 × 1012.01 × 1012.32 × 1011.46 × 101
126.62 × 1036.82 × 1036.04 × 1036.22 × 1035.86 × 1035.14 × 1033.84 × 103
139.83 × 1017.23 × 1016.45 × 1016.38 × 1018.91 × 1017.40 × 1013.01 × 101
145.69 × 1013.13 × 1017.48 × 1016.11 × 1015.22 × 1014.17 × 1011.78 × 101
151.45 × 1011.41 × 1011.48 × 1011.52 × 1011.50 × 1011.42 × 1011.43 × 101
167.96 × 1024.33 × 1027.12 × 1026.53 × 1025.44 × 1024.67 × 1022.53 × 102
171.89 × 1021.05 × 1021.78 × 1028.31 × 1012.22 × 1027.99 × 1014.71 × 101
183.68 × 1013.55 × 1023.81 × 1013.39 × 1013.97 × 1013.45 × 1013.02 × 101
191.29 × 1011.28 × 1011.45 × 1011.34 × 1011.38 × 1011.27 × 1011.30 × 101
201.08 × 1029.62 × 1018.77 × 1017.30 × 1017.75 × 1018.91 × 1014.74 × 101
213.46 × 1023.88 × 1023.29 × 1023.14 × 1023.01 × 1022.98 × 1022.66 × 102
221.00 × 1021.00 × 1021.00 × 1021.00 × 1021.00 × 1021.00 × 1021.00 × 102
234.70 × 1024.48 × 1024.96 × 1024.50 × 1024.04 × 1024.10 × 1023.44 × 102
245.68 × 1025.24 × 1025.75 × 1025.33 × 1025.59 × 1025.21 × 1024.75 × 102
253.87 × 1023.88 × 1023.86 × 1023.87 × 1023.89 × 1023.86 × 1023.86 × 102
269.87 × 1029.77 × 1021.23 × 1031.55 × 1032.26 × 1039.54 × 1021.79 × 103
274.93 × 1025.03 × 1025.38 × 1026.06 × 1024.91 × 1024.97 × 1024.96 × 102
283.21 × 1023.17 × 1023.24 × 1023.15 × 1023.20 × 1023.16 × 1023.10 × 102
295.77 × 1025.42 × 1026.37 × 1025.58 × 1025.25 × 1025.12 × 1024.32 × 102
302.08 × 1032.07 × 1032.08 × 1032.15 × 1032.12 × 1032.07 × 1032.06 × 103
Table 22. Wilcoxon signed-rank test of the ATMALS-GSK algorithm with the original GSK and five ATMALS-GSK versions over 51 runs for D = 30 on CEC 2017.
Table 22. Wilcoxon signed-rank test of the ATMALS-GSK algorithm with the original GSK and five ATMALS-GSK versions over 51 runs for D = 30 on CEC 2017.
Methods R+R−p Value+=Dec.
ATMALS-GSK vs.GSK337410.00032225+
ATMALS-GSK-1 321300.00022234+
ATMALS-GSK-2324270.00022432+
ATMALS-GSK-3341370.00032324+
ATMALS-GSK-4363150.00002423+
ATMALS-GSK-5268320.00071955+
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alfadli, N.M.; Oun, E.M.; Mohamed, A.W. Auto-Tuning Memory-Based Adaptive Local Search Gaining–Sharing Knowledge-Based Algorithm for Solving Optimization Problems. Algorithms 2025, 18, 398. https://doi.org/10.3390/a18070398

AMA Style

Alfadli NM, Oun EM, Mohamed AW. Auto-Tuning Memory-Based Adaptive Local Search Gaining–Sharing Knowledge-Based Algorithm for Solving Optimization Problems. Algorithms. 2025; 18(7):398. https://doi.org/10.3390/a18070398

Chicago/Turabian Style

Alfadli, Nawaf Mijbel, Eman Mostafa Oun, and Ali Wagdy Mohamed. 2025. "Auto-Tuning Memory-Based Adaptive Local Search Gaining–Sharing Knowledge-Based Algorithm for Solving Optimization Problems" Algorithms 18, no. 7: 398. https://doi.org/10.3390/a18070398

APA Style

Alfadli, N. M., Oun, E. M., & Mohamed, A. W. (2025). Auto-Tuning Memory-Based Adaptive Local Search Gaining–Sharing Knowledge-Based Algorithm for Solving Optimization Problems. Algorithms, 18(7), 398. https://doi.org/10.3390/a18070398

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop