Next Article in Journal
LW-YOLO: A Lightweight Enhanced YOLOv8-Based Model for Mosquito Detection
Previous Article in Journal
A Feature Selection Method for Yarn Quality Prediction Based on SHAP Interpretation
Previous Article in Special Issue
Metaheuristic Algorithms in Optimal Design of Engineering Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Surrogate-Assisted Slime Mould Algorithm Considering a Dual-Based Merit Criterion for Global Database Management

Instituto de Telecomunicações and University of Beira Interior, Calçada Fonte do Lameiro, 6201-001 Covilhã, Portugal
*
Author to whom correspondence should be addressed.
Algorithms 2026, 19(4), 265; https://doi.org/10.3390/a19040265
Submission received: 31 January 2026 / Revised: 23 March 2026 / Accepted: 24 March 2026 / Published: 1 April 2026

Abstract

Metaheuristic algorithms, including evolutionary approaches, are vital for solving non-trivial and non-convex optimization problems. However, real-world engineering often involves high-dimensional, expensive problems that deteriorate performance due to the substantial amount of required fitness evaluations. To address this, a growing trend utilizes evolutionary algorithms assisted by surrogate models, which limit the computational burden by providing alternatives to expensive evaluations. Leveraging the exploration capabilities of the recently developed Slime Mould Algorithm—a metaheuristic with only one tuning parameter that ignores personal best information—this work develops its surrogate-assisted counterpart: the Surrogate-Assisted Slime Mould Algorithm (SASMA). This new approach features an original database management strategy and surrogate building mechanism. To confirm its effectiveness and versatility, SASMA is tested on benchmark mathematical functions for 30 and 100 dimensions, as well as a classical truss design problem, against several surrogate-assisted and metaheuristic algorithms. The proposed SASMA achieved statistically significant improvements in both case studies, outperforming the selected benchmark algorithms on most test functions.

1. Introduction

The proliferation of population-based metaheuristic optimization algorithms, a type of stochastic optimization, and their application has been a trend in mathematical and computer optimization for the last few decades [1]. The working principle of these algorithms is to explore subsets of the search space domain, which is otherwise too large to be fully surveyed, using a population of potential solutions rather than a single candidate solution, guided by a series of heuristics that hopefully lead the search towards an optimal solution with a lower computational cost than other traditional optimization algorithms, like iterative methods [2]. Furthermore, unlike many traditional methods, metaheuristic algorithms can also handle non-continuous problems. Nevertheless, depending on the problem complexity and the specific algorithm search space mechanisms, namely the corresponding convergence properties, there is no guarantee that the global optimal solution(s) can always be found [3].
In this crowded field, swarm intelligence (SI) and evolutionary algorithms (EAs) are among the most popular choices given their ability to improve several candidate solutions in a decentralized manner, mirroring diversified types of cooperative strategies observed in nature [4]. In addition, many of these population-based metaheuristics rely on a set of randomly generated variables (within the set of heuristics/strategies) [5]. This stochastic character (randomness) has proven to be effective in the search space task by avoiding an early concentration of search agents in the same space regions, therefore mitigating the possibilities of being trapped in local optima. From the wide range of options, particle swarm optimization (PSO) [6], cuckoo search (CS) [7], differential evolution (DE) [8], whale optimization algorithm (WOA) [9], and gray wolf optimization algorithm (GWO) [10] are amidst the preferred choices to solve many engineering problems. Another feature of these algorithms is the influence of different control parameters and the population size itself, which can greatly influence global search capabilities [11], and consequently, together with the stopping criteria, determine the convergence rate towards the optimal solution.
The search space is surveyed by the population of these metaheuristic algorithms by using mechanisms that balance exploration and exploitation of the domain space [12] until the desired stopping criteria is met, and it is typically defined by a fixed number of fitness evaluations (FEs). Still, the required number of FEs can be considerable: the algorithm heuristics actively explore different regions of space to avoid a premature convergence (exploration) before engaging in the exploitation of potential optimal solutions, thus incurring a significant computation cost when dealing with expensive optimization problems. Nonlinear and non-explicit equations, computational electromagnetics analysis, mechanical design, and finite element analysis, as well as fluid dynamics, are well-known instances of the numerous engineering problems that require costly fidelity simulations [13]. Hence, in these situations the use of metaheuristics that rely upon a sizable number of FEs can be unbearable, a problem known as the curse of dimensionality.
To lessen this time-consuming burden, an increasing amount of research works have been gradually shifting their attention towards the use of surrogate models, i.e., relatively computationally cheap approximated models (surrogates) that in turn replace expensive FEs using the real model/function [14]. These models work by approximating the output from a set of input data based on the behavior of a relatively complex system, yet given the deviation between the actual model and the surrogate model output, it is often necessary to run the original model in a few instances [15].
Accordingly, several surrogate-assisted evolutionary algorithms (SAEAs) have been developed [16,17], and these typically employ polynomial regression models [18]; Gaussian processes (GPs), also known as Kriging models [19,20,21]; support vector machines (SVMs) and support vector regression (SVRs) [22,23]; radial basis functions (RBFs) and radial basis function neural networks (RBFNNs) [24,25]; and other types of neural networks (ANNs) [26,27] to construct the surrogate model. The different traits of these approximation models are explored in a comprehensive manner. In [28], an SAEA is proposed where a more global model with faster convergence is achieved through an RBFNN, whereas a Kriging model is employed to obtain a more local model.
Other examples of SAEAs include a two-layer surrogate-assisted PSO, where by managing a database of historical (particle) positions, a global surrogate is employed in conjunction with several local surrogates to perform the model approximation, and RBFNNs are used to build these surrogates [13]. A similar two-level optimum search approach is followed in [29], where the water cycle algorithm is assisted by a hierarchical surrogate. In addition, a multi-RBF high-fidelity surrogate model is developed in [30] as a suitable option for expensive problems. A decision space partition, where different surrogates are built for different clusters of positions, is suggested in [31] as an effective global search strategy, which is then smoothly complemented with an adaptive selection strategy for the local search. Fuzzy logic has also been applied to this field, and a hierarchical surrogate with a probabilistic PSO search based on global and local assist models is proposed in [32]. The algorithm begins by clustering all precisely evaluated samples to divide the search space into meaningful regions. It then builds dedicated surrogate models within each region, enabling a more accurate representation of the overall fitness landscape and enhancing the algorithm’s exploratory power. Analogously, the generalized multi-factorial evolutionary algorithm was also used in a clustered-based approach, where all the true function evaluations divide the search space into meaningful regions. It then builds dedicated surrogate models within each region, enabling a more accurate representation of the overall fitness landscape and enhancing the algorithm’s exploratory power. Finally, an ensemble surrogate is then used to speed up convergence around the local minima [33].
An engineering application of SAEAs is shown in [34], where the wireless sensor network coverage problem is solved by employing an add-point strategy based on the information from historical surrogate models (RBFs), coupled with a restart strategy. With this methodology, the authors were able to select better candidate (promising) multidimensional positions to evaluate through the real objective function. Another issue concerns the scarcity of training data used to build the surrogate models associated with hard engineering problems; as such, the authors in [35] present a global transfer optimization framework, where similar information is inherited. While the opposite can also occur, for instance in [36], the data samples taken from the feasible design region are abundant, and so they are used to create three surrogate models, capturing the different fitness evaluation traits with an automated multiobjective surrogate-based Pareto finder.
Due to the fitting capabilities of radial basis functions and their fast training, RBFNNs are a common choice to build the dynamically updated surrogate models [37,38,39,40]. A two-phase optimization framework that fuses the exploitation capabilities of surrogates with the exploration of metaheuristics is proposed in [37]. This constitutes another trend in this field, since the combined use of different metaheuristics or surrogate types is often a strength in achieving higher accuracies [13,24,41,42,43]. An uncertainty-based criterion considering the distance and fitness value information simultaneously is proposed in [44], with two prescreening criteria to balance exploration and exploitation, and the surrogate is assisted by the PSO algorithm. A novel evolutionary strategy based on two co-evolutionary mechanisms has been proposed to improve the performance of the Jaya algorithm by replicating optimal directional guidance and historical learning [45], while in [17], the authors propose a separation of the swarms for the different optimization stages during the search space as a better alternative. Additional features like a prescreening criterion, evolution control strategies, and restart strategies are also a trend in the field, as illustrated by the authors in [46].
Considering all of these contributions and development paths, this work proposes a new surrogate-assisted algorithm, designated as SASMA, that explores the balanced Slime Mould Algorithm’s exploration and exploitation capabilities and is applied for the first time in an SAEA variant, coupled with novel thinking behind the database management strategy based on a dual-based update criterion and the surrogate building stages, i.e., the way in which the swarm positions are stored (added and removed) from the global database and how they are subsequently selected to assemble the training data for building the surrogate at each iteration.
The remainder of this paper is organized as follows: Section 2 introduces key aspects of the surrogate model working principles, particularly when coupled with evolutionary algorithms, introducing also pivotal aspects behind the radial basis function neural network model as the (commonly) chosen surrogate model. Section 3 describes the main mechanisms and presents the mathematical formulation behind the chosen metaheuristic, the Slime Mould Algorithm. Section 4 is where the proposed methodology is shown in great detail, together with the subjacent reasoning behind each of the steps. Section 5 introduces both case studies followed by a formal analysis of the different error/test results, where high-dimensional expensive functions and a classical constrained optimization problem are used to evaluate the performance of SASMA against well-known metaheuristics and state-of-the-art SAEAs. Finally, Section 6 outlines the major inferences of the presented work and provides potential research directions.

2. Surrogate Models

Function approximation is an important task in engineering problems. Following the Weierstrass approximation theorem, one can state that for a continuous function f x over a closed domain interval a , b , f x can be approximated by a polynomial φ x of a sufficiently large degree n, such that f x φ n ε , where ε > 0 is an acceptable threshold. With this theoretical background, approximation or surrogate models f ^ x are built upon previously evaluated/known training individuals using the real (objective) function f x , and depending on the quality and representativeness of the training dataset, as well as the chosen method’s ability to properly map the corresponding training space, different magnitudes of the surrogate error Δ f = f ^ x f x will emerge. As seen in the Introduction, these models have attracted researchers from a wide range of fields due to their ability to substitute expensive simulations/complex (real) models [47].
In terms of the surrogate optimization, meaning the use of the approximated model to find good estimates for candidate solutions, there are two main approaches [48]. The first is the offline (direct) approach, where the surrogate f ^ is initially constructed by using a sufficiently large set of well-distributed (expensive) individuals (quite a few pairs of candidate solutions and their corresponding real fitness). For the remainder of the optimization search, the model is kept unchanged (static), which means the optimization is made offline (without further information on the real function/simulation). In contrast, in an online (dynamic) approach, a coarse surrogate model is built initially, i.e., with a relatively small amount of training data, and then, as the iterations progress, the model is augmented with additional (expensive) training samples. This online training procedure is generically illustrated in Figure 1, constituting a type of infill strategy, where the primary objective is to balance between the exploration of unknown search space regions and the exploitation of the present promising regions.
With both approaches to surrogate optimization, it is important to have pairs of training data that are well distributed across the entire search space in order to navigate between the different local optima fairly rapidly, i.e., without the need for extensive exploration, and being less prone to being trapped in one of these optima [13]. This smoothness feature in the approximation curve, which “blinds” the search to the existence of several local optima in comparison with the real objective function, is shown in Figure 2, and it illustrates the “blessing of uncertainty” principle. On the flip side, the “curse of uncertainty” translates the downside due to a bad approximation, where f ^ leads to a wrong (nonexistent) minimum, as can be seen by the cross-sign point highlighted in the same figure.

Radial Basis Function Neural Network

When dealing with higher-dimensional approximation problems, a common trend in SAEAs is to use RBFNNs to build the surrogate model—a supervised machine learning architecture—rather than relying on more conventional ANNs or polynomial-based techniques. As the name suggests, these universal approximators take advantage of the well-documented ability of radial basis functions (RBFs), φ , to adequately interpolate a scattered set of points [49]. This being said, finding an optimal value of the shape parameter in radial basis function interpolation is not a trivial task by any means, as shown by [50].
The fully connected feedforward architecture of RBFNNs, depicted in Figure 3, typically consists of an input layer, a hidden layer in which every neuron implements a radial basis function, and a traditional linear output layer [51]. The single hidden layer is therefore the central component of these networks, containing multiple hidden neurons—commonly referred to as “RBF units” or “radbas”—that implement RBFs upon their activation. By definition, an RBF is any real-valued function ϕ ( · ) that depends solely on the distance of certain points from fixed center coordinates [52], where each φ : R n R attains its maximum activation when the input coincides with the neuron’s center. This contrasts with the activation of a conventional ANN hidden neuron, which is computed from a weighted sum of the input followed by a nonlinear activation function. Together, this distance-based activation mechanism and the characterization of each hidden neuron by its center and width [53] are what distinguish RBFNNs from standard feedforward ANNs [13].
In this regard, the Gaussian RBF is the classic choice for the activation of hidden-layer neurons, also known as the Gaussian kernel, which corresponds to the exponential part of the normal distribution’s probability density function; its canonical form for a given input vector x, Equation (1), is given below:
φ j ( x ) = exp d j ( x ) 2 σ j 2 = exp d j ( x ) 2 2 σ j 2
where φ j ( x ) : R n ( 0 , 1 ] is the j th hidden neuron response; d j ( x ) = x u j denotes the Euclidean distance between the input and the center (or prototype) u j of the Gaussian RBF, which plays a role analogous to the mean of a Gaussian distribution; and σ j > 0 is the spread (width) of the Gaussian, a role analogous to the standard deviation in the Gaussian distribution, determining how quickly the function decays away from its center.
In practical RBFNN implementations, the parameters that define the connection from the input layer to the hidden layer correspond to these two intrinsic quantities of each RBF unit [53]. When expressed in the general ANN notation—where neuron parameters are represented as weights and biases—each radbas unit, from a total of P in the hidden layer, is equipped with trainable center coordinates directly encoded as an input-weight vector u j = [ ω 1 j , , ω n j ] . The spread, on the other hand, is not encoded through a direct substitution; instead, it is conveniently represented by a trainable bias parameter b j that captures the inverse width of the basis function, i.e., b j = 1 2 σ j . Under this representation, the activation of the j th radbas neuron is given by the following [54]:
φ j ( x ) = exp b j 2 i = 1 n ( x i ω i j ) 2
Consequently, as we saw in Equation (1), for a Gaussian RBF, φ j ( x ) = 1 if and only if x i = ω i j for all i. Meanwhile, the weights connecting the hidden layer to the output layer w = [ w 1 , , w P ] form a conventional single-node shifted linear combination, producing the network’s scalar output y ( x ) = j = 1 P w j φ j ( x ) + b  [55].
Nevertheless, due to the shape–parameter sensitivity inherent to Gaussian kernels on the one hand, and the polyharmonic smoothness of cubic splines that enables robust interpolation in complex multimodal, higher-dimensional landscapes on the other, most SAEA implementations—including the one adopted in this work—employ an RBF with a cubic kernel [56,57], i.e., φ j ( x i ) = d j ( x i ) 3 .
For completeness, we note that the surrogate training procedure is used to approximate the input data by a least-squares RBF approximant [58] and follows the standard RBFNN fitting workflow: all input vectors are first normalized to [ 0 , 1 ] per dimension to improve numerical conditioning; the hidden-layer centers are taken directly from the selected training positions; and the output weights are obtained through a least-squares solution of the linear system, defined by the RBF activations. In the publicly available RBFNN construction routine (‘rbfcreate’) commonly used in the SAEA literature, the so-called ‘RBFConstant’ parameter, later codified in Section 4 as σ , does not represent a Gaussian width; instead, it acts as a numerical regularization constant in the generalized cubic kernel ϕ ( r ) = ( r 2 + c 2 ) 3 / 2 , which improves stability when the interpolation matrix becomes ill-conditioned. Likewise, the ‘RBFSmooth’ parameter corresponds to the (linear) regularization term added to the least-squares system, functioning as the approximation error goal ε RBF and providing an implicit stopping criterion that prevents overfitting when the training set is sparse.

3. Slime Mould Algorithm

The Slime Mould Algorithm (SMA) is a fairly recent swarm intelligence metaheuristic proposed in [59]. It mimics the morphological and foraging behaviors of an acellular slime mould protist organism, Physarum polycephalum, that inhabits humid and cold places. Particularly of interest is its food-seeking mechanism, where the cytoplasmic flux of slime mould surrounds and digests food by searching with a front end that resembles a fan shape, interconnected trough a venous network. Then, through a sequence of negative and positive feedback, depending on the odor, the slime mould finds its way to the target food by adjusting its propagation wave to alter the cytoplasmic flow in its veins, as is illustrated in Figure 4.
In terms of mathematical modeling, the SMA replicates the two main stages of this foraging behavior, namely food approach and food wrapping. In accordance, the algorithm begins i t e r = 0 by randomly initializing the population, X i t e r = X ( 0 ) = l b + rand u b l b = X n P o p × n D i m s , in a search space defined by the respective problem’s lower and upper bounds at each dimension, l b and u b , with a total of n P o p (different) search agents (population size) spread all across the problem’s dimensions, n D i m s . To improve clarity, we explicitly highlight that this initialization corresponds to a uniform sampling of the entire search space, ensuring that the initial slime mould positions are well distributed before the algorithm begins its exploration–exploitation dynamics.
The slime mould position update is then modeled by the expressions in Equation (3), where X i , d i t e r + 1 indicates the updated position of the ith search agent (swarm individual) in a given dimension d; l b d and u b d represent the lower and upper bound of the search space in a given dimension d 1 , n D i m s ; r 1 , r 2 and rand denote uniformly distributed random values in the interval of 0 , 1 ; z is a sensitive small threshold parameter that ensures diversification in the population by assigning a newly generated random value to the aforementioned updated position; X b e s t , d i t e r utilizes the index b e s t , which stands for the best swarm individual (the slime mould position with the highest odor concentration) encountered so far, and the index d, which refers to a specific dimension of this individual; and D F is its corresponding (swarm’s best) fitness value. This separation between the best-known position and the remaining agents is essential, as it defines the attraction mechanism that drives the population towards promising regions of the search space.
v b is a randomly uniformly distributed vector that is generated for each ith search agent for all n D i m s dimensions (implying a different v b when updating different slime mould positions), oscillating within the range of a , a , such as v b = a + 2 a · rand , where a = tanh 1 i t e r i t e r max + 1 , which means it gradually decreases towards zero, lim a i t e r i t e r i t e r max = 0 ; v c is an analogous vector to v b since it also converges to zero and is generated for each swarm individual, but instead oscillates in the range of 1 , 1 , i.e., v c = b + 2 b · rand , where b = i t e r max i t e r i t e r max . Both v b and v c control the oscillatory movement characteristic of slime mould behavior, gradually reducing their amplitude as the algorithm converges, which naturally shifts the search from exploration to exploitation.
W i is a weight vector assigned to each slime mould position and is computed using Equation (4). Together with v b and v c , it is one of the parameters that drives the swarm individual’s search; X A i t e r and X B i t e r represent the notation used to denote two randomly selected (current) swarm individuals, i.e., with indexes A B 1 , n P o p , which are selected to update each current position when r 2 < p i . Lastly, p i is an individual threshold value that determines which of the two main position update equations is selected to perform the position update, given by p i = tanh S i D F , where S i denotes the fitness value of the ith agent. This probability term p i is central to SMA: it adaptively controls whether an agent is attracted toward the best solution or contracts around its current position, effectively balancing global and local search. An overview of the SMA space search mechanisms and the balance between exploration and exploitation is given in Figure 5.
X i , d i t e r + 1 = l b d + rand · u b d l b d , r 1 < z . X b e s t , d i t e r + v b , d · W i · X A , d i t e r X B , d i t e r , r 2 < p i r 1 z . v c , d · X i , d i t e r , r 2 p i r 1 z .
W i = W S m e l l I n d e x i = 1 + r 3 log 10 b F S m e l l I n d e x i b F w F , if i p o p s i z e p o p s i z e 2 2 . 1 r 3 log 10 b F S m e l l I n d e x i b F w F , otherwise .
In turn, to calculate the weight vector for each search agent, the variable S m e l l I n d e x i , which is nothing more than the sorted fitness values of the entire population, namely S m e l l I n d e x = sort S , is introduced, and then the log of the quotient between its difference to the current worst fitness w F and the difference between the current best, b F , and the same w F is obtained. The random value r 3 is responsible for the uncertainty in the venous contraction phase and is multiplied by the log of the quotient. For the swarm individuals ranked in the best/higher half of the population, this term is added by 1 for a larger weight and subtracted from 1 for the lower half of the population. This weighting mechanism reinforces the influence of better-performing agents while still allowing weaker agents to contribute stochastic perturbations, which is a key biological analogy of slime mould pulsation behavior.

Bound Checking

An important implementation detail concerns the bound checking after each position update (Equation (3)), given the possibility of agents traveling beyond the established (test problem) bounds l b , u b , i.e., the defined search space. As such, in this work, the commonly used deterministic back confinement [60], given by Equation (5), is considered not only for the SASMA but also for the benchmark algorithms.
X i , d i t e r + 1 = min X i , d i t e r + 1 , u b d , i = 1 , n P o p d = 1 , n D i m s X i , d i t e r + 1 = max X i , d i t e r + 1 , l b d , i = 1 , n P o p d = 1 , n D i m s

4. Proposed Methodology

In terms of methodology, an overall view of the SASMA main mechanisms and the flow of information between them is provided in Figure 6. This starts with the familiar Latin hypercube sampling (LHS) initialization of the search agents’ positions (shared with the benchmark algorithms), followed by the computation of their respective real objective value, which is then assigned to the global database, forming the initial set of training data for the first surrogate build. Thus, emphasizing this throughout the iterations, only positions deemed as “promising” are assigned their real objective value, and from these, only a handful will enter the DB, grounded on a dual-based criterion explained in Section 4.1, a key feature to ensure a more accurate function approximation.
The subsequent stage involves the chosen metaheuristic model, SMA, which is used in an adapted manner to find the optimal solution of each test problem, while the surrogate model is tasked with replacing the objective function, meaning that the swarm agents navigate through the search space based on approximate model information, f ^ x , rather than the usual expensive real test problem. Section 4.2 dives into the specifics of the surrogate’s construction and how the training (positions) and target (real objective values) are selected. These stages will run in a cyclic way until the chosen stop criterion is met, which, as is common in the SAEA literature, is a maximum number of function evaluations m a x F E s . Additionally, as we saw in Section 2, the added uncertainty of using an estimate of the real value can also work in our benefit by “flattening” some of the real model’s “valleys” during the search space process, i.e., minimizing the local minima phenomena.

4.1. Global Database Management Strategy

To build a surrogate model that evaluates the different potential solutions (slime mould positions), it is necessary to have a dynamic database, working as a stack of (available) positions at each iteration. As illustrated in Figure 1, these stored (search agents) positions, as well as their respective fitness, constitute the input data and are the target data. In this work, a fixed length, l e n g t h D B , is considered for the DB, independent of the test problem dimension (after performing several test runs). As we saw earlier, the database is initialized with the initial positions and their respective real fitness, as well as the stored position age, i.e., the iteration number in which this given position entered the database. This control variable is used to evaluate the possible stagnation of the stored swarm agents, and thus, it can also work as a secondary update criterion if needed.
Then, with these iterations, new positions are added/replaced in the DB, which means the accuracy of the surrogate model will improve over time; otherwise, it would require a very large population size. Unlike more traditional surrogate-assisted approaches, the criteria used to add/replace positions in the DB are based on a best weighted metric [61] that considers two different factors rather than only one. The first criterion takes into consideration the surrogate value of the recently updated search agents (via SMA), f ^ X i i t e r + 1 or simply f ^ X i , whose objective is to minimize, and it scales this value based on the current best, f ^ min = min f ^ X 1 , f ^ X 2 , , f ^ X n P o p , and worst, f ^ max = max f ^ X 1 , f ^ X 2 , , f ^ X n P o p , surrogate values, i.e., the current positions that achieved the minimum and maximum fitness approximation values. So, for each swarm agent, i, its scaled surrogate value, S X i i t e r + 1 , is given by Equation (6).
S X i = f ^ X i f ^ min f ^ max f ^ min , S 0 , 1 i 1 , n P o p
The second criterion is centered around the distance to the existing DB positions, since when evaluating the potential of an updated search agent to enter the global DB, it is not only important to have a low (estimated) fitness value but also to be as far away as possible from the previously evaluated positions, consequently promoting a more global search space, which in turn helps avoid local minima problems. Thence, we start off by obtaining the Euclidian distance of every position of the current swarm with respect to all DB stored positions, i.e., for each swarm position, X i i t e r + 1 , compute its distance to every j position in the DB, X j D B , such that d i j = X i i t e r + 1 X j DB . Then, from this matrix of Euclidean distances, we not only find the overall minimum and maximum distances, d min = min d 11 , d 21 , , d n P o p , 1 , d 1 , 2 , , d n P o p , l e n g t h D B and d max = max d 11 , d 21 , , d n P o p , 1 , d 1 , 2 , , d n P o p , l e n g t h D B , respectively, but we also obtain a vector with the minimum distances of every ith position of the current swarm to any given stored position in the DB, i.e., d s i n g l e X i = min d i , 1 , d i , 2 , , d i , l e n g t h D B . With this, a scaled distance metric, D X i , can be obtained using Equation (7).
D X i = d max d s i n g l e X i d max d min , D 0 , 1
With both metrics, a value closer to zero indicates a better candidate to enter the DB; however, they can be somewhat conflicting, i.e., a certain position can have a small (good) scaled surrogate value but at the same time have a large scaled distance value (be very close to the existing DB positions), or vice-versa. For this reason, the merit function, f merit , is a combination of the two metrics, as seen in Equation (8), where through a weight, 0 < ϕ < 1 , it is possible to modulate which metric should receive higher priority.
f merit X i = ϕ S X i + 1 ϕ D X i
A small value of ϕ denotes a greater importance to the more distant positions, which is key to the exploration stage, while a larger value of ϕ implies a prioritization of the smaller scaled surrogate values, which in turn favors the later exploitation phase. To achieve this outcome, a linear increasing function, ϕ = ϕ min + ϕ max ϕ min i t e r max i t e r , is considered in this work, rather than the often-suggested four cycles of constant ϕ values.
Calculating the f merit is just the first step. The entire DB management, as a key feature of the proposed SAEA, comprises a series of steps that must be carried out at each iteration in order to initiate and dynamically update the DB, i.e., the variables X D B , f i t n e s s D B and a g e D B . These recurring steps are fully described in a flowchart form in Figure 7, and they involve the following: sorting the current SASMA swarm based on the respective f merit X i , which leads to a sorted set of positions, X s o r t ; a verification to see if the position is already stored in the DB; and two “semi-elitist” conditions that verify if the potential DB entries are “worthy”. So, the updated search agents ranked in the first 15% in terms of f merit , and 25% of the time, comprising the top quarter, the ranked positions were deemed as valid thresholds, which strengthens the diversity in the DB management strategy.
Finally, even if it passes these conditions, the DB is checked to see if it is already full (having reached its l e n g t h D B limit), and, if so, we need to decide which position to remove, or conversely, maintain, in the DB. Accordingly, the considered criterion is to use the real fitness value as an indicator of the position potential to the subsequent surrogate building stage, so the worst-ranked stored position, max f i t n e s s D B , arbitrarily identified by an index k, X k D B , is compared with the real fitness value of the potential DB entry, f X i s o r t , and in case it is higher, the position is replaced; otherwise, X i s o r t is disregarded.

4.2. Surrogate Building

A crucial factor when assembling the surrogate model is the chosen training data, meaning what positions are handpicked to fit the RBFNNs. In other words, as we saw previously, it is important to have a dispersed set of positions to “accurately map” the search space regions, particularly in the exploration phase, which means that the training data length also matters. But also, the “worthiness” of those positions also matters to properly narrow in on the optimum during the exploitation phase, i.e., in regions with bigger slopes in terms of fitness accuracy (smaller variations in the positioning leading to bigger fitness variations), it is more important to have more scattered positions around these, thus ensuring a better fitted surrogate model, which in turn can mitigate the “curse of uncertainty”.
Understanding that the global DB tends (through the course of the iterations) to have a much wider distribution of positions, given its “global nature”, it is important to narrow down the search space defined to effectively fit the surrogate. This implies that after the first iteration, having concluded the first position update, probably only a portion of the global DB data is used to train the surrogate model, which also saves some computational time. This is an adequate approach to find the balance between the current swarm positioning and the overall upper and lower limits of the minimization problem, therefore allowing the surrogate model to shift its global nature to a more local nature in the closing stages of the search, i.e., with a concentration of swarm agents around the most promising subspace regions. This methodology also includes a safeguard mechanism: when no surrogate-suggested improvement is confirmed by real FEs, all agents are evaluated with the real objective function to prevent misleading surrogate drift during early iterations, ensuring robustness without significantly compromising the computational budget.
To accomplish this outcome, and inspired by the fitness approximation strategy proposed in [13], we start by finding the upper and lower limits of the current swarm agents, X i t e r + 1 , i.e., the subspace defined by the current maximum and minimum values of all the n P o p agents at each dimension d, denoted as X c u r r u b and X c u r r l b , and computed as in Equations (9) and (10).
X c u r r u b i t e r + 1 = max X 1 , d i t e r + 1 , X 2 , d i t e r + 1 , , X n P o p , d i t e r + 1 , d = 1 , 2 , , n D i m s
X c u r r l b i t e r + 1 = min X 1 , d i t e r + 1 , X 2 , d i t e r + 1 , , X n P o p , d i t e r + 1 , d = 1 , 2 , , n D i m s
Then, we can compute the subspace, X s s p a c e , used to select the training data from the global DB, i.e., the positions of the DB that are within certain upper, X s s p a c e u b , and lower, X s s p a c e l b , bounds, as computed in Equations (11) and (12).
X s s p a c e u b i t e r + 1 = min X c u r r u b i t e r + 1 + α X c u r r u b i t e r + 1 X c u r r l b i t e r + 1 , u b
X s s p a c e l b i t e r + 1 = max X c u r r l b i t e r + 1 α X c u r r u b i t e r + 1 X c u r r l b i t e r + 1 , l b
where α is a spread coefficient (and should not be confused with the RBFNN spread parameter mentioned in Section 2), between 0 and 1, that determines how farmbeyond the current positioning the selected subspace goes. A value of 1 implies that the entire DB (all positions in the search space) is considered for building the surrogate, while a value of 0 implies that only the current swarm positioning determines the subspace bounds.
Modifying the original approach to this spread coefficient, we opted for an exponentially decreasing coefficient, i.e., α = α max e γ i t e r + 1 , hence, fulfilling the objective of a wider subspace in the exploration phase versus a reduced promising region in the later exploitation stages. In addition, the minimum and maximum functions in Equations (9) and (10) tend to usually return the first term (expression), constituting the bound values, and this explains why we said earlier that the surrogate model neglects the less “interesting” positions of the DB and is driven by the current swarm evolution.
An illustration of how Equations (11) and (12), with their varying spread coefficient, determine the subspace used to build the surrogate is provided in Figure 8. Importantly, this figure showcases the difference between the resultant subspace domain bounds, X s s p a c e , and the effective training domain, X t r a i n s p a c e , which is defined by the range between the upper and lower bounds per dimension of the n selected DB positions inside the computed subspace, i.e., max X 1 , d t r a i n , X 2 , d t r a i n , , X n , d t r a i n min X 1 , d t r a i n , X 2 , d t r a i n , , X n , d t r a i n , where d = 1 , 2 , , n D i m s . To ensure that the resultant training subspace leads us to a balanced training domain for the surrogate model, i.e., there are sufficient data points without a very expensive computation cost in the RBF model fitting, a verification of the X t r a i n length, given by the index n, is made (as stated in Algorithm 1). A failure to comply with the established (balanced) length limits leads to either a truncation or an extension of the  X t r a i n s p a c e .
When employing the RBFNN as the surrogate model, an important control parameter concerns its spread, σ j , given in Equation (1). A larger spread will require many neurons to find a rough fitness function approximation, while, for a smaller spread, the large number of neurons is the prerequisite to get a smooth function approximation, but it risks overfitting. To achieve this fine balance, an adaptative s p r e a d , as shown in Equation (13), based on the current training set X t r a i n subspace (bounds) is used [13].
σ = min max X 1 , d t r a i n , X 2 , d t r a i n , , X n , d t r a i n min X 1 , d t r a i n , X 2 , d t r a i n , , X n , d t r a i n , d = 1 , 2 , , n D i m s
Moreover, as hinted in Section 3, an error goal is defined as a proxy of the approximation accuracy of the surrogate model. In this regard, a typical approach is to use two surrogate levels [13,42], i.e., a more global surrogate model where a greater error is allowed, and a more refined model for each or individual parts of the swarm agents, often defined as the local surrogate model, where accuracy is targeted, to find a nearby solution. Acknowledging the benefits of this approach, in this work, we opted for a single global surrogate that transmutes to a more local type of surrogate model by defining a linearly decreasing error goal, as shown in Equation (14). Hence, our model constitutes a slightly different approach that shares the same principle. In other words, the model targets a lower accuracy when the DB is filled with a more dispersed set of positions (exploration phase), while favoring more accuracy in its later stages, when the DB is composed of neighboring positions, given the convergence properties of the SMA.
ε R B F = ε max ε max ε min i t e r max i t e r
Algorithm 1 Pseudocode of the Surrogate-Assisted Slime Mould Algorithm (SASMA) and the employed parameters.
Input: 
n P o p = 30 ; n D i m s = 30 or n D i m s = 100 ; m a x F E s , f, l b , u b are test function/problem dependent; i t e r = F E s = 0 ; z = 0.03 (SMA parameter); l e n g t h D B = 1000 ; ϕ min = 0.35 and ϕ max = 0.95 ; α max = 0.305 and γ = 1.5 × 10 3 ; l e n g h t X t r a i n l b = min length X t r a i n s p a c e = n P o p and l e n g h t X t r a i n u b = max length X t r a i n s p a c e = 5 × n P o p ; ε min = 0.01 and ε max = 0.1
1:
Initialize (using LHS) the population, X, where the search agents or positions are denoted as X i , i 1 , n P o p ; X t e m p = X
2:
Evaluate the real fitness of the initialized agents S i = f X i ; X b e s t | b e s t = arg min i 1 , n P o p S i ; S m e l l I n d e x = Sort S | S m e l l I n d e x i S m e l l I n d e x i + 1 , i 1 , n P o p ; D F = b F = min i 1 , n P o p S i ; w F = max i 1 , n P o p S i
3:
Store this information into the DB X D B , f i t n e s s D B , a g e D B , according to the flowchart in Figure 7
4:
Assemble the training set, X t r a i n using Equations (9)–(12); Compute the spread, σ , using Equation (13); Compute the error goal, ε R B F , using Equation (14); Build the cubic RBF surrogate (approximation) model, f ^
5:
while  F E s < m a x F E s  do
6:
    if  i t e r > 0  then
7:
         S m e l l I n d e x = Sort S | S m e l l I n d e x i S m e l l I n d e x i + 1 , i 1 , n P o p
8:
         b F = min i 1 , n P o p S i = S m e l l I n d e x 1 ; w F = max i 1 , n P o p S i = S m e l l I n d e x n P o p
9:
    end if
10:
    if  S m e l l I n d e x 1 < D F  then
11:
         D F = S m e l l I n d e x 1
12:
    end if
13:
    for  i = 1 : n P o p  do
14:
        Compute the weight of each individual slime mould, W i , using Equation (4)
15:
    end for
16:
    Compute the auxiliary SMA variables: a , b
17:
    for  i = 1 : n P o p  do
18:
        Calculate the variables p , v b , v c
19:
        Update each swarm agent (position) and store it in a temporary variable, X i t e m p , according to Equation (3)
20:
        Bound checking of the updated position X i t e m p , using Equation (5)
21:
    end for
22:
     r e a l F E c o u n t e r = 0
23:
    for  i = 1 : n P o p  do
24:
        Compute the surrogate value for each updated position, f ^ X i t e m p
25:
        if  f ^ X i t e m p < S i  then
26:
           The updated position is deemed as “promising”, so it is accepted, X i = X i t e m p
27:
           The real fitness value is therefore computed, S i = f X i
28:
            r e a l F E c o u n t e r = r e a l F E c o u n t e r + 1
29:
        end if
30:
    end for
31:
    if  r e a l F E c o u n t e r = = 0  then   ▹ none of the updated positions were deemed as “promising”
32:
        for  i = 1 : n P o p  do
33:
           Accept the updated position regardless, X i = X i t e m p ▹ as in the original SMA code
34:
           Compute its real fitness and store the value, S i = f X i
35:
            r e a l F E c o u n t e r = r e a l F E c o u n t e r + 1
36:
        end for
37:
    end if
38:
    for  i = 1 : n P o p  do
39:
        Compute the variables S X i , D X i , ϕ , according to Equations (6) and (7) and the inline ϕ equation in Section 4.1
40:
        Compute the dual based merit metric, f merit X i , Equation (8)
41:
    end for
42:
    Update the DB, i.e., X D B , f i t n e s s D B , a g e D B , based on the f merit and S, according to several stages depicted in the flowchart in Figure 7
43:
    Assemble the new training set, X t r a i n using Equations (9)–(12); Update the spread, σ , and the error goal, ε R B F , using Equation (13) and Equation (14), respectively; Update the cubic RBF surrogate (approximation) model, f ^
44:
     F E s = F E s + r e a l F E c o u n t e r
45:
     i t e r = i t e r + 1
46:
end while
Output: 
X b e s t ; f X b e s t
Overall, the careful fine-tuning of the control parameter ϕ regarding the database management strategy, as well as the surrogate-building-related parameters γ , α , σ , and ε R B F , which, with the exception of the first, all vary along the course of the iterations, allows us to gradually shift from a more global to a more localized surrogate model. This shift is crucial to capture the error differences, with increasingly smaller position updates during the exploitation phase.

4.3. A Novel Surrogate-Assisted Metaheuristic: SASMA

Having described the holistic approach and key aspects behind the proposed SASMA, with a focus on the novelties regarding the DB management strategy and the surrogate-building approach, it is now useful to introduce a more in-depth account of the flow between the main mechanisms illustrated in Figure 6, namely by comprising all the involved variables and the individual stages needed to adapt the original SMA to its SAEA form in pseudocode (as shown in Algorithm 1).

5. Case Studies and Benchmarking

To conduct a fair evaluation of the proposed SASMA, well-known SAEAs with publicly available code—namely TLSAPSO [13], from which many of the SASMA features were inspired, as mentioned in Section 4, SAEA-RFS [62], SHPSO [40], TL-SSLPSO [63], CALSAPSO [64], and GORS-SSLPSO [39]—were selected, and their code was modified accordingly. For verification and fair replication purposes, the same initial positioning is seen by each algorithm for every test run, the main control parameters (for the benchmark algorithms) are shown in Table 1, and the full test code can be accessed in [65].
With regards to the optimization, a total of 35 runs are performed, which means that the same 35 different initial positions are used to benchmark all the algorithms, thus dismissing initial positioning bias. Moreover, this allows us to assess not only the methodology’s accurateness but also its precision (through the standard deviation), i.e., to judge the ability of each method to consistently find the best solution. The swarm size, n p o p , is set to 30, and these individual agents will survey the problem’s search space up until a total of 330 or 1000 (real) FEs for 30D and 100D, respectively, are reached (following the literature standard). Unlike in traditional metaheuristic algorithms, where there is an equivalence between the number of iterations and the FEs as a stop criterion, this is no longer the case when using surrogate-assisted algorithms, since most of the time, on a given iteration, we end up evaluating only a limited amount of the updated positions, and not the entire population.
The difference between the best (real) fitness value achieved in each run by each algorithm and the problem/test function global minimum, represented as f min in Table 2, is the chosen error metric. To analyze the error performance of the proposed SASMA versus the chosen benchmark algorithms, beyond the common descriptive statistics, i.e., the mean, the standard deviation, and the minimum and maximum error values, we followed the common approach of using the non-parametric Wilcoxon signed rank and the Friedman statistical tests with 5% significance [66].
The first test gauges if there are substantial differences in the central tendency of two competing data series (SASMA vs. all the benchmark algorithms). A one-on-one comparison of the signed errors in the 35 runs for all the problems/test functions allows us to say if the algorithm’s error accuracy differs substantially (null hypothesis), and if so, attributing the symbol “=”, or on the contrary, if there is a definite difference in terms of central tendency (alternative hypothesis). A smaller mean error value for the SASMA versus the benchmarked algorithms means a better performance of the proposed SASMA, and thus, the symbol “+”, whereas the opposite leads us to the symbol “−” in the result analysis (tables). Meanwhile, the second test allows us to perform a multiple-algorithm comparison for the non-normally distributed algorithms’ signed error for each of the 35 runs, revealing if there are any significant differences in the obtained problem/test function errors (alternative hypothesis). The same is conducted by examining the (Friedman) mean rank, rather than the true value. It is expected that for similar distributions, the mean ranks will be approximately identical (null hypothesis).

5.1. Case Study I: Mathematical Test Functions and Optimization Results

Seventeen classical benchmark test functions, shown in Table 2 alongside their global minimum f min and properties, are used to evaluate the proposed SASMA, including its scalability, by considering two different numbers of high dimensions, n D i m s = 30 , 100 . Among these test functions, F1–F4 and F14 are continuous unimodal functions; F5 is the Rosenbrock function, which can be considered multimodal when the problem dimension is greater than 3 [68]; F6 is a discontinuous step unimodal function; F7 is a noisy unimodal function; and F8–F13 and F15–F17 are multimodal functions, with many local minima, thus rendering them more difficult to optimize [69]. A detailed account of the optimization bounds and function mathematical expressions is given in Table A1.
As mentioned before, key descriptive statistics and two non-parametric tests are used to assess the algorithm’s performance. As such, for all the testing functions F1–F17, the mean, standard deviation (STD), and minimum error values of all the benchmark SAEAs and the proposed SASMA are presented in Table 3 and Table A2, respectively, for 30D and 100D. CALSAPSO’s very slow convergence for 100D made the simulation cost very heavy, which is why it was not considered in the test case. The best results per metric are shaded in grey, clearly indicating the superior performance of SASMA, which only failed to achieve the best mean error value (lowest value) on four functions for 30D, and on three occasions for 100D, a scenario that is pretty similar in terms of the minimum error value, where we assess the ability of SASMA to be closest to the global function minimum in at least one of the runs. Despite being a harder problem, SASMA’s performance for 100D is even better than for 30D, which is explained by the larger number of function evaluations, and this is why for nine out of the seventeen functions, we see that SASMA is able to consistently find the global minimum (an STD of zero), while also being very close to this target value for F12.
The same can also be confirmed by the results of the Wilcoxon signed ranked test at a 5% significance level, where SASMA is compared against every other algorithm. For both 30D and 100D, SASMA records significantly better error accuracies in most of the seventeen test functions (overwhelmingly scoring wins) against the benchmarked algorithms. The few exceptions mostly occur in the last three test functions.
For 30D, GORS-SSLPSO was the only benchmark algorithm to show a statistically better error performance in four functions, particularly in F1 and F6, with values in the order of 10 5 , while the remainder were only able to surpass or tie with SASMA in F15–F17, which means that for at least fourteen out of the seventeen functions, SASMA was better. SHPSO and TL-SSLPSO showed consistent standard deviations, and this last SAEA achieved the best performance in the Shifted Rotated Rastrigin function. Meanwhile, for 100D, the advantage of SASMA is even more clear, overall recording more wins, one less loss, and only one tie. SHPSO is its closest competitor, showing an outstanding performance for F6, F15, and F16, while for the remainder, it is clearly outmatched by SASMA, partially confirming the trend already seen for 30D.
This improved (lower) error accuracy is also attested by the Friedman test, which was performed for each function, taking into consideration each test run error. By analyzing the computed mean rank, we can then individually order each algorithm, where the lowest mean rank score indicates a better performance. The results for 30D are shown in Table 4, where we can see that SASMA is consistently the best ranked algorithm, with the exception of four functions (F1, F6, F15, and F16). Nevertheless, for the first of the two, GORS-SSLPSO closely follows as the second-best SAEA in terms of mean rank. For most of the test functions, the top three ranked algorithms are SASMA, GORS-SSLPSO, and TL-SSLPSO. Meanwhile, on the opposite end, TLSAPSO and CALSAPSO are the worst ranked SAEAs.
As for the 100D results, shown in Table A3, the accuracy of SASMA is even better, achieving the best mean rank in fourteen of the test functions, and as was the case for 30D, it is GORS-SSLPSO that achieved the best mean rank for F6, while for F15 and F16, it is now SHPSO that is ranked the best in detriment of TL-SSLPSO. Completing the top three in this test metric are again SASMA and GORS-SSLPSO, which are now joined by SHPSO. Contrariwise, TLSAPSO and SAE-ARFS occupy the lower (ranked) end of the benchmark algorithms. Notwithstanding all the comparisons, the comprehensive error analysis not only confirms SASMA’s ability to mitigate the initial positioning bias but also its ability to substantially outperform TLSAPSO. This is a very significant result given that many of SASMA’s features are inspired by it.
To visually illustrate the aforementioned superiority of SASMA in comparison with the benchmarked SAEAs, the convergence curve of the best SASMA run, i.e., the run where the proposed methodology achieved the minimum error value, is shown together with the convergence curves for all the selected algorithms, thus ensuring that all start from the same initial positioning. Figure 9 depicts the results for 30D and the unimodal test functions, and it is no surprise to see that SASMA, apart from F6 where GORS-SSLPSO excelled, achieves the minimum fitness value (log), and most of the time in far fewer FEs.
For the multimodal test functions, shown in Figure 10, still considering 30D, the picture is very similar, with SASMA consistently reaching the lowest fitness value with a faster convergence (sharp drop) in test functions F8, F12, and F13. It is again followed by GORS-SSLPSO and SAHO. The very complicated multimodal test functions shown in Figure 11 reveal a more challenging scenario where SASMA greatly outperforms all the SAEAs in F9 and, despite slower convergence, achieves the same in F17. Meanwhile, for F15 and F16, it is TL-SSLPSO that shows a greater performance, confirming the Wilcoxon and Friedman test results.
The very complicated multimodal test functions are shown in Figure 11.
For 100D, the convergence curves for the unimodal test functions are shown in Figure A1, the multimodal test functions in Figure A2, and the very complicated multimodal test functions in Figure A3. Likewise, we observe that the superiority of SASMA is even more evident, as the convergence curves for F3, F4, F10, F11, and F9 reveal. An exception occurs in F6 and F16, where GORS-SSLPSO and SHPSO, respectively, are able to outperform it. Noticeably, SASMA maintains its fast convergence characteristics, and it even reaches the global minimum in F11, and so, the log curve vanishes from this point onwards. In all the convergence curve figures, we can also confirm the trend revealed by the several error metrics and non-parametric tests, where GORS-SSLPSO and SHPSO are the closest competitors of SASMA, while the other SAEAs tend to display premature convergence and stagnation features.
In terms of computational cost, a relevant feature in large-scale optimization problems, the relative time taken by the benchmark SAEAs in comparison with SASMA is shown in Figure 12, and is measured as the ratio between the time taken by each of the SAEAs and the time taken by SASMA. As we can see, the x-scale is given in a logarithmic scale, given that most of the algorithms take several orders of magnitude more than SASMA, with the exception of TL-SSLPSO and SAE-ARFS, for 30D and 100D, respectively, which are slower but roughly in the same order of magnitude.
To complement the results analysis, the SASMA DB properties after each individual run for each test function are analyzed via a bivariate histogram in Figure 13. The bar plots on the left side highlight the relative frequency of the final DB length and average age of the stored positions, whereas the bar plots on the right highlight the distribution of the rules activated in order to enter the DB for all the performed runs. The distribution for 30D unveils that the most frequent class for the DB size is between 133 and 144 stored positions, while the average age of the positions is around seven to eight iterations, meaning that they are fairly new, and that with the considered small number of FEs, the DB is being filled with rapidly improving positions. This scenario increases significantly for the 100D case, where the most frequent class for the DB size is between 374 and 401 stored positions, while the average age of the positions is around 26 to 27 iterations, thus indicating a more mature level of stored positions, which translates into more constancy in the surrogate approximation.
These results are in line with the intended semi-elitist nature of the DB entry rules, designed in Algorithm 1 and shown in Figure 7. The same can be verified by the right-side bivariate distributions, where we see that independent of the number of problem dimensions, rule 1 and rule 2 are responsible for roughly between 85/89% and 15/11% of the DB position entries, respectively, for more than half of the times (most frequent classes).

5.2. Case Study II: 25 Truss Bar Design (Continuous) Problem

A truss design problem was chosen to provide an additional real engineering case study, in a field known as structural optimization, to validate the proposed SASMA. As the name suggests, it concerns the design optimization of truss bar structures with the purpose of minimizing its weight while still fulfilling displacement and stress constraints. The underlying mathematical formulation behind the constrained objective function and the penalty method is provided in Appendix B.
The 25-bar transmission tower shown in Figure 14 is one of the most broadly used truss design problems. As such, it is a perfect fit to compare the different algorithms and verify the numerous design methodologies through which each one ensures the minimum weight, while preserving its structural integrity. The elements that form the 25-bar truss are organized into eight groups, meaning eight design variables, and all the members in the same group share the same cross-sectional and material properties. The three-dimensional coordinates for the ten nodes/elements shown in Figure 14, namely unit weight, modulus of elasticity, loading conditions, and the respective member grouping, can be found in [47].
The continuous design variable variant for the 25-bar truss problem assumes that the cross-sectional areas are continuous decision variables (no need to map to a set of feasible discrete design variables) and considers multiple loading conditions, i.e., different stress constraints for each node, with minimum and maximum cross-sectional areas of 0.01 and 3.40 in2 [allowable bounds in Equation (A2)], respectively.
Therefore, this constrained continuous truss problem presents a different challenge in comparison with the first case study. That is, although the number of dimensions is smaller, there are now a set of relationships between the different design parameters that imply that even solutions (cross-sectional areas), X, inside the allowable bounds may possibly violate the other restrictions, as shown in Equation (A1). Hence, this problem tests the SASMA exploration capabilities to guide the search agents towards the global optimum based on the changes given by the penalty function, Equation (A2).

Optimization Results

So, to further justify the use of SMA as the base algorithm and to verify that the different SASMA mechanisms enhance the SMA optimization capabilities, regardless of the type of optimization problem, a secondary evaluation is made with a constrained problem with inner relations between the variables, where SASMA is put against the original SMA, as well as with common metaheuristics used to solve the truss design problem, namely, a PSO with an inertia factor and with constriction factor, WOA, GWO, and other ordinary metaheuristics like gravitational search algorithm (GSA) [70], flower pollination algorithm (FPA) [71], bat algorithm (BA) [72], and gaining-sharing knowledge-based algorithm (GSK) [73]. The respective control parameters (Table A4) are used to evaluate, under the same error accuracy metrics/tests assumptions, thus providing an additional comparison of SASMA optimization capabilities. Yet, since we are dealing with a single optimization problem, the results of the two non-parametric statistical tests, which gauge the differences in the errors of the different algorithms, are written together alongside the descriptive statistics in Table 5. The convergence curves for all the algorithms, regarding the test run where SASMA reached the minimum value, are shown in Figure 15.
The non-parametric test results reveal that GSK is the best performing algorithm, with a lower mean rank, as well as the lowest mean, standard deviation, and minimum error values of 5.4516 × 102, 1.9313 × 10−4, and 5.4516 × 102, respectively. The proposed SASMA then follows as a close second-best ranked algorithm, as proven by the mean rank, and with fairly nearby mean and minimum values of 5.4585 × 102 and 5.4523 × 102, respectively. With the exception of GSK, SASMA is able to outscore all the other algorithms as attested by the “+” signal from the Wilcoxon signed rank test in all the columns in Table 5. Completing the top three is the SMA, presenting a standard deviation with a negative exponent, which is only achieved by GSK, SASMA, SMA, and GWO (in this ranking order). Unlike in Case Study I, WOA and GSA are now among the bottom ranked algorithms (together with BA), both in the error metrics and the mean rank, highlighting perfectly the “No Free Lunch theorem”, i.e., algorithms that work well in a class of problems will fail on another class of problems.
Figure 15 highlights a similar trend to the one observed in Case Study I, namely that SMA tends to present a faster convergence in the initial phase of the search space, certainly with fewer problem dimensions, proving that the intended balance mechanisms are effectively slowing down its surrogate-assisted counterpart convergence by favoring a more global nature search in this phase, with the hopes of bearing the fruits later on. Both the PSO variants and GSK are very good in the initial phase, whereas GSK is the only one capable of outperforming SASMA and SMA in the latter stages. Unlike what was seen in Case Study I, WOA now shows a compromised stagnation almost right from the get-go, which may suggest an inadequacy of its control parameters, which once again underscores the importance of using a more (control) parameter-independent algorithm like SMA.

6. Conclusions

Solving complex real-world optimization problems with computationally efficient algorithms is a major trend of the evolutionary computation field. Traditional metaheuristics tend to rely on many FEs, which is troubling when the problem requires a hefty computational cost (model simulation). To address this challenge, surrogate models, which constitute an approximation model of the real objective function, are used in lieu of these expensive FEs, thus reducing the computational cost. With the advent of many SAEAs, the focus has been not only in the diversification of the base metaheuristic but more importantly on the mechanisms that control the flow of information between the algorithm’s swarm and surrogate building, i.e., how and which are the positions and respective fitness selected from the evolving swarm to build the approximated version of the objective function. And to this end, it is crucial that the database, as an intermediary layer, has the adequate information to ensure a general level of accurateness in the surrogate-building stage.
As such, the proposed SASMA relies on a dual-based criterion that prioritizes simultaneously the surrogate value of a given (discovered) position and its distance to the already stored positions as the chosen update mechanism of the DB. With a balance that changes with the course of the iterations from the distance to the surrogate value (minimum), SASMA ensures that the priority is to have a more spread set of stored positions in the exploration phase, i.e., mapping the search space, while it targets more accuracy in the latter phase, when the SAEA is already converging (exploitation). By using this approach together with SMA, which does not need to use local information of the individual swarm agents, we can avoid the constraints posed by a double-layered surrogate, where often times the algorithm’s local best position, as in PSO, is used to assemble a second local surrogate, which runs in parallel to the global surrogate, often requiring an independent DB(s), thus playing a key role in the refining stages of the search space process with an added computational cost. Therefore, in this work, only a global DB and surrogate is considered, and so we use the several control parameters to shift the balance of both the DB update and the assembling training data selection for the surrogate building as a way to move from a more global surrogate in the first phase to a more local surrogate in the last phase of the optimization.
To validate the performance of the proposed SASMA, we compared it against six well-regarded SAEAs on seventeen widely used benchmark mathematical functions (unconstrained optimization) with dimensionalities of 30 and 100, as well as with a set of MHs. Moreover, an additional case study, where a classical truss design (constrained) problem is considered, was included. The experimental results indicate that the proposed SASMA takes advantage of the SMA versatility as an effective population-based metaheuristic, coupled with a novel DB management strategy and surrogate building approach to accurately perform stochastic optimization. These unique traits explain why SASMA was able to outmatch or closely trail almost all the best results in terms of the mean, standard deviation, and minimum error, inheriting the good convergence properties of SMA, as can be seen in the convergence curves, with an accuracy that increases substantially around the 1000-FE mark. And, for the same token, this explains why it was best ranked in the vast majority of all test functions (Case Study I) in terms of both the Wilcoxon signed rank and Friedman test (mean rank). It is also worth noting that the additional overhead associated with database management and surrogate building mechanisms proved to be dominated by the surrogate’s least-squares fit and the DB distance computations, both scaling with the DB size and with the subset size. In comparison with the original SMA, the computational overhead of SASMA remained modest throughout our experiments and, importantly, was consistently in line with—or even more favorable than—the computation time profiles observed for the other SAEAs (in all test functions).
With fewer FEs, the increased initial phase variance explains why the good overall mean error is not matched with a similar performance in terms of STD, which is an expected consequence of the followed balanced approach. The shift from a global to a local surrogate nature, i.e., a more spread set of positions in this initial phase (when only a couple hundred FEs have passed) both in the SASMA swarm and in the DB, with the focus gradually changing (linearly) from favoring distant solutions towards quality solutions impacts the surrogate building. Meanwhile, from the half point onwards (with further evaluations), we can see the benefits of this balanced approach, with SASMA effectively moving towards the solution, particularly with 100D, which attests its adequacy for expensive optimization problems. Case Study II, the constrained optimization, with a smaller number of problem dimensions, further proved SASMA’s capabilities with competitive descriptive statistical and non-parametric results, outmatching all the tested metaheuristics, except for GSK. Finally, it is important to acknowledge certain limitations of the present study. The scalability analysis was conducted up to 100D, a commonly adopted upper benchmark in SAEA research, and our results confirm that SASMA preserves the performance trends previously observed with 30D, although exploring even higher dimensionalities remains an open direction for future work. Likewise, the adoption of a cubic RBF proved not to be an obstacle but rather a robust, parameter-free surrogate choice that aligns well with the dynamics of SASMA, although—as with any modelling option—future studies may explore alternative kernels such as Gaussian functions or even other surrogate models (e.g., Kriging).

Author Contributions

Conceptualization, P.B. and J.P.; methodology, P.B.; software, P.B. and H.N.; validation, J.P., H.N. and S.M.; formal analysis, P.B., J.P., M.C. and S.M.; investigation, P.B., J.P., H.N., M.C. and S.M.; visualization, H.N. and M.C.; supervision, J.P., M.C. and S.M.; writing—original draft preparation, P.B. and J.P.; writing—review and editing, H.N., M.C. and S.M. All authors have read and agreed to the published version of the manuscript.

Funding

Several preliminary conducive works reflected in this paper were supported by FCT/MCTES through national funds and, when applicable, were co-funded through EU funds under the project UIDB/50008/2020.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries regarding data or the code can be directed to the corresponding author.

Acknowledgments

Pedro Bento gives his special thanks to the Fundacao para a Ciencia e a Tecnologia (FCT), Portugal, for their Ph.D. Grant (SFRH/BD/140371/2018).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

As mentioned in Section 5.1, the mathematical test functions used for Case Study I, i.e., its expression, name, optimization range, and global minimum, are shown in Table A1.
The descriptive statistics and the results of the non-parametric Wilcoxon and Friedman tests for 100D, relative to Case Study I, are presented below in Table A2 and Table A3 for the primary test case, while the respective convergence curves are shown in Figure A1 for the unimodal test functions, Figure A2 for the multimodal test functions, and Figure A3 for the very complicated multimodal test functions. The control parameters of the benchmark algorithms used in Case Study II are shown in Table A4.
Table A1. Case Study I: Benchmark test functions.
Table A1. Case Study I: Benchmark test functions.
FunctionNameRange ( n ) f min
F 1 ( x ) = i = 1 n x i 2 Sphere [ 100 , 100 ] 0
F 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | Schwefel 2.22 [ 10 , 10 ] 0
F 3 ( x ) = i = 1 n j = 1 i x j 2 Schwefel 1.2 [ 100 , 100 ] 0
F 4 ( x ) = max i = 1 , , n | x i | Schwefel 2.21 [ 100 , 100 ] 0
F 5 ( x ) = i = 1 n 1 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 Rosenbrock [ 30 , 30 ] 0
F 6 ( x ) = i = 1 n | x i + 0.5 | 2 Step [ 100 , 100 ] 0
F 7 ( x ) = i = 1 n i x i 4 + α i , α i rand ( [ 0 , 1 ] ) Quartic [ 1.28 , 1.28 ] 0
F 8 ( x ) = i = 1 n x i sin | x i | Schwefel 2.26 [ 500 , 500 ] 418.9829 × n D i m s
F 9 ( x ) = i = 1 n 10 + x i 2 10 cos ( 2 π x i ) Rastrigin [ 5.12 , 5.12 ] 0
F 10 ( x ) = 20 e 0.2 1 n i = 1 n x i 2 e 1 n i = 1 n cos ( 2 π x i ) + 20 + e Ackley [ 32 , 32 ] 0
F 11 ( x ) = 1 + 1 4000 i = 1 n x i 2 i = 1 n cos x i i Griewank [ 600 , 600 ] 0
F 12 ( x ) = π n 10 sin ( 2 π y 1 ) + i = 1 n 1 ( y i 1 ) 2 1 + 10 sin 2 ( π y i + 1 ) + ( y n 1 ) 2 + i = 1 n u ( x i , 10 , 100 , 4 )
y i = 1 + x i + 1 4
u ( x i , a , k , m ) = k ( x i a ) m , x i > a 0 , a x i a k ( x i a ) m , x i < a
Penalized 1 [ 50 , 50 ] 0
F 13 ( x ) = 0.1 sin 2 ( 3 π x 1 ) + i = 1 n ( x i 1 ) 2 1 + sin 2 ( 3 π x i + 1 ) + ( x n 1 ) 2 1 + sin 2 ( 2 π x n )
+ i = 1 n u ( x i , 5 , 100 , 4 )
Penalized 2 [ 50 , 50 ] 0
F 14 ( x ) = i = 1 n i x i 2 Ellipsoid [ 100 , 100 ] 0
F 15 ( x ) = i = 1 n [ 10 + z i 2 10 cos ( 2 π z i ) ] + f b i a s
z = ( x o ) M , f b i a s = 330
Shifted Rotated Rastrigin
(F10 in [67])
[ 5 , 5 ] 330
F 16 ( CF ) : Rotated Hybrid Composition
f 1 , 2 = Rastrigin , f 3 , 4 = Weierstrass , f 5 , 6 = Griewank ,
f 7 , 8 = Ackley , f 9 , 10 = Sphere ; f b i a s = 120 ;
σ i = 1 ; λ = [ 1 , 1 , 10 , 10 , 5 / 60 , 5 / 60 , 5 / 32 , 5 / 32 , 5 / 100 , 5 / 100 ] ;
M i identity
Rotated Hybrid Composition
(F16 in [67])
[ 5 , 5 ] 120
F 17 ( CF ) : Rotated Hybrid Composition
f 1 , 2 = Ackley , f 3 , 4 = Rastrigin , f 5 , 6 = Sphere ,
f 7 , 8 = Weierstrass , f 9 , 10 = Griewank ; f b i a s = 10 ;
σ i = [ 1 , 2 , 1.5 , 1.5 , 1 , 1 , 1.5 , 1.5 , 2 , 2 ] ;
λ = [ 10 / 32 , 5 / 32 , 2 , 1 , 10 / 100 , 5 / 100 , 20 , 10 , 10 / 60 , 5 / 60 ] ;
M i are rotation matrices
Rotated Hybrid Composition
(F19 in [67])
[ 5 , 5 ] 10
Table A2. Statistical results of the proposed SASMA and the benchmark SAEAs on 100D mathematical test functions.
Table A2. Statistical results of the proposed SASMA and the benchmark SAEAs on 100D mathematical test functions.
Test Function/MetricGORS-SSLPSOSAE-ARFSSHPSOTL-SSLPSOTLSAPSOSASMA
Mean 1.429 × 10 3 1.065 × 10 5 6.281 × 10 1 7.452 × 10 3 1.243 × 10 5 9.287 × 10 29
F1St. Dev. 4.299 × 10 3 + 1.121 × 10 4 + 1.953 × 10 1 + 3.197 × 10 3 + 1.531 × 10 4 + 2.344 × 10 28
Min 1.069 × 10 4 7.513 × 10 4 2.863 × 10 1 2.312 × 10 3 9.326 × 10 4 3.395 × 10 39
Mean 6.761 × 10 31 2.640 × 10 9 1.826 × 10 15 3.132 × 10 12 2.640 × 10 38 6.749 × 10 15
F2St. Dev. 4.000 × 10 32 + 1.449 × 10 10 + 8.767 × 10 15 + 1.799 × 10 13 + 7.400 × 10 38 + 1.480 × 10 14
Min 1.052 × 10 2 2.908 × 10 2 1.578 × 10 2 2.439 × 10 2 3.492 × 10 23 1.595 × 10 20
Mean 1.442 × 10 5 4.064 × 10 5 2.504 × 10 5 2.285 × 10 5 3.963 × 10 5 9.798 × 10 16
F3St. Dev. 7.018 × 10 4 + 6.533 × 10 4 + 5.408 × 10 4 + 5.307 × 10 4 + 7.280 × 10 4 + 4.145 × 10 15
Min 4.291 × 10 4 2.742 × 10 5 1.417 × 10 5 1.212 × 10 5 2.500 × 10 5 2.371 × 10 33
Mean 7.814 × 10 1 8.319 × 10 1 4.313 × 10 1 7.757 × 10 1 9.266 × 10 1 7.664 × 10 14
F4St. Dev. 3.531 × 10 0 + 2.529 × 10 0 + 5.503 × 10 0 + 4.761 × 10 0 + 3.424 × 10 0 + 1.466 × 10 13
Min 6.999 × 10 1 7.760 × 10 1 3.173 × 10 1 6.773 × 10 1 8.292 × 10 1 1.066 × 10 16
Mean 2.020 × 10 5 2.977 × 10 8 4.839 × 10 4 2.669 × 10 6 6.314 × 10 8 9.651 × 10 1
F5St. Dev. 3.715 × 10 4 + 6.150 × 10 7 + 3.117 × 10 4 + 2.023 × 10 6 + 8.800 × 10 7 + 1.046 × 10 1
Min 1.254 × 10 5 1.881 × 10 8 1.328 × 10 4 6.084 × 10 5 4.451 × 10 8 4.343 × 10 1
Mean 1.426 × 10 3 1.084 × 10 5 6.607 × 10 1 7.415 × 10 3 1.233 × 10 5 1.750 × 10 1
F6St. Dev. 3.544 × 10 3 + 1.120 × 10 4 + 3.117 × 10 1 2.875 × 10 3 + 1.226 × 10 4 + 6.667 × 10 0
Min 8.937 × 10 5 8.731 × 10 4 2.933 × 10 1 3.330 × 10 3 8.511 × 10 4 9.927 × 10 2
Mean 1.826 × 10 0 3.976 × 10 2 1.133 × 10 0 1.070 × 10 1 1.070 × 10 3 2.193 × 10 3
F7St. Dev. 3.753 × 10 1 + 6.799 × 10 1 + 4.985 × 10 1 + 7.321 × 10 0 + 1.514 × 10 2 + 1.736 × 10 3
Min 1.146 × 10 0 2.651 × 10 2 5.251 × 10 1 3.419 × 10 0 7.234 × 10 2 6.724 × 10 5
Mean 1.946 × 10 4 2.264 × 10 4 3.159 × 10 4 2.211 × 10 4 3.020 × 10 4 3.056 × 10 3
F8St. Dev. 1.683 × 10 3 + 1.095 × 10 3 + 1.512 × 10 3 + 2.012 × 10 3 + 1.420 × 10 3 + 4.036 × 10 3
Min 1.567 × 10 4 2.035 × 10 4 2.854 × 10 4 1.865 × 10 4 2.790 × 10 4 1.903 × 10 1
Mean 3.447 × 10 2 1.042 × 10 3 7.663 × 10 2 5.165 × 10 2 1.423 × 10 3 4.582 × 10 6
F9St. Dev. 6.295 × 10 1 + 5.005 × 10 1 + 1.009 × 10 2 + 6.831 × 10 1 + 6.909 × 10 1 + 1.280 × 10 5
Min 2.129 × 10 2 9.500 × 10 2 5.722 × 10 2 4.049 × 10 2 1.235 × 10 3 0.000 × 10 0
Mean 1.861 × 10 1 1.951 × 10 1 4.199 × 10 0 1.575 × 10 1 2.039 × 10 1 1.537 × 10 14
F10St. Dev. 5.462 × 10 1 + 2.553 × 10 1 + 5.341 × 10 1 + 1.145 × 10 0 + 1.293 × 10 1 + 3.398 × 10 14
Min 1.670 × 10 1 1.897 × 10 1 3.444 × 10 0 1.284 × 10 1 2.006 × 10 1 4.441 × 10 16
Mean 1.570 × 10 1 9.917 × 10 2 9.750 × 10 1 6.393 × 10 1 9.979 × 10 1 0.000 × 10 0
F11St. Dev. 4.100 × 10 1 + 8.318 × 10 1 + 4.646 × 10 2 + 2.747 × 10 1 + 1.390 × 10 1 + 0.000 × 10 0
Min 7.129 × 10 2 8.201 × 10 2 8.485 × 10 1 2.568 × 10 1 7.361 × 10 1 0.000 × 10 0
Mean 2.859 × 10 2 5.301 × 10 8 2.988 × 10 3 1.379 × 10 6 1.392 × 10 9 5.005 × 10 1
F12St. Dev. 6.397 × 10 2 + 1.440 × 10 8 + 8.741 × 10 3 + 1.759 × 10 6 + 2.812 × 10 8 + 4.198 × 10 1
Min 1.425 × 10 1 2.186 × 10 8 8.739 × 10 0 3.766 × 10 4 6.339 × 10 8 4.151 × 10 4
Mean 6.285 × 10 4 1.173 × 10 9 2.135 × 10 4 5.801 × 10 6 2.573 × 10 9 6.895 × 10 0
F13St. Dev. 2.217 × 10 4 + 2.290 × 10 8 + 5.463 × 10 4 + 5.531 × 10 6 + 4.791 × 10 8 + 4.174 × 10 0
Min 1.120 × 10 4 6.274 × 10 8 1.272 × 10 2 4.045 × 10 5 1.666 × 10 9 5.662 × 10 2
Mean 7.918 × 10 3 4.391 × 10 6 1.662 × 10 4 3.397 × 10 5 5.543 × 10 6 9.533 × 10 26
F14St. Dev. 1.857 × 10 4 + 4.554 × 10 5 + 3.097 × 10 3 + 1.615 × 10 5 + 6.585 × 10 5 + 3.178 × 10 25
Min 1.041 × 10 3 3.447 × 10 6 1.041 × 10 4 1.650 × 10 5 4.357 × 10 6 1.228 × 10 40
Mean 1.464 × 10 3 2.480 × 10 3 1.172 × 10 3 1.491 × 10 3 2.519 × 10 3 2.315 × 10 3
F15St. Dev. 1.058 × 10 2 2.535 × 10 2 + 9.584 × 10 1 1.312 × 10 2 2.345 × 10 2 + 9.103 × 10 1
Min 1.272 × 10 3 1.995 × 10 3 1.047 × 10 3 1.223 × 10 3 2.017 × 10 3 2.080 × 10 3
Mean 6.388 × 10 2 7.446 × 10 2 3.989 × 10 2 4.961 × 10 2 9.037 × 10 2 9.161 × 10 2
F16St. Dev. 6.302 × 10 1 7.359 × 10 1 2.929 × 10 1 5.488 × 10 1 9.710 × 10 1 = 8.486 × 10 1
Min 5.078 × 10 2 6.188 × 10 2 3.517 × 10 2 3.895 × 10 2 7.458 × 10 2 7.173 × 10 2
Mean 1.483 × 10 3 1.482 × 10 3 1.409 × 10 3 1.407 × 10 3 1.500 × 10 3 9.000 × 10 2
F17St. Dev. 4.642 × 10 1 + 3.113 × 10 1 + 3.634 × 10 1 + 3.282 × 10 1 + 3.949 × 10 1 + 0.000 × 10 0
Min 1.364 × 10 3 1.428 × 10 3 1.331 × 10 3 1.333 × 10 3 1.435 × 10 3 9.000 × 10 2
Win15 16 14 15 16
Tie0 0 0 0 1
Lose2 1 3 2 0
According to the Wilcoxon signed rank test at a 5% significance level, the symbol “+”, “=”, or “−” symbolizes that the performance of SASMA is better, similar, or worse than that of other SAEAs, respectively. In addition, the best results for all metrics (from the 35 runs) for each test function are highlighted in grey.
Table A3. Mean rank scores (based on the Friedman test) of the SASMA and the benchmark SAEAs on 100D mathematical test functions.
Table A3. Mean rank scores (based on the Friedman test) of the SASMA and the benchmark SAEAs on 100D mathematical test functions.
Friedman TestGORS-SSLPSOSAE-ARFSSHPSOTL-SSLPSOTLSAPSOSASMA
Mean RankF12.315.232.773.915.771.00
F22.113.714.004.176.001.00
F32.265.493.463.405.401.00
F43.544.802.003.715.941.00
F53.005.002.004.006.001.00
F61.435.141.913.895.862.77
F72.865.002.144.006.001.00
F82.173.575.773.265.231.00
F92.005.003.943.066.001.00
F104.064.942.003.006.001.00
F112.376.002.864.034.741.00
F122.715.002.294.006.001.00
F132.915.002.094.006.001.00
F142.065.092.944.005.911.00
F152.405.201.062.545.314.49
F163.094.001.062.005.345.51
F174.634.742.802.695.141.00
According to the Friedman test, the best mean rank (from the 35 runs) for each function is highlighted in grey.
Figure A1. Convergence curves (fitness value): GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 100D unimodal test functions (F1–F4, F6, F7, and F14).
Figure A1. Convergence curves (fitness value): GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 100D unimodal test functions (F1–F4, F6, F7, and F14).
Algorithms 19 00265 g0a1
Table A4. Parameter settings for the benchmark algorithms applied in the second case study.
Table A4. Parameter settings for the benchmark algorithms applied in the second case study.
AlgorithmParameter Settings
PSO w/inertia factor ω min = 0.4 , ω max = 0.9 , c 1 = c 2 = 2.05
PSO w/constriction factor c 1 = c 2 = 2.05
WOA p = 0.5 , a linearly decreasing from 2 to 0
GWO a linearly decreasing from 2 to 0
GSA α = 20 , G 0 = 100 , R power = 1
R norm = 2 , ElitistCheck = 1 ( True )
FPA β = 1.5 , s 0 = 0.01 , p = 0.8
BA f min = 0 , f max = 2 , α = 0.95
r min = r 0 = 0.25 , γ = 0.015
GSK K = 10 , K r = 0.9 , K f = 0.5 , p = 0.1
SASMA z = 0.03
Figure A2. Convergence curves (fitness value): GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple) and SASMA (solid pink) on 100D multimodal test functions (F5, F8 and F10–F13).
Figure A2. Convergence curves (fitness value): GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple) and SASMA (solid pink) on 100D multimodal test functions (F5, F8 and F10–F13).
Algorithms 19 00265 g0a2
Figure A3. Convergence curves (fitness value): GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 100D very complicated multimodal test functions (F9 and F15–F17).
Figure A3. Convergence curves (fitness value): GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 100D very complicated multimodal test functions (F9 and F15–F17).
Algorithms 19 00265 g0a3

Appendix B

In a truss design problem, the objective is to find a cross-sectional area for each member, namely a fixed truss geometry, ensuring its structural integrity and minimizing the total weight of the structure [74]. Cross-sectional areas constitute the design variables and are picked from a list of allowable sections. As such, this constrained optimization problem, marked as Equation (A1), can be generically formulated as follows [75]:
minimize W X = u = 1 z ρ u X u L u subject to σ min σ u σ max , u 1 , z δ min δ k δ max , k 1 , n X u allowable section
where X denotes the truss design variables, i.e., the members of the cross-sectional areas; therefore, W X represents the weight of the studied truss; z is the number of members; ρ u is the material density; L u is the length of the members; σ min and σ max stand for the lower and upper stress limits, respectively; σ u is the stress of the member u; n is the number of nodes; δ min and δ max stand for the lower and upper displacement limits, respectively; δ k is the displacement/deflection of the node k; and X u denotes the cross-sectional area of member u, and as stated, this value comes from an allowable list of sections that are ordered in an ascending manner. To deal with these restrictions, a penalty function is commonly used to transform this into an unconstrained optimization problem. The infeasible truss designs are penalized by multiplying the objective function (weight) by a cumulative penalty that stands for the sum of the stress Φ σ and displacement Φ δ violations. Thus, we transform the problem given in Equation (A1) to an unconstrained one, P, using Equation (A2).
minimize P X , Φ σ , Φ δ = W X 1 + ξ 1 Φ σ + Φ δ ξ 2
where ξ 1 and ξ 2 are the penalty function coefficients, which in this context are usually set to 1; Φ σ = u = 1 z Φ u σ is the total stress violation term; and Φ δ = k = 1 n Φ k δ is the total displacement penalty. The individual stress and displacement penalties, Φ u σ and Φ k δ , can be the computed according to the piecewise expressions presented in Equation (A3).
Φ u σ = 0 , σ min σ u σ max σ min σ u σ min , σ u < σ min σ max σ u σ max , σ u > σ max Φ k δ = 0 , δ min δ k δ max δ min δ k δ min , δ k < δ min δ max δ k δ max , δ k > δ max

References

  1. Liu, W.; Wang, J. Recursive elimination current algorithms and a distributed computing scheme to accelerate wrapper feature selection. Inf. Sci. 2022, 589, 636–654. [Google Scholar] [CrossRef]
  2. Ozbay, F.A.; Alatas, B. A Novel Approach for Detection of Fake News on Social Media Using Metaheuristic Optimization Algorithms. Elektron. Elektrotech. 2019, 25, 62–67. [Google Scholar] [CrossRef]
  3. Yang, X.S. Nature-Inspired Optimization Algorithms, 2nd ed.; Academic Press: Cambridge, MA, USA, 2020; pp. 1–310. [Google Scholar] [CrossRef]
  4. Molina, D.; LaTorre, A.; Herrera, F. An Insight into Bio-inspired and Evolutionary Algorithms for Global Optimization: Review, Analysis, and Lessons Learnt over a Decade of Competitions. Cogn. Comput. 2018, 10, 517–544. [Google Scholar] [CrossRef]
  5. Gutjahr, W.J.; Montemanni, R. Stochastic Search in Metaheuristics. Int. Ser. Oper. Res. Manag. Sci. 2019, 272, 513–540. [Google Scholar] [CrossRef]
  6. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of ICNN’95—International Conference on Neural Networks 1995; IEEE: Piscataway, NJ, USA, 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
  7. Yang, X.S.; Deb, S. Cuckoo search: Recent advances and applications. Neural Comput. Appl. 2014, 24, 169–174. [Google Scholar] [CrossRef]
  8. Das, S.; Mullick, S.S.; Suganthan, P.N. Recent advances in differential evolution—An updated survey. Swarm Evol. Comput. 2016, 27, 1–30. [Google Scholar] [CrossRef]
  9. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  10. Li, Y.; Lin, X.; Liu, J. An improved gray wolf optimization algorithm to solve engineering problems. Sustainability 2021, 13, 3208. [Google Scholar] [CrossRef]
  11. Piotrowski, A.P.; Napiorkowski, J.J.; Piotrowska, A.E. Population size in Particle Swarm Optimization. Swarm Evol. Comput. 2020, 58, 100718. [Google Scholar] [CrossRef]
  12. Morales-Castañeda, B.; Zaldívar, D.; Cuevas, E.; Fausto, F.; Rodríguez, A. A better balance in metaheuristic algorithms: Does it exist? Swarm Evol. Comput. 2020, 54, 100671. [Google Scholar] [CrossRef]
  13. Sun, C.; Jin, Y.; Zeng, J.; Yu, Y. A two-layer surrogate-assisted particle swarm optimization algorithm. Soft Comput. 2015, 19, 1461–1475. [Google Scholar] [CrossRef]
  14. Tang, Z.; Xu, L.; Luo, S. Adaptive dynamic surrogate-assisted evolutionary computation for high-fidelity optimization in engineering. Appl. Soft Comput. 2022, 127, 109333. [Google Scholar] [CrossRef]
  15. Srithapon, C.; Fuangfoo, P.; Ghosh, P.K.; Siritaratiwat, A.; Chatthaworn, R. Surrogate-Assisted Multi-Objective Probabilistic Optimal Power Flow for Distribution Network with Photovoltaic Generation and Electric Vehicles. IEEE Access 2021, 9, 34395–34414. [Google Scholar] [CrossRef]
  16. Tong, H.; Huang, C.; Minku, L.L.; Yao, X. Surrogate models in evolutionary single-objective optimization: A new taxonomy and experimental study. Inf. Sci. 2021, 562, 414–437. [Google Scholar] [CrossRef]
  17. Li, F.; Li, Y.; Cai, X.; Gao, L. A surrogate-assisted hybrid swarm optimization algorithm for high-dimensional computationally expensive problems. Swarm Evol. Comput. 2022, 72, 101096. [Google Scholar] [CrossRef]
  18. Zhou, Z.; Ong, Y.S.; Nguyen, M.H.; Lim, D. A study on polynomial regression and Gaussian Process global surrogate model in hierarchical surrogate-assisted evolutionary algorithm. In 2005 IEEE Congress on Evolutionary Computation; IEEE: Piscataway, NJ, USA, 2005; Volume 3, pp. 2832–2839. [Google Scholar] [CrossRef]
  19. Han, D.; Du, W.; Wang, X.; Du, W. A surrogate-assisted evolutionary algorithm for expensive many-objective optimization in the refining process. Swarm Evol. Comput. 2022, 69, 100988. [Google Scholar] [CrossRef]
  20. Liu, Q.; Jin, Y.; Heiderich, M.; Rodemann, T. Surrogate-assisted evolutionary optimization of expensive many-objective irregular problems. Knowl.-Based Syst. 2022, 240, 108197. [Google Scholar] [CrossRef]
  21. Tian, J.; Sun, C.; Tan, Y.; Zeng, J. Granularity-based surrogate-assisted particle swarm optimization for high-dimensional expensive optimization. Knowl.-Based Syst. 2020, 187, 104815. [Google Scholar] [CrossRef]
  22. Clarke, S.M.; Griebsch, J.H.; Simpson, T.W. Analysis of support vector regression for approximation of complex engineering analyses. J. Mech. Des. Trans. ASME 2005, 127, 1077–1087. [Google Scholar] [CrossRef]
  23. Volz, V.; Rudolph, G.; Naujoks, B. Investigating uncertainty propagation in surrogate-assisted evolutionary algorithms. In GECCO ’17: Proceedings of the Genetic and Evolutionary Computation Conference; Association for Computing Machinery, Inc.: New York, NY, USA, 2017; Volume 8, pp. 881–888. [Google Scholar] [CrossRef]
  24. Liu, J.; Wang, Y.; Sun, G.; Pang, T. Multisurrogate-Assisted Ant Colony Optimization for Expensive Optimization Problems With Continuous and Categorical Variables. IEEE Trans. Cybern. 2021, 52, 11348–11361. [Google Scholar] [CrossRef]
  25. Cai, X.; Qiu, H.; Gao, L.; Jiang, C.; Shao, X. An efficient surrogate-assisted particle swarm optimization algorithm for high-dimensional expensive problems. Knowl.-Based Syst. 2019, 184, 104901. [Google Scholar] [CrossRef]
  26. Eason, J.; Cremaschi, S. Adaptive sequential sampling for surrogate model generation with artificial neural networks. Comput. Chem. Eng. 2014, 68, 220–232. [Google Scholar] [CrossRef]
  27. Zhang, T.; Li, F.; Zhao, X.; Qi, W.; Liu, T. A Convolutional Neural Network-Based Surrogate Model for Multi-objective Optimization Evolutionary Algorithm Based on Decomposition. Swarm Evol. Comput. 2022, 72, 101081. [Google Scholar] [CrossRef]
  28. Wang, W.; Liu, H.L.; Tan, K.C. A Surrogate-Assisted Differential Evolution Algorithm for High-Dimensional Expensive Optimization Problems. IEEE Trans. Cybern. 2022, 53, 2685–2697. [Google Scholar] [CrossRef] [PubMed]
  29. Chen, C.; Wang, X.; Dong, H.; Wang, P. Surrogate-assisted hierarchical learning water cycle algorithm for high-dimensional expensive optimization. Swarm Evol. Comput. 2022, 75, 101169. [Google Scholar] [CrossRef]
  30. Yu, M.; Liang, J.; Zhao, K.; Wu, Z. An aRBF surrogate-assisted neighborhood field optimizer for expensive problems. Swarm Evol. Comput. 2022, 68, 100972. [Google Scholar] [CrossRef]
  31. Liu, Y.; Liu, J.; Tan, S. Decision space partition based surrogate-assisted evolutionary algorithm for expensive optimization. Expert Syst. Appl. 2023, 214, 119075. [Google Scholar] [CrossRef]
  32. Chu, S.C.; Du, Z.G.; Peng, Y.J.; Pan, J.S. Fuzzy Hierarchical Surrogate Assists Probabilistic Particle Swarm Optimization for expensive high dimensional problem. Knowl.-Based Syst. 2021, 220, 106939. [Google Scholar] [CrossRef]
  33. Li, H.; Chen, L.; Zhang, J.; Li, M. A Multi-Surrogate Assisted Multi-Tasking Optimization Algorithm for High-Dimensional Expensive Problems. Algorithms 2025, 18, 4. [Google Scholar] [CrossRef]
  34. Pan, J.S.; Zhang, L.G.; Chu, S.C.; Shieh, C.S.; Watada, J. Surrogate-Assisted Hybrid Meta-Heuristic Algorithm with an Add-Point Strategy for a Wireless Sensor Network. Entropy 2023, 25, 317. [Google Scholar] [CrossRef]
  35. Chen, W.; Dong, H.; Wang, P.; Wang, X. Surrogate-assisted global transfer optimization based on adaptive sampling strategy. Adv. Eng. Inform. 2023, 56, 101914. [Google Scholar] [CrossRef]
  36. Younis, A.; Dong, Z. High-Fidelity Surrogate Based Multi-Objective Optimization Algorithm. Algorithms 2022, 15, 279. [Google Scholar] [CrossRef]
  37. Dong, H.; Wang, P.; Yu, X.; Song, B. Surrogate-assisted teaching-learning-based optimization for high-dimensional and computationally expensive problems. Appl. Soft Comput. 2021, 99, 106934. [Google Scholar] [CrossRef]
  38. Chen, G.; Zhang, K.; Xue, X.; Zhang, L.; Yao, C.; Wang, J.; Yao, J. A radial basis function surrogate model assisted evolutionary algorithm for high-dimensional expensive optimization problems. Appl. Soft Comput. 2022, 116, 108353. [Google Scholar] [CrossRef]
  39. Yu, H.; Tan, Y.; Sun, C.; Zeng, J. A generation-based optimal restart strategy for surrogate-assisted social learning particle swarm optimization. Knowl.-Based Syst. 2019, 163, 14–25. [Google Scholar] [CrossRef]
  40. Yu, H.; Tan, Y.; Zeng, J.; Sun, C.; Jin, Y. Surrogate-assisted hierarchical particle swarm optimization. Inf. Sci. 2018, 454–455, 59–72. [Google Scholar] [CrossRef]
  41. Hu, P.; Pan, J.S.; Chu, S.C.; Sun, C. Multi-surrogate assisted binary particle swarm optimization algorithm and its application for feature selection. Appl. Soft Comput. 2022, 121, 108736. [Google Scholar] [CrossRef]
  42. Dong, H.; Li, X.; Yang, Z.; Gao, L.; Lu, Y. A two-layer surrogate-assisted differential evolution with better and nearest option for optimizing the spring of hydraulic series elastic actuator. Appl. Soft Comput. 2021, 100, 107001. [Google Scholar] [CrossRef]
  43. Ji, X.; Zhang, Y.; Gong, D.; Sun, X. Dual-Surrogate-Assisted Cooperative Particle Swarm Optimization for Expensive Multimodal Problems. IEEE Trans. Evol. Comput. 2021, 25, 794–808. [Google Scholar] [CrossRef]
  44. Li, F.; Shen, W.; Cai, X.; Gao, L.; Wang, G.G. A fast surrogate-assisted particle swarm optimization algorithm for computationally expensive problems. Appl. Soft Comput. 2020, 92, 106303. [Google Scholar] [CrossRef]
  45. Zhao, F.; Zhang, H.; Wang, L.; Ma, R.; Xu, T.; Zhu, N.; Jonrinaldi. A surrogate-assisted Jaya algorithm based on optimal directional guidance and historical learning mechanism. Eng. Appl. Artif. Intell. 2022, 111, 104775. [Google Scholar] [CrossRef]
  46. Pan, J.S.; Liu, N.; Chu, S.C.; Lai, T. An efficient surrogate-assisted hybrid optimization algorithm for expensive optimization problems. Inf. Sci. 2021, 561, 304–325. [Google Scholar] [CrossRef]
  47. Loshchilov, I. Surrogate-Assisted Evolutionary Algorithms. Ph.D. Thesis, Université Paris Sud–Paris XI, Paris, France, 2013. [Google Scholar]
  48. Dong, H.; Dong, Z. Surrogate-assisted grey wolf optimization for high-dimensional, computationally expensive black-box problems. Swarm Evol. Comput. 2020, 57, 100713. [Google Scholar] [CrossRef]
  49. Dash, C.S.K.; Behera, A.K.; Dehuri, S.; Cho, S.B. Radial basis function neural networks: A topical state-of-the-art survey. Open Comput. Sci. 2016, 6, 33–63. [Google Scholar] [CrossRef]
  50. Cavoretto, R.; Rossi, A.D.; Mukhametzhanov, M.S.; Sergeyev, Y.D. On the search of the shape parameter in radial basis functions using univariate global optimization methods. J. Glob. Optim. 2021, 79, 305–327. [Google Scholar] [CrossRef]
  51. Zendehboudi, A.; Saidur, R.; Mahbubul, I.M.; Hosseini, S.H. Data-driven methods for estimating the effective thermal conductivity of nanofluids: A comprehensive review. Int. J. Heat Mass Transf. 2019, 131, 1211–1231. [Google Scholar] [CrossRef]
  52. Bagheri, S.; Konen, W.; Bäck, T. Comparing Kriging and Radial Basis Function Surrogates. In Proceedings of the Proceedings-27. Workshop Computational Intelligence; Hoffmann, F., Huellermeier, E., Mikut, R., Eds.; Scientific Publishing: Singapore, 2017; pp. 243–259. [Google Scholar]
  53. Vahabli, E.; Rahmati, S. Application of an RBF neural network for FDM parts’ surface roughness prediction for enhancing surface quality. Int. J. Precis. Eng. Manuf. 2016, 17, 1589–1603. [Google Scholar] [CrossRef]
  54. Demuth, H.B.; Beale, M.H.; Jess, O.D.; Hagan, M.T. Neural Network Design, 2nd ed.; Martin Hagan: Cramlington, UK, 2014; p. 800. [Google Scholar]
  55. Jawad, J.; Hawari, A.H.; Javaid Zaidi, S. Artificial neural network modeling of wastewater treatment and desalination using membrane processes: A review. Chem. Eng. J. 2021, 419, 129540. [Google Scholar] [CrossRef]
  56. Bornatico, R.; Hüssy, J.; Witzig, A.; Guzzella, L. Surrogate modeling for the fast optimization of energy systems. Energy 2013, 57, 653–662. [Google Scholar] [CrossRef]
  57. Regis, R.G. Multi-objective constrained black-box optimization using radial basis function surrogates. J. Comput. Sci. 2016, 16, 140–155. [Google Scholar] [CrossRef]
  58. Cavoretto, R.; Rossi, A.D.; Lancellotti, S. Bayesian approach for radial kernel parameter tuning. J. Comput. Appl. Math. 2024, 441, 115716. [Google Scholar] [CrossRef]
  59. Li, S.; Chen, H.; Wang, M.; Heidari, A.A.; Mirjalili, S. Slime mould algorithm: A new method for stochastic optimization. Future Gener. Comput. Syst. 2020, 111, 300–323. [Google Scholar] [CrossRef]
  60. Clerc, M. Confinements and Biases in Particle Swarm Optimization; HAL Open Science: Villeurbanne, France, 2006. [Google Scholar]
  61. Regis, R.G.; Shoemaker, C.A. A Stochastic Radial Basis Function Method for the Global Optimization of Expensive Functions. Informs J. Comput. 2007, 19, 497–509. [Google Scholar] [CrossRef]
  62. Fu, G.; Sun, C.; Tan, Y.; Zhang, G.; Jin, Y. A surrogate-assisted evolutionary algorithm with random feature selection for large-scale expensive problems. In Parallel Problem Solving from Nature—PPSN XVI; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020; Volume 12269, pp. 125–139. [Google Scholar] [CrossRef]
  63. Yu, H.; Kang, L.; Tan, Y.; Sun, C.; Zeng, J. Truncation-learning-driven surrogate assisted social learning particle swarm optimization for computationally expensive problem. Appl. Soft Comput. 2020, 97, 106812. [Google Scholar] [CrossRef]
  64. Wang, H.; Jin, Y.; Doherty, J. Committee-Based Active Learning for Surrogate-Assisted Particle Swarm Optimization of Expensive Problems. IEEE Trans. Cybern. 2017, 47, 2664–2677. [Google Scholar] [CrossRef]
  65. Bento, P. SASMA Benchmark Code. 2023. Available online: https://gitfront.io/r/pedrobento/hWud6E83nsJt/SASMAsims/ (accessed on 11 January 2026).
  66. Kassoul, K.; Zufferey, N.; Cheikhrouhou, N.; Belhaouari, S.B. Exponential Particle Swarm Optimization for Global Optimization. IEEE Access 2022, 10, 78320–78344. [Google Scholar] [CrossRef]
  67. Suganthan, P.N.; Hansen, N.; Liang, J.J.; Deb, K.; Chen, Y.P.; Auger, A.; Tiwari, S. Problem Definitions and Evaluation Criteria for the CEC 2005 Special Session on Real-Parameter Optimization; Technical Report January; Nanyang Technological University: Singapore, 2005. [Google Scholar]
  68. Ma, J.; Li, H.; Ma, J.; Li, H. Research on Rosenbrock Function Optimization Problem Based on Improved Differential Evolution Algorithm. J. Comput. Commun. 2019, 7, 107–120. [Google Scholar] [CrossRef]
  69. Guo, Z.; Huang, H.; Deng, C.; Yue, X.; Wu, Z. An Enhanced Differential Evolution with Elite Chaotic Local Search. Comput. Intell. Neurosci. 2015, 2015, 583759. [Google Scholar] [CrossRef]
  70. Bala, I.; Yadav, A. Gravitational Search Algorithm: A State-of-the-Art Review. In Proceedings of the Harmony Search and Nature Inspired Optimization Algorithms; Yadav, N., Yadav, A., Bansal, J.C., Deep, K., Kim, J.H., Eds.; Springer: Singapore, 2019; pp. 27–37. [Google Scholar]
  71. Abdel-Basset, M.; Shawky, L.A. Flower pollination algorithm: A comprehensive review. Artif. Intell. Rev. 2019, 52, 2533–2557. [Google Scholar] [CrossRef]
  72. Fister, I.; Fister, I.; Yang, X.S.; Fong, S.; Zhuang, Y. Bat algorithm: Recent advances. In Proceedings of the 2014 IEEE 15th International Symposium on Computational Intelligence and Informatics (CINTI); IEEE: Piscataway, NJ, USA, 2014; pp. 163–167. [Google Scholar] [CrossRef]
  73. Mohamed, A.W.; Abutarboush, H.F.; Hadi, A.A.; Mohamed, A.K. Gaining-Sharing Knowledge Based Algorithm with Adaptive Parameters for Engineering Optimization. IEEE Access 2021, 9, 65934–65946. [Google Scholar] [CrossRef]
  74. Camp, C.V.; Farshchin, M. Design of space trusses using modified teaching–learning based optimization. Eng. Struct. 2014, 62–63, 87–97. [Google Scholar] [CrossRef]
  75. Ghannadiasl, A.; Zarbilinezhad, M. CGO and SNS Optimization Algorithm for the Structures with Discontinuous and Continuous Variables. Comput. Intell. Neurosci. 2022, 2022, 4211707. [Google Scholar] [CrossRef]
Figure 1. Online surrogate model working scheme (generic).
Figure 1. Online surrogate model working scheme (generic).
Algorithms 19 00265 g001
Figure 2. Surrogate model: smoothing effect. (The × symbols denote fictitious approximated points (“curse of uncertainty”), whereas the + symbols indicate matching "real" approximated points (“bless of uncertainty”).
Figure 2. Surrogate model: smoothing effect. (The × symbols denote fictitious approximated points (“curse of uncertainty”), whereas the + symbols indicate matching "real" approximated points (“bless of uncertainty”).
Algorithms 19 00265 g002
Figure 3. RBFNN architecture.
Figure 3. RBFNN architecture.
Algorithms 19 00265 g003
Figure 4. Slime mould foraging mechanism.
Figure 4. Slime mould foraging mechanism.
Algorithms 19 00265 g004
Figure 5. SMA: Balance between different space search mechanisms. The dots represent the positional behavior of the swarm for each problem dimension, and the red dot denotes X b e s t , d .
Figure 5. SMA: Balance between different space search mechanisms. The dots represent the positional behavior of the swarm for each problem dimension, and the red dot denotes X b e s t , d .
Algorithms 19 00265 g005
Figure 6. Proposed Surrogate-Assisted SMA (SASMA): Overview of the flow between the main mechanisms.
Figure 6. Proposed Surrogate-Assisted SMA (SASMA): Overview of the flow between the main mechanisms.
Algorithms 19 00265 g006
Figure 7. Detailed flowchart of the global database management: DB update and integrity check mechanisms.
Figure 7. Detailed flowchart of the global database management: DB update and integrity check mechanisms.
Algorithms 19 00265 g007
Figure 8. Selected subspace for surrogate modeling (spread coefficient effect).
Figure 8. Selected subspace for surrogate modeling (spread coefficient effect).
Algorithms 19 00265 g008
Figure 9. Convergence curves (fitness value): CALSAPSO (dashed orange), GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 30D unimodal test functions (F1–F4, F6, F7, and F14).
Figure 9. Convergence curves (fitness value): CALSAPSO (dashed orange), GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 30D unimodal test functions (F1–F4, F6, F7, and F14).
Algorithms 19 00265 g009
Figure 10. Convergence curves (fitness value): CALSAPSO (dashed orange), GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 30D multimodal test functions (F5, F8, and F10–F13).
Figure 10. Convergence curves (fitness value): CALSAPSO (dashed orange), GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 30D multimodal test functions (F5, F8, and F10–F13).
Algorithms 19 00265 g010
Figure 11. Convergence curves (fitness value): CALSAPSO (dashed orange), GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 30D very complicated multimodal test functions (F9 and F15–F17).
Figure 11. Convergence curves (fitness value): CALSAPSO (dashed orange), GORS-SSLPSO (dashed red), SAE-ARFS (dashed blue), SHPSO (dashed cyan), TL-SSLPSO (dashed brown), TLSAPSO (dashed purple), and SASMA (solid pink) on 30D very complicated multimodal test functions (F9 and F15–F17).
Algorithms 19 00265 g011
Figure 12. Relative time comparison between SASMA and the other SAEAs for 30D and 100D.
Figure 12. Relative time comparison between SASMA and the other SAEAs for 30D and 100D.
Algorithms 19 00265 g012
Figure 13. SASMA database: Result analysis for 30D (first line) and 100D (second line).
Figure 13. SASMA database: Result analysis for 30D (first line) and 100D (second line).
Algorithms 19 00265 g013
Figure 14. Configuration of the 25-bar truss. The numbers represent the node (joint) labels.
Figure 14. Configuration of the 25-bar truss. The numbers represent the node (joint) labels.
Algorithms 19 00265 g014
Figure 15. Convergence curves (fitness value) of the algorithms PSO w/inertia (dot-dashed grey), PSO w/constr (dot-dashed red), WOA (dot-dashed cyan), GWO (dot-dashed blue), GSA (dot-dashed light green), FPA (dot-dashed dark green), BA (dot-dashed brown), GSK (dot-dashed orange), SMA (dashed pink), and SASMA (solid purple) on the 25 truss bar continuous design problem.
Figure 15. Convergence curves (fitness value) of the algorithms PSO w/inertia (dot-dashed grey), PSO w/constr (dot-dashed red), WOA (dot-dashed cyan), GWO (dot-dashed blue), GSA (dot-dashed light green), FPA (dot-dashed dark green), BA (dot-dashed brown), GSK (dot-dashed orange), SMA (dashed pink), and SASMA (solid purple) on the 25 truss bar continuous design problem.
Algorithms 19 00265 g015
Table 1. Parameter settings for the benchmark algorithms and the proposed SASMA.
Table 1. Parameter settings for the benchmark algorithms and the proposed SASMA.
AlgorithmParameter Settings
TLSAPSO ω = 0.7 , c 1 = c 2 = 2.05 , v min = 100 , v max = 100 f i t t r e s h o l d = 10 4 , n b t r e s h o l d = 10 3 g l o b a l s u r r R B F g o a l = 10 1 , g l o b a l s u r r R B F g o a l = 10 2 R B F m a x n e u r o n s = 20 , R B F a d d n e u r o n s = 5
SAE-ARFS F = 0.8 ( scaling factor ) , C R = 1 ( crossover rate ) Exponential Crossover , φ j x i = d j ( x i ) 3 l e n g h t X t r a i n l b = 200 , n u m s u b p r o b l e m s = 5 max d i m s i z e s u b p r o b l e m = 20 , n u m i t e r s s u b s u b p r o b l e m = 5
SHPSO ω = 0.7298 , c 1 = c 2 = 2.05 , v min = l b , v max = u b l e n g h t X t r a i n l b = 100 , n D i m s < 100 l e n g h t X t r a i n l b = 200 , n D i m s 100 M = 100 , β = 10 2
TL-SSLPSO ω = 1 , c 1 = c 2 = rand , ω = 1 , φ j x i = d j ( x i ) 3
CALSAPSO θ = g 100 , ω = 0.9 θ 2 c 1 = c 2 = 1.49445 , l e n g h t X t r a i n l b = 100
GORS-SSLPSO ω = 1 , c 1 = c 2 = rand , ω = 1 , φ j x i = d j ( x i ) 3
SASMA (SMA) z = 0.03
Table 2. Case Study I: List of unimodal and multimodal test functions.
Table 2. Case Study I: List of unimodal and multimodal test functions.
Function No.Function NamefminProperties
F1Sphere0Unimodal
F2Schwefel 2.220Unimodal
F3Schwefel 1.20Unimodal
F4Schwefel 2.210Unimodal
F5Rosenbrock0Multimodal with narrow valley
F6Step0Unimodal
F7Quartic0Unimodal
F8Schwefel 2.26 418.9829 × n D i m s Multimodal
F9Rastrigin0Very complicated Multimodal
F10Ackley0Multimodal
F11Griewank0Multimodal
F12Penalized (1)0Multimodal
F13Penalized (2)0Multimodal
F14Ellipsoid0Unimodal
F15Shifted Rotated Rastrigin
(F10 in  [67])
−330Very complicated Multimodal
F16Rotated Hybrid Composition
(F16 in [67])
120Very complicated Multimodal
F17Rotated Hybrid Composition
(F19 in [67])
10Very complicated Multimodal
Table 3. Statistical results of the proposed SASMA and the benchmark SAEAs on 30D mathematical test functions.
Table 3. Statistical results of the proposed SASMA and the benchmark SAEAs on 30D mathematical test functions.
Test Function/MetricCALSAPSOGORS-SSLPSOSAE-ARFSSHPSOTL-SSLPSOTLSAPSOSASMA
Mean 2.212 × 10 2 3.180 × 10 5 2.288 × 10 4 1.492 × 10 1 3.880 × 10 2 2.272 × 10 4 1.226 × 10 2
F1St. Dev. 4.883 × 10 2 + 2.293 × 10 5 5.438 × 10 3 + 1.015 × 10 1 + 9.043 × 10 2 + 4.500 × 10 3 + 3.917 × 10 2
Min 1.940 × 10 1 4.090 × 10 6 1.362 × 10 4 3.171 × 10 0 1.678 × 10 0 1.282 × 10 4 5.095 × 10 10
Mean 4.826 × 10 12 6.090 × 10 4 5.821 × 10 5 1.142 × 10 4 5.332 × 10 1 1.610 × 10 9 3.258 × 10 4
F2St. Dev. 9.374 × 10 12 + 1.781 × 10 5 + 2.815 × 10 6 + 4.704 × 10 4 + 1.483 × 10 1 + 6.128 × 10 9 + 4.560 × 10 4
Min 1.762 × 10 9 4.288 × 10 1 7.892 × 10 1 3.304 × 10 1 3.866 × 10 1 7.760 × 10 1 2.777 × 10 6
Mean 1.417 × 10 5 1.686 × 10 4 7.193 × 10 4 4.247 × 10 4 2.689 × 10 4 5.702 × 10 4 2.470 × 10 1
F3St. Dev. 4.632 × 10 4 + 1.230 × 10 4 + 1.124 × 10 4 + 1.190 × 10 4 + 7.087 × 10 3 + 1.303 × 10 4 + 7.522 × 10 1
Min 7.582 × 10 4 5.484 × 10 3 4.358 × 10 4 1.946 × 10 4 1.508 × 10 4 3.447 × 10 4 5.311 × 10 9
Mean 7.628 × 10 1 4.721 × 10 1 7.398 × 10 1 3.264 × 10 1 4.088 × 10 1 6.386 × 10 1 5.160 × 10 2
F4St. Dev. 5.811 × 10 0 + 8.021 × 10 0 + 6.278 × 10 0 + 6.421 × 10 0 + 7.809 × 10 0 + 6.450 × 10 0 + 1.317 × 10 1
Min 6.579 × 10 1 2.850 × 10 1 5.888 × 10 1 2.159 × 10 1 2.957 × 10 1 5.264 × 10 1 1.892 × 10 4
Mean 6.988 × 10 5 1.346 × 10 5 4.185 × 10 7 7.158 × 10 4 1.015 × 10 5 1.182 × 10 8 2.921 × 10 1
F5St. Dev. 8.578 × 10 5 + 2.920 × 10 4 + 1.559 × 10 7 + 7.039 × 10 4 + 2.317 × 10 5 + 4.750 × 10 7 + 8.241 × 10 1
Min 9.871 × 10 4 4.274 × 10 4 1.475 × 10 7 8.585 × 10 3 2.488 × 10 3 3.704 × 10 7 2.895 × 10 1
Mean 1.646 × 10 2 3.536 × 10 5 2.235 × 10 4 1.842 × 10 1 2.271 × 10 2 2.629 × 10 4 6.372 × 100
F6St. Dev. 2.796 × 10 2 + 2.435 × 10 5 4.153 × 10 3 + 1.129 × 10 1 + 4.275 × 10 2 + 5.763 × 10 3 +1.214 × 100
Min 4.659 × 10 0 5.083 × 10 6 1.431 × 10 4 5.132 × 10 0 8.742 × 10 2 1.584 × 10 4 5.842 × 10−4
Mean 8.883 × 10 1 5.761 × 10 1 2.025 × 10 1 5.772 × 10 1 5.688 × 10 1 1.266 × 10 2 9.476 × 10 3
F7St. Dev. 4.258 × 10 1 + 2.317 × 10 1 + 8.888 × 10 0 + 3.047 × 10 1 + 3.945 × 10 1 + 2.628 × 10 1 + 7.626 × 10 3
Min 3.190 × 10 1 2.548 × 10 1 9.262 × 10 0 1.161 × 10 1 1.847 × 10 1 7.957 × 10 1 3.169 × 10 4
Mean 8.448 × 10 3 5.380 × 10 3 6.428 × 10 3 9.051 × 10 3 5.335 × 10 3 7.651 × 10 3 3.094 × 10 3
F8St. Dev. 1.371 × 10 3 + 7.296 × 10 2 + 6.831 × 10 2 + 5.593 × 10 2 + 7.648 × 10 2 + 1.056 × 10 3 +2.298 × 103
Min 5.041 × 10 3 4.288 × 10 3 5.011 × 10 3 7.926 × 10 3 4.029 × 10 3 5.732 × 10 3 1.372 × 10 1
Mean 1.282 × 10 2 1.332 × 10 2 3.115 × 10 2 2.749 × 10 2 1.096 × 10 2 3.729 × 10 2 1.432 × 10 1
F9St. Dev. 4.302 × 10 1 + 3.656 × 10 1 + 3.806 × 10 1 + 3.760 × 10 1 + 2.674 × 10 1 + 2.893 × 10 1 +2.745 × 101
Min 4.873 × 10 1 8.159 × 10 1 2.380 × 10 2 1.757 × 10 2 5.623 × 10 1 3.070 × 10 2 1.930 × 10 7
Mean 2.032 × 10 1 1.434 × 10 1 1.852 × 10 1 8.688 × 10 0 7.867 × 10 0 2.009 × 10 1 3.363 × 10 3
F10St. Dev. 3.964 × 10 1 + 6.077 × 10 0 + 1.265 × 10 0 + 6.990 × 10 1 + 4.311 × 10 0 + 2.177 × 10 1 + 3.216 × 10 3
Min 1.913 × 10 1 3.299 × 10 0 1.468 × 10 1 7.183 × 10 0 2.759 × 10 0 1.912 × 10 1 1.120 × 10 4
Mean 3.796 × 10 0 2.660 × 10 0 1.992 × 10 2 1.116 × 10 0 4.080 × 10 0 2.736 × 10 1 6.739 × 10 4
F11St. Dev. 2.915 × 10 0 + 1.529 × 10 1 + 3.328 × 10 1 + 7.297 × 10 2 + 4.817 × 10 0 + 7.017 × 10 0 + 1.253 × 10 3
Min 1.610 × 10 0 1.113 × 10 2 1.197 × 10 2 1.039 × 10 0 9.886 × 10 1 1.636 × 10 1 4.747 × 10 7
Mean 2.482 × 10 5 1.661 × 10 3 6.228 × 10 7 8.456 × 10 3 1.248 × 10 4 1.921 × 10 8 7.546 × 10 1
F12St. Dev. 7.293 × 10 5 + 2.571 × 10 3 + 4.416 × 10 7 + 2.719 × 10 4 + 4.020 × 10 4 + 1.090 × 10 8 + 3.324 × 10 1
Min 2.943 × 10 1 1.662 × 10 1 9.392 × 10 6 5.588 × 10 0 2.555 × 10 1 4.059 × 10 7 2.928 × 10 6
Mean 8.496 × 10 5 1.283 × 10 5 1.335 × 10 8 4.071 × 10 4 4.430 × 10 4 4.071 × 10 8 2.578 × 10 0
F13St. Dev. 1.239 × 10 6 + 6.569 × 10 4 + 6.399 × 10 7 + 1.580 × 10 5 + 6.827 × 10 4 + 1.703 × 10 8 + 9.339 × 10 1
Min 1.676 × 10 3 1.234 × 10 4 3.547 × 10 7 1.971 × 10 1 6.551 × 10 2 1.005 × 10 8 1.252 × 10 1
Mean 9.632 × 10 3 4.486 × 10 2 2.740 × 10 5 5.951 × 10 3 8.209 × 10 3 3.211 × 10 5 8.987 × 10 2
F14St. Dev. 1.118 × 10 4 + 3.414 × 10 2 + 7.496 × 10 4 + 2.136 × 10 3 + 9.588 × 10 3 + 6.634 × 10 4 + 2.538 × 10 1
Min 2.079 × 10 1 3.969 × 10 1 1.296 × 10 5 2.669 × 10 3 4.844 × 10 2 1.932 × 10 5 1.479 × 10 8
Mean 3.994 × 10 2 2.115 × 10 2 4.647 × 10 2 2.837 × 10 2 1.505 × 10 2 5.226 × 10 2 4.893 × 10 2
F15St. Dev. 5.489 × 10 1 5.557 × 10 1 5.948 × 10 1 = 2.471 × 10 1 3.792 × 10 1 5.321 × 10 1 = 8.253 × 10 1
Min 3.108 × 10 2 1.294 × 10 2 3.792 × 10 2 2.236 × 10 2 7.785 × 10 1 4.219 × 10 2 3.236 × 10 2
Mean 6.799 × 10 2 4.803 × 10 2 5.799 × 10 2 4.114 × 10 2 3.548 × 10 2 6.584 × 10 2 5.975 × 10 2
F16St. Dev. 1.177 × 10 2 + 1.782 × 10 2 8.952 × 10 1 = 8.985 × 10 1 1.428 × 10 2 1.317 × 10 2 + 1.315 × 10 2
Min 4.811 × 10 2 2.330 × 10 2 4.489 × 10 2 3.057 × 10 2 1.360 × 10 2 4.424 × 10 2 3.607 × 10 2
Mean 1.190 × 10 3 1.004 × 10 3 1.115 × 10 3 9.806 × 10 2 9.673 × 10 2 1.110 × 10 3 9.620 × 10 2
F17St. Dev. 1.574 × 10 2 + 6.867 × 10 1 = 5.178 × 10 1 + 2.229 × 10 1 = 3.050 × 10 1 = 5.151 × 10 1 + 8.828 × 10 1
Min 9.785 × 10 2 9.177 × 10 2 1.022 × 10 3 9.464 × 10 2 9.257 × 10 2 1.017 × 10 3 9.000 × 10 2
Win16 12 15 14 14 16
Tie0 1 2 1 1 1
Lose1 4 0 2 2 0
According to the Wilcoxon signed rank test at the 5% significance level, the symbol “+”, “=”, or “−” symbolizes that the performance of SASMA is better, similar, or worse than that of the other SAEAs, respectively. In addition, the best results for all metrics (from the 35 runs) for each test function are highlighted in grey.
Table 4. Mean rank scores (based on the Friedman test) of the SASMA and the benchmark SAEAs on 30D mathematical test functions.
Table 4. Mean rank scores (based on the Friedman test) of the SASMA and the benchmark SAEAs on 30D mathematical test functions.
Friedman TestCALSAPSOGORS-SSLPSOSAE-ARFSSHPSOTL-SSLPSOTLSAPSOSASMA
Mean RankF14.231.466.493.434.346.511.54
F27.003.944.543.462.235.831.00
F36.972.315.744.062.835.091.00
F46.573.716.232.233.115.141.00
F54.833.866.002.832.497.001.00
F64.291.006.293.463.976.712.29
F74.173.406.003.203.237.001.00
F85.802.463.916.572.465.201.60
F93.203.065.895.292.696.831.06
F106.714.064.632.942.466.201.00
F114.662.117.003.174.115.941.00
F124.203.716.112.603.496.891.00
F134.544.116.062.402.946.941.00
F143.772.206.264.034.006.741.00
F154.491.975.342.831.206.405.77
F165.773.264.542.311.865.464.80
F175.943.115.632.972.345.662.34
According to the Friedman test, the best mean rank (of the 35 runs) for each test function is highlighted in grey.
Table 5. SASMA descriptive statistical errors and non-parametric test results versus other state-of-the-art algorithms on the 25C truss design problem.
Table 5. SASMA descriptive statistical errors and non-parametric test results versus other state-of-the-art algorithms on the 25C truss design problem.
AlgorithmMeanSTDMinMean RankSigned Rank
PSO w/inertia 5.534 × 10 2 1.577 × 10 1 5.452 × 10 2 4.80+
PSO w/constr 5.472 × 10 2 4.508 × 10 0 5.453 × 10 2 4.11+
WOA 6.233 × 10 2 3.911 × 10 1 5.710 × 10 2 8.97+
GWO 5.469 × 10 2 9.491 × 10 1 5.455 × 10 2 4.83+
GSA 5.624 × 10 2 1.443 × 10 1 5.458 × 10 2 7.43+
FPA 5.500 × 10 2 3.058 × 10 0 5.457 × 10 2 6.60+
BA 6.883 × 10 2 5.561 × 10 1 5.915 × 10 2 9.97+
GSK 5.452 × 10 2 1.930 × 10 4 5.452 × 10 2 1.00
SMA 5.465 × 10 2 8.644 × 10 1 5.453 × 10 2 4.17+
SASMA 5.459 × 10 2 5.056 × 10 1 5.452 × 10 2 3.11+
According to the Wilcoxon signed rank test at the 5% significance level, the symbol “+”, or “−” indicates that SASMA performs better, or worse than the corresponding metaheuristic, respectively. The best results (Mean, STD, Min) and the best Friedman mean rank are highlighted in grey.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bento, P.; Pombo, J.; Nunes, H.; Calado, M.; Mariano, S. Surrogate-Assisted Slime Mould Algorithm Considering a Dual-Based Merit Criterion for Global Database Management. Algorithms 2026, 19, 265. https://doi.org/10.3390/a19040265

AMA Style

Bento P, Pombo J, Nunes H, Calado M, Mariano S. Surrogate-Assisted Slime Mould Algorithm Considering a Dual-Based Merit Criterion for Global Database Management. Algorithms. 2026; 19(4):265. https://doi.org/10.3390/a19040265

Chicago/Turabian Style

Bento, Pedro, José Pombo, Hugo Nunes, Maria Calado, and Sílvio Mariano. 2026. "Surrogate-Assisted Slime Mould Algorithm Considering a Dual-Based Merit Criterion for Global Database Management" Algorithms 19, no. 4: 265. https://doi.org/10.3390/a19040265

APA Style

Bento, P., Pombo, J., Nunes, H., Calado, M., & Mariano, S. (2026). Surrogate-Assisted Slime Mould Algorithm Considering a Dual-Based Merit Criterion for Global Database Management. Algorithms, 19(4), 265. https://doi.org/10.3390/a19040265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop