1. Introduction
Optimization problems are omnipresent across numerous fields within physics and engineering, and their importance cannot be overstated [
1,
2]. Especially in the fields of applied mechanics and engineering, finding the optimal or near-optimal solution becomes the key to enhancing the efficiency of the system, reducing the consumption of resources, and even determining the success or failure of the technology or product [
3,
4]. The development and application of optimization algorithms as mathematical tools play a crucial role in solving these problems [
5,
6]. With the rapid increase in computing power and the arrival of the significant data era, artificial intelligence technologies, such as deep learning, have developed rapidly in recent years. At the same time, they also bring new challenges and opportunities for the research and application of optimization algorithms. In deep learning, the gradient descent method and its variant algorithms, such as stochastic gradient descent (SGD), Adam, etc., have become mainstream methods in neural network training and optimization [
7,
8]. However, these algorithms often need to be improved to solve complex issues, such as nonconvex optimization and multimodal problems. As a class of stochastic optimization methods based on intuitive or empirical constructions, the Metaheuristic Algorithm (MHA) has shown significant advantages in solving practical optimization designs in a wide range of domains due to its lack of a strict gradient requirement, as well as its parallel search, fewer parameters, and strong search capability [
9,
10].
MHAs are a class of optimization tools designed for simulating natural phenomena or processes [
11]. They can accurately seek optimal or near-optimal solutions in complex search spaces, an ability that is made possible by the simulation of complex behaviors, such as biological evolution, physical motions, and chemical reactions. These algorithms are capable of global search and are flexible in adapting their strategies to diverse optimization problems [
12,
13]. Depending on the simulation objects and mechanisms, MHAs can be categorized into four primary groups: evolution-based, physics-based, population-based, and human-based [
14]. Inspired by the theory of biological evolution, evolution-based methods explore and optimize problem solution space by simulating biological evolutionary mechanisms, such as natural selection, heredity, and mutation. Among them, genetic algorithms (GA) [
15] are the pioneers in this field, and they mimic the process of biological inheritance and mutation. In addition, differential evolutionary algorithms (DEs) [
16], and Evolution Search (ES) [
17] are well-known evolutionary algorithms. Physical phenomena and laws inspire physics-based MHAs. The Simulated Annealing Algorithm (SA) simulates the physical process of heating an object to melt and then its gradually cooling to optimize the problem solution [
18]. Similarly, the Fick’s Law Algorithm (FLA) is based on Fick’s first law of diffusion [
19], while the Gravitational Search Algorithm (GSA) simulates the law of gravity in the physical world [
20]. On the other hand, population-based metaheuristic algorithms simulate the behavior or group intelligence of populations in nature [
21]. They represent the solution of a problem as individuals in a population and find the optimal solution through interaction, competition, and cooperation among individuals. The Particle Swarm Algorithm (PSO) [
22] is the earliest traditional algorithm, and it simulates the social behavior of animal groups. Other well-known algorithms include the Gray Wolf Optimization (GWO) [
23], the Harris Hawk Optimization (HHO) [
24], and the Black-Winged Kite Optimization Algorithm (BKA) [
25]. Finally, human-based methods are mainly generated based on human behaviors, mindsets, and cognitive processes. The Political Optimization Algorithm (PO) [
26], Human Evolutionary Optimization Algorithm (HEOA) [
27], and Rich-Poor Optimization Algorithm (PRO) [
28] are outstanding representatives in this field. These algorithms provide new ideas and methods for solving complex optimization problems by simulating human behaviors or cognitive processes.
YDSE is a method based on physical principles proposed in 2022 [
29]. It has effectively addressed many realistic engineering problems due to its advantages of high flexibility, high convergence accuracy, and global search capability. Compared with some classical and new algorithms, YDSE is highly competitive and asymptotic when facing engineering optimization problems. However, YDSE cannot effectively balance the equilibrium between global and local searches, as well as having slower convergence to the optimal solution, and tends to trap in the local best, thus limiting its application scope. Some improved versions of YDSE also try to solve these problems and verify the effectiveness of the improvement through engineering optimization problems. For example, Hu et al. presented an enhanced form of YDSE. They tried to introduce four efficient mechanisms to enhance the optimization ability of YDSE and demonstrated better optimization performance in some difficult engineering optimizations [
30]. A smart weighting tool for optimizing dissolved oxygen levels based on Young’s double-slit optimizer was presented by Dong et al. [
31].
Although YDSE and some existing enhanced versions still demonstrate better convergence performance and results when facing engineering optimization problems, they still do not effectively balance the global and local searches and are especially prone to becoming bogged down in localized solutions when facing complex issues. This result leads to unsatisfactory optimization results obtained by the YDSE optimizer when solving difficult optimization issues. Therefore, this study proposes a fractional-order boosted hybrid YDSE optimizer to further alleviate the shortcomings of existing YDSE algorithms and attempts to solve realistic engineering optimization problems of higher complexity. In addition to the above motivations, bringing in operators and single strategies, while improving the manifestations of one aspect of the algorithm, can indirectly impair the performance of other aspects. Therefore, we jointly improve the YDSE problem by introducing multi-strategy mechanisms and minimizing this indirect damage. Second, many original and improved algorithms have been proposed that are performance-efficient and effective. However, based on the No Free Lunch (NFL) statement [
32], it has been shown that existing methods are not capable of comprehensively addressing all conceivable optimization challenges. This scenario suggests that a certain category of optimization algorithms, while capable of yielding satisfactory outcomes for particular problems, may result in unacceptable results for others [
33]. In addition, the fractional-order strategy is considered a strategy to efficiently address the balance between exploration and exploitation. It has been applied to improve a variety of algorithms and has demonstrated efficient experimental performance. For example, a JS algorithm based on fractional-order improvement has been proposed to address the inability of the JS algorithm to effectively balance between exploration and exploitation, and the effectiveness of the strategy has been verified through a variety of engineering optimization problems and real-world economic forecasting problems [
34]. These reasons motivate this study to propose a fractional-order boosted hybrid YDSE optimizer. Solteiro Pires et al. proposed a convergence rate control method for Particle Swarm Optimization algorithms based on fractional-order calculus [
35]. Luo et al. proposed a Jellyfish Search Algorithm (JSA) for the optimization and tuning of the control gains of the developed strategy to obtain a high-quality global optimum [
36]. Zhang et al. proposed an improved algorithm of the Bird Flock Algorithm for solving parameter estimation of fractional-order chaotic control and synchronization [
37]. These reasons motivate this study to propose a fractional-order boosted hybrid YDSE optimizer.
Therefore, this study proposes a fractional-order boosted hybrid YDSE optimizer. First, a fractional-order strategy is introduced into the dark edge position updating of FYDSE to ensure more effective exploitation of the search potential of a single neighborhood space while reducing the possibility of getting bogged down in localized solutions. Secondly, piecewise chaotic mapping is constructed in the generation of the initial solution to gain better distributed initial individuals and to enhance the speed of convergence to the most individuals. Moreover, the low exploration space of the algorithm is extended by using a dynamic inversion strategy, which improves the probability of obtaining a globally optimal solution. Finally, by introducing the vertical operator, FYDSE can better balance global exploration and local exploitation to explore new unknown areas. In addition, the presented FYDSE is compared with a series of the most efficient comparative algorithms in a comparative experiment on the CEC2022 test. The experimental outcomes and statistical analysis validate the efficient performance of FYDSE in addressing optimization issues. Further, to verify the efficient ability of the FYDSE in facing real-world engineering application problems, comparative experiments of FYDSE are conducted on three complex truss topology optimization problems. The superior adaptability and high efficiency of the FYDSE in complex engineering issues validate the ability of the FYDSE in dealing with similar applied mechanics and engineering problems.
The key contributions of this research are enumerated as follows:
- (1)
In this article, a fractional-order boosted hybrid Young’s double-slit experimental optimization algorithm is proposed.
- (2)
A multi-strategy mechanism is employed to jointly improve the YDSE problem to mitigate the negative impact of a single strategy on other aspects of the algorithm and to improve its ability to solve complex problems.
- (3)
Experimental results of numerical and 20-bar, 24-bar, and 72-bar topology optimization validate the performance advantages of the proposed FYDSE approach.
The remainder of the article is arranged as described below.
Section 2 describes the specific details of the YDSE optimizer.
Section 3 describes a fractional-order boosted hybrid improved YDSE optimization method. In
Section 4, the ability of FYDSE on the CEC2022 test suit is discussed.
Section 5 describes the performance of YDSE on the 20-bar, 24-bar, and 72-bar topology problems. Finally, the conclusions of this research are introduced.
2. The Theory of YDSE
YDSE is an innovative physics-based meta-heuristic method presented in 2022 that provides high-quality solutions to numerical and engineering optimization problems based on the fluctuating properties of light [
29]. YDSE has already demonstrated excellent performance in the original and its improved versions, effectively addressing real-world mechanical optimization issues. The workflow of YDSE consists of three core steps: The first part is to initialize the position of the population. In the second part, YDSE simulates the fluctuating behavior of light according to Huygens’ principle and guides the algorithm in finding the optimal solution by constantly updating the traveling waves and path differences. The third part is the exploration and development phase. In the exploration stage, YDSE attempts to discover new potential optimal solutions by expanding the search range. Meanwhile, in the exploitation phase, the current optimal solution is fine-tuned to obtain a more highly interpretable solution. The ability to effectively balance these two processes dramatically determines the performance of the algorithm [
29].
YDSE initializes each individual by simulating the process of projecting a monochromatic light wave source through two closely spaced slits into the barrier. To simulate this process, an original light source composed of
M waves is constructed in the solution space as follows [
29]:
where
Mi,j represents the
ith wave in dimension
j. In addition,
n means the population size and
d means the dimension of the issue. For each
j dimension,
ubj and
lbj denote the maximum and minimum range of the issue. In addition,
rand means a randomly generated value between 0 and 1.
After initialization, waves next pass through the slit barrier and diverge in many different directions. The resulting dispersion allows for constant changes in the position of the wave source and the wave center. The YDSE method does this by setting the points on the wavefronts passing through the two slits to be equal and set to
n. In a simplified way, the wavefronts of
n points at two slits can be calculated by the following two equations [
29]:
where
FSij and
SSij denote the
jth dimension coordinates of the
ith point on the wavefront flowing out of the first and second slits, respectively. To simulate the random scattering phenomenon that may occur after a light wave passes through the slit, we introduce two random variables,
rand1 and
rand2, both of which take values in the range of [−1, 1].
Length means the length from the light source to the barrier.
SAvg means the mean value of the current light wave group, which is calculated as follows [
29]:
where
Mi,: means the
ith original light source.
Points on the wavesurface propagating from the two narrow gaps produce interference patterns. Specifically, constructive and damaging interference generate light and dark streaks. Therefore, the position of each individual is updated by constructed wave fronts (
FS and
SS) to realize the interference behavior and path differences between light and dark streaks. The specific equations are as follows [
29]:
where ∆
P means the path length from the
FS point to the
SS point. In addition,
FSi,: and
SSi,: mean the
ith
FS point to
SS point. The introduction of ∆
P is utilized to distinguish different interference orders. Specifically, for zero-order or even-order interference, bright fringes are produced in the interference region. Meanwhile, for odd-order interference, dark fringes are produced. The specific equations are given below [
29]:
where
m means the order of interference. In addition,
λ denotes wavelength.
The bright fringe emerges due to the combination of two waves, creating a wave of greater amplitude. The amplitude of this bright fringe is then progressively updated as follows [
29]:
where
Ab means the mean amplitude of bright fringes. In addition, cosh(·) means the cosh function.
t represents the now iteration and
Tmax means the maximum iterations.
If the peak of one wave meets the trough of the other when two waves are added together, they both cancel out and produce dark fringes. Similarly, the amplitude of the dark fringe is iteratively updated as follows [
29]:
where
φ represents a constant with a value of 0.38. In addition, tanh
−1(·) means the inverse of the tanh function.
Both bright and dark fringes represent potential candidate solutions when exploring solutions. In particular, the center fringe location is considered to be the optimal solution. Due to the effect of destructive interference, the fitness values of the dark fringe candidate solutions are lower than those of the bright fringe candidate solutions. This interference causes the positions in the dark fringe to be inferior to objective function than those in the light fringe. Therefore, YDSE emphasizes the priority of exploring for hopeful solutions in the dark zone during the exploration phase. A dark edge update is introduced to guide the search process and ensure that the solution space can be explored more efficiently. The specific update equation is as follows [
29]:
where
XN-odd means the dark fringe, and
rand denotes a randomly selected number between 0 and 1.
LiN-odd denotes the dark edge intensity, which is utilized to reflect the brightness change of the dark fringe and
Xbest denotes the best solution. In addition,
Z means a trial vector of
d, as follows [
29]:
where
h means a randomly created number in the range of [−1, 1]. In addition, the definition of
LiN-odd is as follows [
29]:
where
yd is employed to measure the length from the center stripe to the dark stripe, as follows [
29]:
In addition,
Limax denotes the maximum center stripe intensity, as follows [
29]:
where
C is taken as 10
−20, which represents the maximum intensity.
When the order is even, YDSE pays special attention to the bright edge regions and finds promising candidate solutions in them. It assumes that these regions are where the best solutions are potentially located, which, in turn, generates constructive interference in favor of the solution process. Therefore, during the development phase, the YDSE optimizer spares no effort in exploring all possible and promising candidate solution regions in the bright edge region. Its position update equation is concretely given as below [
29]:
where X
N-even means the bright fringe. In addition,
Y denotes the difference between two randomly acquired stripes and
LiN-even denotes the intensity of the bright stripe.
g means a randomly created number in the range of [0, 1]. The definition of
Y is as follows [
29]:
where X
N−r1 and X
N−r2 represent two random fringes. In addition, the definition of
LiN-even is as follows [
29]:
where
Limax is defined through Equation (13).
Finally, the specific formula for the position update method of the center fringe is as follows [
29]:
where
XN-zero means the central fringe and
rand means a random number between 0 and 1. In addition, X
rb represents the bright fringe selected based on a random even number of
rb.
3. Proposed FYDSE Algorithm
The YDSE method suffers from slow convergence, is easily caught up in localized solutions, and is unable to effectively balance exploration and exploitation when facing complex optimization problems in engineering. In this work, a fractional-order improved optimization method for a multi-strategy YDSE is proposed. The specific modifications of the proposed FYDSE are shown below.
3.1. Fractional-Order Modified Mechanism
Fractional calculus elevates the fundamental order of calculus by extending its realm from integers to fractions. It systematically derives the solution limit by employing differential approximations of integer calculus, specifically through the differentiation and integration of fractional orders [
34,
35]. The introduction of the fractional-order strategy in the population position updating process of FYDSE, which extends
X(
t) to fractions through the fractional-order formula, guarantees a more efficient exploration of the potential within a single neighborhood space, thereby minimizing the likelihood of falling into a local optimum. The definition of fractional-order discretization based on the Grunwald–Letnikov (G-L) method [
38] is summarized as follows [
34]:
where
β is the fractional element in the common signal
ω(
t), Γ denotes the gamma function, and
T is the truncation term.
r stands for the order.
The results of the fractional derivative are closely related to the value of the current term and the previous state, with the effect of past events diminishing over time. Then, the above equation is modified as follows [
34]:
where
β is set to 4 [
34] in the specific implementation.
During the exploration phase, given YDSE’s emphasis on identifying promising solutions in the dark region, a fractional-order correction strategy was implemented to enhance the updating formula for dark streaks, thus refining our search approach.
Figure 1 shows the effect of fractional-order correction on the update. The dark fringe update formula based on fractional-order correction is as follows:
Notably, upon multiplying the terms of 1/120, 1/720, or greater by the remaining components of the equation, their respective results become insignificant, scarcely influencing location updates. Consequently, these higher-order terms are disregarded.
3.2. Piecewise Chaotic Map Strategy
YDSE guides the updating of candidate solutions by iterative a priori information. Therefore, a better initial value captures valid information faster and thus searches for the optimal solution location faster [
39]. The initialization of the established methods is through Gaussian distribution in order to generate randomly. However, this initialization is highly contingent, and there are more invalid solutions, thus limiting the effective updating of the solution. To obtain a better distribution of the initial solution space, FYDSE introduces a piecewise chaotic mapping distribution to initialize the solution space [
40]. The population initialization in the piecewise chaotic mapping is achieved through chaotic variables rather than relying on random variables, thereby facilitating a full and efficient exploration of the solutions space [
41]. The updated formula for the piecewise chaotic mapping is as follows [
41]:
where
d means a random value in the scope of (0, 0.5).
means the piecewise chaotic mapping function. Therefore, the initialization method is updated as follows:
3.3. Dynamic Opposition (DO)
When solving complex optimization problems, YDSE often fails to efficiently traverse the search region of low exploration, which results in the algorithm failing to capture the global optimal solution. Therefore, a dynamic opposition (DO) strategy is introduced in the YDSE optimizer [
42]. It extends the search space of low exploration by using a dynamic inversion strategy, thereby enhancing the likelihood of getting close to the global solution.
Considering that frequent DO strategy updates can affect the development performance of the algorithm, a jump rate (
Jr) determines DO policy in the current iteration or not. If the number created at random is less than
Jr, the DO implementation strategy is applied to all individuals. Consequently,
Jr governs the likelihood of executing DO [
43]. This approach, the Randomized Jump Strategy (
RJS), facilitates the balanced exploration and development of methods in an effective manner through iterative selection. Furthermore, implementing DO during the selection iteration leads to an enhanced utilization of the search space, thus maximizing its efficiency. The DO strategy first requires generating the opposite solution in the neighborhood of the candidate solution
X as below [
42]:
where
represents the inverse solution.
The position of the candidate solution undergoes a shift due to the influence of a randomly generated number within the range of [0, 1], thereby yielding a new position designated as
Xrand [
42]:
where
rand means a random number.
When the candidate solution
X is different from the moving direction of
Xrand guided by the random number, a new candidate position based on the DO policy is constructed, as follows [
43]:
where
represents the new position after dynamic opposition update.
To further illustrate the specific implementation of the DO strategy, Algorithm 1 provides pseudo-code for the DO strategy.
Algorithm 1: Dynamic Opposition Strategy |
1: input: Jumping rate Jr, solution size N, candidate solution location X. |
2: if rand < Jr |
3: for i = 1 to N do |
4: |
5: |
6: |
7: end for |
8: end if |
9: output: New location on dynamic dichotomy. |
3.4. Vertical Crossover Operator
To address the tendency of the YDSE algorithm to converge prematurely into local optimal solutions while tackling intricate problems, we develop an update strategy based on the vertical operation mechanism [
44,
45]. Specifically, the vertical operator performs an arithmetic crossover operation between two dimensions of the candidate solutions to generate new candidate individuals. By integrating the vertical operator, the FYDSE method gains enhanced global exploration and local exploitation abilities, allowing it to delve into novel and unexplored domains [
46]. Furthermore, the adoption of the updated foraging strategy mitigates the issue of stagnant updates across iterations over many times. Assuming that the vertical crossover operation is performed with the
d1 and
d2 dimensions of the
ith candidate solution, the generated conditioning solution
is as follows [
45]:
where
c means uniformly distributed random values ranging from 0 to 1,
n means the quantity of agents, and
d signifies the dimensionality of the population. In addition, X
id1 and X
id2 denote the
d1st and
d2nd dimensions of the
ith position.
M(1,
n) represents a function that selects individuals from all populations.
The vertical operator strategy is normalized for individuals, considering the upper and lower bounds for decision variables. The vertical operation of each is tailored specifically to a single individual, ensuring a focused search. This single-individual approach prevents the disruption of potentially globally optimal dimensions by abruptly departing from a locally optimal yet stagnant dimension. Furthermore, introducing a competition operator fosters healthy competition between the new and original candidate solutions. Specifically, the competition operator works as follows [
46]:
where
Vvc represents the candidate solution after the vertical operator operation.
Xnew means new position.
f(·) represents the fitness function.
3.5. The Proposed Enhanced FYDSE Meticulous Steps
Given the vast search space encountered in engineering optimization problems, attaining global convergence for intricate optimization challenges poses a significant challenge. To address this, this paper introduces a fractional-order improved boosted hybrid YDSE, explicitly tailored for solving complex optimization issues. First, a fractional-order strategy is introduced into the dark fringe position updating of FYDSE to ensure a more elegant search potential in a single neighborhood space, and also to minimize the risk of drifting into local optimality. Second, during the initialization phase, a piecewise chaotic mapping is introduced to yield a superior-quality initial population, thereby enhancing the convergence speed of the algorithm. Moreover, the lower exploration space for algorithms is extended by using a dynamic inversion strategy, which improves the probability of obtaining a globally optimal solution. Finally, by introducing the vertical operator, FYDSE can enhance equilibrium global exploration and local exploitation and explore new unknown areas.
The primary processes of FYDSE are described below:
Step 1: Initialization of the relevant elements of the FYDSE: the stripe quantity n, the dimension d, the lower and upper limits of the variable lb and ub, the maximum iteration number Tmax, the fraction coefficient β and the jump rate Jr;
Step 2: Initialize wavefronts by piecewise chaotic mapping. Obtain the position of each fringe based on the wavefronts and path difference ∆L and compute the objective function value.
Step 3: Update Limax by utilizing Equation (13) and pick the optimal location Xbest.
Step 4: When m is 0, the central stripe amplitude is updated using Equation (7) and the central stripe position is also updated according to the amplitude and the best position Xbest by Equation (17).
Step 5: When m is odd, the dark streak intensity and amplitude are updated using Equations (7) and (11). Afterwards, a new location is obtained by the dark streak update formula based on the fractional-order correction.
Step 6: When m is even, Equations (8) and (16) are utilized to update the amplitude and intensity of the bright streaks and further update the location of the bright streaks.
Step 7: Examine the lower and upper bounds of the variables and evaluate the objective function of the updated solution.
Step 8: If rand < Jr, update the dynamic opposite solution for all the stripes by Equations (23)–(25) and compute the fitness value.
Step 9: A vertical operator is utilized to intersect the two dimensions of a randomly selected location and the availability of that new candidate solution is determined by a competitive operator.
Step 10: Consider the limitations of the assessment variables, both the upper and lower boundaries, and evaluate the solution of the objective function accordingly. Output the optimal fringe position Xbest.
The figure depicted in
Figure 2 outlines the proposed FYDSE flowchart; furthermore, Algorithm 2 provides the corresponding pseudo-code for a clearer understanding of its implementation.
Algorithm 2: Fractional-order boosted hybrid Young’s double-slit experiment optimizer |
1: Input: Initialize the related parameters of the proposed FYDSE algorithm: number of fringes (n), upper and lower bounds (lb and ub), dimension (d), maximum number of iterations (Tmax), the fraction coefficient β and jump rate Jr; |
2: Initialize a monochromatic light source consisting of n waves using Equation (22) by piecewise chaotic mapping. |
3: Calculate the wavefronts (FS and SS) of the two slits. |
4: Obtain the position of each fringe based on wavefronts and path difference ∆L. |
5: Examine the upper and lower bounds of the variables and evaluate the objective function of the updated solution. |
6: While t < TMax do |
7: Update Limax by utilizing Equation (13) and pick the optimal location Xbest |
8: For i =1 to n do |
9: Update Z with Equation (10). |
10: If m = 0 (Central fringe) |
11: The intensity and amplitude of the central fringe are updated with Equations (17) and (7). |
12: Update the position of the center stripe Xm-zero based on the amplitude and the optimal position Xbest. |
13: Else if m = even (bright fringe) |
14: The intensity and amplitude of the bright fringe are updated with Equations (14) and (7). |
15: Update the bright fringe Xm-even by Equation (14). |
16: Else if m = odd (dark fringe) |
17: The intensity and amplitude of the dark fringe are updated with Equations (9) and (8). |
18: Update the new position Xm-odd of dark fringe based on the fractional-order modification with Equation (20). |
19: End if |
20: Check the bounds of the variables. |
21: End for |
22: If rand < Jr |
23: For i = 1 to n do |
24: |
25: |
26: |
27: End for |
28: End if |
29: B = permutate(d). |
30: For i = 1 to d/2 do |
31: Update a uniformly random value p. |
32: If p< P then let no1 = B(2i−1), and no2 = B(2i). |
33: For j = 1 to n do |
34: Construct a random value c in the range of [0, 1]. |
35: Update the position of a stripe according to the vertical operator. |
36: End for |
37: End if |
38: End for |
31: Update the best position Xbest. |
32: t = t + 1. |
33: End while |
34: Output: the best solution Xbest. |
3.6. The Computational Complexity of FYDSE
An important metric for evaluating optimization algorithms is the time complexity. The original YDSE has a complexity of
O(
Tmax×
M×
n+
Tmax×
n×
d), where
M represents the number of evaluations of the objective function. The proposed FYDSE includes four effective strategies. First, the segmented mapping initialization has the same complexity of
O(
n×
d) as the Gaussian-based initialization of YDSE. Second, the fractional-order improved dark streak updating strategy has no added time complexity. Since the fractional-order improved strategy needs to save the experimental results of the first
β iterations, the space complexity increases by
O(
β×
n×
d). Second, the DO strategy produces dynamic inverse solutions for all candidate solutions. However, given the limitation imposed by
rand <
Jr, the computational complexity of the DO strategy in the crap scenario of fulfilling the equation stands at
O(
Tmax×
n×
d). Finally, the complexity of the vertical operator is
O(
Tmax×
n/2×
d). Thus, the overall complexity of FYDSE is:
4. Numerical Simulations
To comprehensively measure the capabilities of FYDSE, by contrast to some other SOTA methods, in this section a comparative experiment on the CEC2022 set of tests is described. All the methods are set with the same parameters in the experiments, i.e., the maximum iteration number is 1000, and the population size is 30. We chose the mean, worst solution, optimal solution, and standard deviation as the evaluation indexes to visualize the experimental results. The experimental results are presented and analyzed in various forms, such as iterative plots, box plots, and radar plots. All tests were conducted on a personal computer with Matlab-2019b with a 2.11 GHz quad-core Intel(R) Core(TM) i5 and 8.00 GB.
4.1. Experimental Parameter Setting and Benchmark Functions
To deeply verify the effectiveness and superiority of our proposed FYDSE algorithm, we identified and selected a set of advanced and distinctive optimization algorithms for a comprehensive analysis and comparison with the experimental findings of FYDSE as follows: (1) YDSE [
29]; (2) Artificial Hummingbird Algorithm (AHA) [
47]; (3) Snake Optimization (SO) [
48]; (4) Dandelion Optimizer (DO) [
49]; (5) Aquila Optimizer (AO) [
50]; (6) Chimpanzee Optimization Algorithm (ChOA) [
51]; (7) Honey Badger Algorithm (HBA) [
52]; (8) Sand Cat Swarm Optimization (SCSO) [
53]; (9) Salp Swarm Algorithm (SSA) [
54]; (10) White Shark Optimizer (WSO) [
55]; (11) COOT Algorithm (COOT) [
56]. By comparing with these algorithms, we aim to comprehensively demonstrate the advantages and features of the FYDSE algorithm in terms of performance, efficiency, and adaptability, thus further proving its capability to solve sophisticated optimization issues.
Table 1 provides the parameters for the various algorithms.
The CEC2022 test suite contains 12 well-designed single-objective test functions used as benchmark functions for this experiment. These functions cover a wide range of types, such as single-peak, basis, hybrid, and combined functions, which are distinctive in complexity and morphology. Some functions exhibit smooth surface features, while others are full of sharp peaks and steep valleys. Among them, for the cec01 function, as a representation of a single-peaked function, only one globally optimal solution can measure the convergence speed and accuracy within the algorithm. On the other hand, the cec02-cec05 functions represent multimodal functions that possess numerous local optima, among which a single global optimum is designed to test whether the algorithm can successfully avoid traps of local minima and thus find the globally optimal solution. In addition, the hybrid functions cec06-cec08 are used to model the properties of real-world complex problems and allow for the evaluation of the algorithm’s effectiveness and ability to solve problems with hybrid properties. The CEC2022 test suite also introduces several combinatorial functions, such as the cec09-cec12 functions, which incorporate characteristics from a diverse set of optimization problems, presenting more intricate choices for the algorithms to overcome. In addition, the suite contains several constrained optimization problems. These problems search for optimal solutions under specific constraints and require the algorithms to adhere to these constraints strictly during optimization. The value domains of all test functions are set between [−100, 100].
4.2. Analysis of Exploration and Exploitation Behaviors
In the conventional YDSE framework, the refinement of dark fringes emphasized probing unknown areas to foster diversity. In contrast, the center and bright fringe prioritize cultivating optimal solutions within promising intervals. On the other hand, FYDSE improves the dark streak updating process through a fractional-order correction strategy, so that the positional updating of dark fringes focuses more on the search potential of a single neighborhood space, which mitigates the risk of converging to a local optimum and aids in discovering an optimal solution. Meanwhile, performing dynamic opposition for all fringe positions facilitates the exploration of the inverse positions of candidate solutions in the domain of the search, increasing the possibilities for searching undeveloped areas. Furthermore, incorporating the vertical operator enhances the exploration of potentially rewarding regions by augmenting the exploration of stripe positions within the dimensional intersection.
To ascertain the efficacy of the implemented strategy in balancing the exploration-exploitation equilibrium, we delve into the group diversity of FYDSE in the context of CEC2022. Population diversity infers group traits within a condensed environment through variations among individual dimensions [
57]. Specifically, an increase in the difference between dimensions of the stripe location determines the dispersal or distribution in the search area of the population. Conversely, individuals will cluster into a convergence region. The introduction of the vertical operator improves the difference between the dimensions of the stripe positions.
First, we define the diversity (
Divj) of the
jth dimension and further compute the average diversity (
Div) of the dimensions. Further, the proportion of exploration and exploitation is calculated based on the average diversity. The specific formula is as follows [
57]:
where
n represents the number of all light and dark stripes and
d represents the dimension. The median of the
jth dimension across all (
n) stripes is denoted as
Median(
X:,j).
Meanwhile, the definitions of the exploration rate
Explora(%) and the exploitation rate
Exploita(%) are given [
57]:
where max(·) denotes the max function.
This experiment validates the diversity results of FYDSE at the CEC2022 test function with 1000 iterations of FYDSE.
Figure 3 shows the iterative plots of the exploration and exploitation rates of FYDSE at CEC2022 for all 12 test functions. From the analysis of the graphs, it can be concluded that for cec02, cec05, and cec07, FYDSE exhibits a high rate of exploration during extended iterations, while maintaining a high utilization rate during the middle and later stages of iterations. This result is attributed to the fact that the piecewise mapping initialization and dynamic opposition strategies enhance the search potential of YDSE. Meanwhile, the fractional-order modification strategy enhances the exploration performance by improving the update of dark streaks and effectively utilizing the existing prior knowledge. When dealing with cec01, cec03, cec09, and cec11, most iterations tend to be in a more intensive exploitation phase, with only a brief exploration phase in the initial stages. This is primarily due to the vertical operator enhancement of the exploitation stage. Additionally, the lower exploration rate during shorter iteration times indicates that the modified strategy aids the algorithm in swiftly locating the best solution.
4.3. Results Analysis with the Latest Methods in CEC2022
Table 2 provides the experimental outcomes of the FYDSE with the other latest comparison algorithms on the CEC2022 test dataset. To comprehensively evaluate the performance of these algorithms, we have chosen five key evaluation metrics, including the mean, the best value, the worst value, the standard deviation, and the ranking, to consider these algorithms comprehensively. These metrics provide us with a multi-dimensional view of the algorithm performance, which helps us to understand more clearly the performance differences and advantages and disadvantages between the FYDSE algorithm and other algorithms. As is evident from
Table 2, the FYDSE algorithm exhibits the superior average performance among the eight test functions (cec02, cec06, cec07, cec08, cec09, cec10, cec11, cec12) under consideration, significantly outperforming the other algorithms. Not only that, but FYDSE also demonstrates remarkable competitiveness in other test functions. Comparatively, the original algorithm YDSE provides the best average results only on the specific test function cec05. The HBA algorithm, on the other hand, achieves the best average performance on two test functions, cec01 and cec03. In addition, the WSO algorithm gives the best average results for the cec04 test function. And for the cec01 function, both the HBA and SSA algorithms find the optimal solution. When exploring the cec01 single-peak function, both the HBA and SSA algorithms can converge to the global optimal solution. Meanwhile, our proposed FYDSE algorithm can also reach the sub-optimal solution level, further verifying that the FYDSE algorithm is not only fast-converging but also highly accurate. When confronted with these multimodal functions from cec02 to cec05, FYDSE achieves the best mean result in the cec02 function. Although the best mean value is not obtained on several other functions, the test results of FYDSE show the most minor standard deviation, demonstrating the stable exploration and development capability of FYDSE when facing multimodal functions. It is especially worth mentioning that the FYDSE algorithm can also find the global optimal solution when solving combinatorial and hybrid function problems. This achievement highlights the global solid search capability of FYDSE when solving complex, multivariate optimization problems and demonstrates its capability to achieve a balance in exploration and exploitation. In particular, when handling higher-dimensional, multimodal optimization tasks, the outstanding performance of FYDSE allows it to address the intricacies of real-world optimization problems with more excellent proficiency.
Table 3 provides a comprehensive overview of the statistical outcomes obtained from the Wilcoxon rank sum test conducted on FYDSE and alternative algorithms for comparison, with a predefined significance level of 0.05. Here, we describe the performance difference between the algorithms based on the size of the
p-value. Specifically, when the
p-value is less than 0.05, the difference between FYDSE and the comparison algorithm is significant. In contrast, when the
p-value is greater than or equal to 0.05, it implies that the difference in performance between the two is insignificant. In addition, the symbols “-”, “=”, and “+” in
Table 3 provide us with intuitive performance comparison information. Among them, “-” indicates that other meta-heuristic algorithms are not as practical as FYDSE in the corresponding test function. “=” indicates that FYDSE and the comparison algorithm have comparable performance in this test function, and both of them have the same effect. “+” indicates that the other algorithms obtained better results than FYDSE.
Upon examining
Table 3, we can discern the distribution of the outcomes for all algorithms in the Wilcoxon test. Specifically, the results of the AHA, SO, HBA, and SSA algorithms are each 0/1/11. The statistical results imply that FYDSE does not have statistically worse results than AHA and SO in this set of tests and demonstrates better yield outcomes in a majority of the test functions. On the other hand, the results of the DO, SCSO, AO, and ChOA algorithms are 0/0/12. The statistical results imply that these algorithms neither show better performance than FYDSE nor comparable performance to FYDSE in the tests, but instead show worse performance 12 times. In addition, the Wilcoxon test results for both WSO and COOT are 0/2/10, indicating that they outperform the proposed FYDSE algorithm for no test function in the statistical test, and that their performance is comparable for the two test functions. Finally, from the statistical results of the original YDSE, it can be found that only one test function exists for YDSE that is statistically superior to FYDSE. However, in the other test function, FYDSE obtains superior experimental results. These detailed test results illustrate the performance differences between the algorithms.
Figure 4 showcases the convergence performance of the FYDSE in comparison to other approaches when evaluated on the cec2022 test function set. The x-axis represents the number of iterations, whereas the y-axis represents the fitness values, where some function results are expressed in logarithmic form with a base of ten. All algorithms start from the same initial point (i.e., iteration zero) to ensure a fair comparison. Observing
Figure 4, for the cec01-cec03 functions, FYDSE exhibits a high convergence rate during the initial iteration phase, followed by a gradual localization of the optimal position and validation of the previous findings by updating the position. For the cec04-cec06 functions, FYDSE subsequently obtains the best solution at a satisfactory speed and executes a precise search optimization in proximity to the optimal solution. This performance proves its reliability in avoiding local optima. For the cec07-cec08 function, FYDSE can steer the search over a large area during the initial iteration phase to find potentially high-quality regions in the search space. FYDSE shifts to a localized search as the algorithm proceeds, updating to the optimal location over a smaller area. For the cec09-cec12 functions, FYDSE can swiftly transition between the initial search phase and later stages of development, converging to a solution that is close to optimal at an early iteration. Afterward, it continues pinpointing the optimal position and validating previous observations by updating the results. In summary, FYDSE performs well in all four types of test functions and maintains a significant advantage in most of them. Moreover, these compelling results also demonstrate that the FYDSE algorithm effectively strikes a harmonious balance between exploratory and developmental search strategies.
Figure 5 shows the distribution of the best mean values of FYDSE and other comparison algorithms in multiple test functions through box-and-line diagrams. From the figure, we can see that in most test scenarios, the distribution of the best mean values of FYDSE is more concentrated and compact, fully demonstrating the excellent performance of FYDSE. In addition, this distribution characteristic further proves that FYDSE has excellent consistency and stability under different testing conditions, making it stand out amongst the algorithms. Specifically, when facing the test functions of cec03, cec04, cec07, cec08, and cec12, the box-and-line plot box of FYDSE is narrower, which implies that the performance of FYDSE fluctuates less during multiple iterations, demonstrating its stability and reliability. As for the test functions cec01, cec02, cec05, cec09, cec10, and cec11, the box-and-line diagram of FYDSE shows a red line, which means that FYDSE can solve the problem effectively and achieve a high-performance level during each iteration. This performance once again proves the superiority and usefulness of FYDSE.
Figure 6 visually depicts the capability rankings of the FYDSE against a string of comparative algorithms across the twelve functions of the CEC2022 benchmark test set, utilizing an intuitive radar chart format. Upon examining
Figure 6, it becomes evident that the FYDSE exhibits a significantly smaller coverage area within the radar chart, which not only accentuates its outstanding performance across various optimization scenarios but also attests to its exceptional optimization capabilities and stability. Regardless of whether it tackles simple or complex optimization problems, FYDSE can swiftly discover satisfactory solutions through efficient and accurate search strategies.
4.4. Comparison of Algorithm Complexity and Running Results
In order to validate the proposed FYDSE algorithm against other algorithms in terms of complexity and running results, we provide the algorithmic complexity of the proposed algorithm and other methods in
Table 4, where
Tmax represents the maximum iteration,
M represents the number of function evaluations,
n represents the number of populations, and
d represents the dimension.
The table shows that the YDSE, SO, HBA, SSA, WSO, and COOT algorithms have the same complexity. Also, the DO, AO and SCSO algorithms have the same complexity. The proposed algorithm, only in the second term of the complexity, has a difference in coefficients with YDSE, SO, HBA, SSA, WSO, and COOT, mainly due to the introduced crossover strategy. In fact, real-world optimization problems are complex and nonlinear. Therefore, when confronted with real-world optimization problems, the complexity of the proposed FYDSE is comparable to that of YDSE, SO, HBA, SSA, WSO, COOT, and AHA and outperforms DO, AO, and SCSO and significantly outperforms ChOA. To demonstrate this conclusion more effectively, we provide the runtime results for the ten functions in the cec2022 suite in
Table 5.
From the results in the table, the overall operational results of the proposed FYDSE are better than those of the ChOA, SCSO, and DO algorithms. Also, the gap between the running results of the proposed algorithm and the other algorithms is considered small. At the same time, this gap will gradually decrease as the complexity of the objective function of the problem under study increases. Therefore, the complexity of the proposed FYDSE is improved within the acceptable range, and it is important to note that it cannot be ignored that the proposed algorithm obtains better performance results in these benchmark suites.
4.5. Search Capability Analysis of Global Optimal Solutions
In order to verify the improvement in the introduced initialization and exploration strategies on the global solution searchability of the YDSE algorithm, we provide different methods to obtain the probability of the global optimal solution. First, considering that some test functions are difficult to converge to the optimal position, we consider that the candidate solution of the solution is considered optimal if it converges to the position of the neighborhood of the optimal solution, i.e., the relative error ((Candidate Solution − Optimal Solution)/Candidate Solution) is < 10
−2. We provide the results of the probability of reaching the optimal solution for the twenty runs of the proposed FYDSE and YDSE, as well as the other methods, in
Table 6.
The results in
Table 6 show that the proposed FYDSE achieves the best probability of reaching the optimal solution for most of the test functions and has the best final average ranking. Compared with YDSE, the proposed FYDSE has a higher probability of searching for the optimal solution in cec02, cec07, cec08, and cec10. The main reason is that the probability of obtaining the global optimal solution is improved by introducing a dynamic adversarial strategy that effectively extends the full exploration of seldom-searched regions.
5. An Example of Complex Engineering Optimization: The Four-Stage Gearbox Problem
To further validate the performance of the proposed FYDSE in solving complex engineering optimization problems, the proposed algorithm is experimentally compared with other algorithms, including the original YDSE, in a complex engineering optimization example: a four-stage gearbox problem. A schematic diagram of the four-stage gearbox problem is shown in
Figure 7. The objective of the problem is to minimize the weight of the gearbox, which contains 22 discrete independent variables [
58]. They are categorized into four types, including gear position, pinion position, billet thickness, and number of teeth. In addition, the problem contains 86 nonlinear design constraints related to pitch, kinematics, contact ratio, gear strength, gear assembly, and gear size [
44]. The mathematical model is shown below:
Table 7 gives the minimum, worst, and average gearbox weights and values of the design variables obtained after 20 runs of the proposed FYDSE with YDSE, AHA, SO, DO, AO, ChOA, HBA, SCSO, SSA, WSO, and COOT. Also, the running times of all the methods are given at the end of the table. The gearbox weight box plots for FYDSE and other methods at 20 runs are given in
Figure 8. From the results, the difference between the running time of the proposed FYDSE method and YDSE is small when facing complex optimization problems due to the high complexity of the objective function. Also, the proposed algorithm obtains the best minimum and average gearbox weight with a significant difference compared to other methods and YDSE. Therefore, the introduced four strategies can effectively improve the algorithm’s ability to face complex engineering and realistic optimization problems.
6. Specific Truss Topology Optimization Problems (TTOP)
To validate the efficacy of FYDSE in tackling intricate engineering optimization challenges, it is employed to optimize the topology of the truss structure. Compared with the cross-section optimization of trusses, topology optimization attempts to streamline the mass of the truss by eliminating unnecessary members and nodes. Therefore, topology optimization can effectively reduce the cost when the cost of nodes is large. In our implementation, we impose frequency constraints on the intrinsic frequencies to ensure that they are effectively limited. Conventional optimization techniques, including sensitivity analysis, have garnered widespread application in addressing truss optimization challenges and have yielded notable outcomes. However, the efficacy of traditional methods in tackling intricate optimization challenges remains in need of enhancement. Meta-heuristic algorithms are an effective solution to this problem since they can fully explore the nonlinear and nonconvex space of topological optimization and maintain the topological results. Therefore, in this section, the efficacy of the FYDSE algorithm in addressing highly nonlinear and nonconvex TTOP problems is further validated. Moving forward, we present the mathematical model for topology optimization.
6.1. TTOP Model
The primary goal of the TTOP was to ascertain the best truss structure and layout to ensure that the ground structure could carry the minimum possible loads. This ground structure comprises both necessary and discretionary nodes. Necessary nodes are typically regarded as promptly bearing structural, load, and nodal stresses [
59]. Conversely, including discretionary nodes enhances the stress distribution among the various components. To embark on this endeavor, we first define the key constraints and objective functions of the TTOP model.
Suppose variable
X= {
A1,
A2, …,
Ap} and the target function of TTOP is formulated as follows:
where
p means numbers of variables and
q means numbers of nodes.
ρi,
Lengthi,
Ai signifies the elemental density, length, and transverse area of the
ith variable.
bj denotes the quantity of the
jth node. We set a critical region with a small positive value to determine which elements to discard [
60]. When the cross-sectional area of an element falls below a certain minimal threshold, it becomes necessary to discard it.
Further, we give multiple constraints for the TTOP problem, including stress, displacement, Euler buckling, intrinsic frequency, and upper and lower cross-section constraints. The specific definitions are shown below:
where
σi implies the stress of the ith node.
and
represent the upper and lower boundary for stress constraints.
δj means the displacement of the
ith knot.
and
represent the upper and lower boundary for displacement constraints, respectively.
fr means the intrinsic frequency of the truss structure in
rth mode. In addition,
and
denote the scope of variables, respectively.
In order for the variables to satisfy the above constraints, we introduce a penalty term in the objective function [
61]. As illustrated in Equation (39), when the constraints are obeyed, the penalty term will be reduced to zero. However, a positive penalty will be imposed in the event of a constraint violation. The following outlines the specific penalty function in detail:
where the
penalty term is set to zero if the independent variable adheres to the specified constraint.
F(
A) means the weighted objective function. However, if the constraint is violated, a positive penalty is imposed. The precise formulation of this penalty function is detailed as follows:
where
g denotes the tendency of a variable to violate a constraint and be subject to a
,
ac representation of active restraint.
ε is set to 2 [
62].
In order to fully validate the performance and usefulness of FYDSE in solving the optimized TTOPs for complex projects, three TTOPs with different numbers of trusses are discussed in this section. In these three TTOPs, all independent variables are considered to be continuous, and the Euler buckling coefficients and nodal masses are fixed to 4 kg and 5 kg. The cross-section size range is set between Amax and −Amax. To fairly assess the engineering optimization capabilities of FYDSE, we compare it with the original YDSE as well as other latest optimization algorithms, including AHA, SO, DO, AO, ChOA, HBA, SCSO, SSA, WSO, and COOT. To ensure the fairness and repeatability of the optimization TTOP, we made the population size of all algorithms 30 with a maximum iteration of 1000. In addition, all algorithms are tested with 30 independent repetitions.
6.2. 20-Truss Topology Optimization
Figure 9 illustrates the initial schematic ground architecture of a 20-rod truss consisting of 20 members and 9 knots. In this figure, all 20 members and 9 nodes are clearly labeled. Of particular note is that the weights of this truss structure are mainly carried by the supports of node 1 and node 9.
Furthermore, to enhance the optimization of this truss structure, we have compiled a comprehensive list of detailed constraints and inherent material properties in
Table 8. Additionally,
Table 9 comprehensively displays the experimental outcomes obtained by various comparative methods in addressing the 20-truss optimization problem. Among the experimental results, according to a model introduced for TTOP, we further optimized the design of the truss structure by deleting some unwanted parts (indicated by “-”) according to the requirements.
The outcomes of the independent variables and objective functions of YDSE, AHA, SO, DO, AO, ChOA, HBA, SCSO, SSA, WSO, COOT and the proposed FYDSE are given in
Table 9. Upon examination of the research results outlined in
Table 9, it is evident that the FYDSE and the AHA, HBA, and COOT all attained the desired minimum optimal weight, measuring precisely 154.799. However, among these algorithms, the proposed FYDSE algorithm shows its unique advantage, with an average weight of only 164.718, significantly smaller than the average weights of other algorithms. In contrast, the SSA and SO algorithms ranked second and third, with their optimal structural qualities of 155.347 and 155.574, which failed to surpass the performance of the FYDSE algorithm despite their proximity. This result indicates that the FYDSE algorithm has higher efficiency and stability in weight optimization.
Figure 10 illustrates the optimal topology achieved through the FYDSE method and a comparative analysis with alternative algorithms during optimization. Specifically, all eight algorithms, FYDSE, YDSE, AHA, SO, DO, HBA, SSA, and COOT, select eight members as the basis of the topology, and these members have the same numbering in their respective algorithms. However, the CHOA algorithm employs a different strategy for retaining only six members. Although the CHOA algorithm requires a more streamlined component structure than that suggested by the FYDSE algorithm, it is more stringent in terms of the quality of the structure. As shown in
Figure 10a, YDSE, AHA, SO, DO, HBA, SSA, COOT, and FYDSE present consistency in topology, and their main difference lies in the fact that there are different sizes of cross-sections in the structure. These algorithms provide a valuable reference for solving practical problems, especially when applying metaheuristics. It is worth mentioning that our proposed FYDSE algorithm shows excellent performance in solving a topology optimization problem containing 20 trusses. In contrast to other SOTA algorithms, FYDSE performs better in terms of optimization efficiency and quality of results, further validating its effectiveness and usefulness in topology optimization.
6.3. 24-Truss Topology Optimization
Figure 11 shows a structural diagram for an original ground configuration of a 24-rod truss consisting of 24 members and 8 knots, all clearly labeled. Notably, node 3 supports a non-structural concentrated mass of 500 kg. While this additional weight does not contribute to the overall weight of the truss, it significantly influences the design considerations for the positioning and dimensions of both nodes and members.
To facilitate the optimization process for the 24-rod truss structure, we have compiled a list of pertinent constraints and the inherent material properties in
Table 10. Moreover, the results of comparing the method with FYDSE in terms of independent variables and minimum weight is presented in
Table 11. Among the experimental results, according to the introduction of the TTOP model, we further optimized the design of the truss structure by deleting some unwanted parts (indicated by “-”) according to the requirements.
The consequences of the independent variables and objective functions of YDSE, AHA, SO, DO, AO, ChOA, HBA, SCSO, SSA, WSO, COOT and the proposed FYDSE are shown in
Table 11. Based on the results of the experiment, we can see that the HBA, COOT, and our proposed FYDSE algorithms achieve the same and the lowest result, i.e., 126.252, in seeking the optimal weights, and none of them violates the set constraints. However, among these algorithms, the proposed FYDSE algorithm shows its unique advantage, with an average weight of only 139.951, significantly smaller than the average weights of other algorithms. In contrast, the HBA and COOT algorithms positioned themselves in the second and third spots, respectively, boasting optimal structural qualities of 178.958 and 146.784, which failed to surpass the performance of the FYDSE algorithm despite their proximity. The FYDSE algorithm performs optimally in terms of average weights, a result that further proves the stability and reliability of the FYDSE algorithm.
In
Figure 12, a comparative analysis is presented, showcasing the optimal topology attained by the FYDSE algorithm in contrast to that of other algorithms. FYDSE, COOT, SO, AO, and SSA successfully retain seven building blocks among these algorithms. FYDSE and COOT corresponded to the same number of building blocks, while SO, AO, and SSA chose different numbers. Although the HBA algorithm retains only six building blocks, this may imply that it is more stringent regarding structural quality. The topology of the FYDSE algorithm is more reasonable regarding layout and connections looking at the structure of the figure. When dealing with the problem of topology optimization with a 24-rod truss, our proposed FYDSE algorithm outperforms other recent methods.
6.4. 72-Truss Topology Optimization
Figure 13 comprehensively illustrates the original ground structure schematic of a 72-rod truss comprising 72 members and 20 knots, with select members and nodes labeled for clarity. It is crucial to consider the impact of the center on the positioning and dimensions of these members and knots when determining the ultimate total weight of this 72-rod truss structure.
In order to optimize this truss structure, we have detailed the materials’ associated constraints and inherent properties in
Table 12.
In addition, the results of comparing the method with FYDSE in terms of independent variables and minimum weight are presented in
Table 13. Among the experimental results, according to the introduction of the TTOP model, we further optimized the design of the truss structure by deleting some unwanted parts (indicated by “-”) according to the requirements.
Drawing from the experimental data presented in
Table 13, we observe that the SCSO, COOT, and our proposed FYDSE algorithm attain the lowest optimal weight value of 450.388. Nevertheless, the FYDSE algorithm stands out in its performance, exhibiting a significantly lower average weight value of 454.209, surpassing all comparable algorithms. These data further highlight the stability and reliability of the FYDSE algorithm. Among the comparison algorithms, the SO algorithm also performs exceptionally well, with an average weight of 458.213, the second highest, and its structural quality is equally satisfactory.
To have a more intuitive understanding of the optimization effect of these algorithms,
Figure 14 exhibits the optimal topologies derived through FYDSE and its comparative algorithms [
63,
64].
Figure 14a indicates that the YDSE, SO, HBA, SCSO, WSO, COOT, and FYDSE algorithms retain the same number of building blocks during optimization. In addition, the AHA and ChOA algorithms also retain the same number of building blocks. These algorithms remain consistent in the number of building blocks and labeling, but the cross-sectional area in the topology varies in size. In addressing the optimization challenge of truss topology, our FYDSE surpasses the performance of other cutting-edge algorithms, exhibiting remarkable efficiency and effectiveness. It proves once again its superiority in the field of optimal structural design.
8. Conclusions
In order to solve the problems that the original YDSE will easily fall into local optimums, converge slowly, and have an imbalance between exploration and exploitation when dealing with sophisticated engineering optimal problems, this paper presents a fractional-order boosted hybrid YDSE for solving complex optimization problems. FYDSE introduces a piecewise chaotic mapping strategy, a fractional-order improvement strategy, a dynamic dyadic strategy, and a vertical operator. First, the fractional-order strategy is introduced in the dark edge position update of FYDSE to ensure that the search potential of a single neighborhood space is exploited more efficiently while decreasing the likelihood of trapping in a local optimum. Secondly, during the initialization phase, the incorporation of piecewise chaotic mapping is intended to yield a high-quality primary population, thereby enhancing algorithm convergence efficiency. Furthermore, the low exploration space of FYDSE is extended by using a dynamic inversion strategy, which improves the probability of obtaining a globally optimal solution. Finally, by introducing the vertical operator, FYDSE can better balance global exploration and local exploitation and explore new unknown areas. Comparative experiments of FYDSE with a series of state-of-the-art algorithms are conducted in the CEC2022 test suite, the four-stage gearbox problem, and three TTOP cases, respectively. The Wilcoxon rank-sum test statistically proves the effectiveness of FYDSE. In addition, the average ranking of the mean result and probability of FYDSE in cec2022 are 1.42 and 1.33, respectively, verifying that FYDSE can better converge to the optimal solution. Although FYDSE increases the constant complexity by introducing improvement strategies, this difference is small when facing complex problems as analyzed through experiments. The efficient performance of the four-stage gearbox problem and the three TTOP cases verifies that FYDSE is an effective method for solving complex engineering optimization challenges.
Many complex, nonlinear, and high-dimensional realistic optimization problems, such as complex path planning for UAVs, neural network parameter optimization, and curved shape parameter optimization problems, are yet to be solved in scientific and practical research. The better experimental results for numerical and engineering optimization problems show that the proposed FYDSE is an optimization method that can be extended in different fields. Therefore, FYDSE can be used as a potential solution to solve various complex real-world problems effectively.
As a direction for future work, we will strive to extend FYDSE to other domain-specific optimization problems, such as UAV path planning and image threshold segmentation. Moreover, we will also explore the integration of FYDSE with machine learning and deep learning techniques, such as using FYDSE to solve feature selection problems or for deep learning parameter optimization.