Next Article in Journal
Bioarchitectonic Nanophotonics by Replication and Systolic Miniaturization of Natural Forms
Next Article in Special Issue
An Improved Binary Walrus Optimizer with Golden Sine Disturbance and Population Regeneration Mechanism to Solve Feature Selection Problems
Previous Article in Journal
Chemical Synthesis and Structure–Activity Relationship Studies of the Coagulation Factor Xa Inhibitor Tick Anticoagulant Peptide from the Hematophagous Parasite Ornithodoros moubata
Previous Article in Special Issue
A Multi-Strategy Improvement Secretary Bird Optimization Algorithm for Engineering Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Osprey Optimization Algorithm Based on Two-Color Complementary Mechanism for Global Optimization and Engineering Problems

School of Mechanical and Precision Instrument Engineering, Xi’an University of Technology, Xi’an 710048, China
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(8), 486; https://doi.org/10.3390/biomimetics9080486
Submission received: 20 June 2024 / Revised: 9 August 2024 / Accepted: 10 August 2024 / Published: 12 August 2024
(This article belongs to the Special Issue Nature-Inspired Metaheuristic Optimization Algorithms 2024)

Abstract

:
Aiming at the problem that the Osprey Optimization Algorithm (OOA) does not have high optimization accuracy and is prone to falling into local optimum, an Improved Osprey Optimization Algorithm Based on a Two-Color Complementary Mechanism for Global Optimization (IOOA) is proposed. The core of the IOOA algorithm lies in its unique two-color complementary mechanism, which significantly improves the algorithm’s global search capability and optimization performance. Firstly, in the initialization stage, the population is created by combining logistic chaos mapping and the good point set method, and the population is divided into four different color groups by drawing on the four-color theory to enhance the population diversity. Secondly, a two-color complementary mechanism is introduced, where the blue population maintains the OOA core exploration strategy to ensure the stability and efficiency of the algorithm; the red population incorporates the Harris Hawk heuristic strategy in the development phase to strengthen the ability of local minima avoidance; the green group adopts the strolling and wandering strategy in the searching phase to add stochasticity and maintain the diversity; and the orange population implements the optimized spiral search and firefly perturbation strategies to deepen the exploration and effectively perturb the local optimums, respectively, to improve the overall population diversity, effectively perturbing the local optimum to improve the performance of the algorithm and the exploration ability of the solution space as a whole. Finally, to validate the performance of IOOA, classical benchmark functions and CEC2020 and CEC2022 test sets are selected for simulation, and ANOVA is used, as well as Wilcoxon and Friedman tests. The results show that IOOA significantly improves convergence accuracy and speed and demonstrates high practical value and advantages in engineering optimization applications.

1. Introduction

Optimization can be formally defined as the process of identifying the optimal solution from among all feasible alternatives within a given problem. More specifically, it involves the maximization or minimization of a multidimensional function subject to a set of constraints. Optimization problems (OPs) are ubiquitous in real-world applications, encompassing areas such as business planning, ecological management, engineering design, and industrial control [1]. However, most real-world optimization problems are inherently complex and challenging to solve, with their complexity manifesting in multiple dimensions, such as multivariate, multi-objective, multi-constrained, nonlinear, multimodal, non-differentiable objective functions, and non-differentiable constraints. Mathematically, these problems are often classified as NP-hard, meaning that finding exact solutions requires substantial computational resources and may not be solvable in polynomial time. In this context, classical optimization methods such as gradient descent, branch-and-bound, cutting plane techniques, and dynamic programming often prove inadequate due to limitations such as susceptibility to local optima and high computational requirements in terms of both time and space. As a result, there is a pressing need for more efficient and robust methodologies to address complex OPs. To tackle OPs, optimization techniques can generally be categorized into three classes: exact methods, heuristic algorithms, and metaheuristics. The use of traditional exact optimization methods is typically impractical for complex problems due to their prohibitive computational demands. Heuristic algorithms, while more computationally feasible, tend to be highly dependent on specific problem characteristics. In contrast, metaheuristic algorithms are more advanced, employing problem-independent techniques, and are applicable to a wide range of optimization problems.
In the past 30 years, metaheuristic algorithms, as a stochastic approximate optimization method, have greatly overcome the above drawbacks and excelled in solving complex OP problems, which have attracted extensive attention and research. Among them, there are evolution-based metaheuristic algorithms, also known as evolutionary algorithms. They arise from the rules and processes of genetic and population evolution in nature and usually require iterative mutation, crossover, selection, and other related evolutionary operators. The most common are evolutionary planning (EP, 1966) [2], genetic algorithms (GA, 1975) [3], and differential evolution (DE, 1997) [4]. There are also physically based metaheuristic algorithms. These algorithms have a different optimization mechanism than evolutionary algorithms. Their search agents will exchange information and update the state in the search space according to physical rules. The most representative ones include Simulated Annealing (SA, 1983) [5], Golden sine algorithm (GSA, 2017) [6], and Chernobyl Disaster Optimizer (CDO, 2023) [7]. Secondly, there are also metaheuristics inspired by human concepts or behaviors. Examples include Cultural Algorithms (CA, 1994) [8], Harmony Search (HS, 2001) [9], Group Search Optimizer (GSO, 2006) [10], Student psychology-based optimization algorithm (SPBO, 2020) [11], mountaineering team-based optimization (MTBO, 2023) [12], etc. Finally, it is common to find population-based metaheuristic algorithms, which simulate the group behavior of a variety of animals and use information between groups of organisms as well as between them and their environment for communication and cooperation. Moreover, the goal of optimization is achieved by simple and limited interactions with experienced and intelligent individuals. The most classical ones are Particle Swarm Optimization (PSO, 1995) [13], Ant Colony Optimization (ACO, 1996) [14], Bat Algorithm (BA, 2012) [15], Sailfish Optimization Algorithm (SFO, 2019) [16], Remora Optimization Algorithm (ROA, 2021) [17], Waterwheel Plant Algorithm (WWPA, 2023) [18], Osprey Optimization Algorithm (OOA, 2023) [19], etc. Among them, OOA is a new swarm intelligence algorithm that is a fast, powerful, and high-performance optimization algorithm.
The algorithm emulates the natural foraging behavior of fish eagles, conferring a significant competitive edge over other heuristic algorithms. Although OOA possesses numerous advantages, it exhibits certain shortcomings when addressing complex engineering optimization problems, such as lower convergence accuracy and a tendency to become trapped in local optima. To enhance the overall performance of OOA, several researchers have introduced modifications aimed at overcoming these limitations. Yongliang Yuan et al. proposed the Attack-Defense Strategy-Assisted Osprey Optimization Algorithm (ADSOOA) [20], which integrates attack and defense phases, utilizing dynamic search parameters to strengthen global exploration capabilities and local exploitation strategies, thereby improving the algorithm’s convergence performance and avoiding local optima. Xiaodong Wen’s innovative multi-strategy fusion Improved Osprey Optimization Algorithm (SOOA) [21] significantly enhances the algorithm’s versatility and search efficiency through chaotic mapping initialization and a dynamically adjustable elite bootstrap mechanism. Yi Zhang et al. introduced an Improved Osprey Optimization Algorithm (IOOA) [22] for mixed-strategy optimization, which employs a Sobol sequence for initialization, a Weibull distribution step factor, and a perturbation strategy based on the principles of the firefly algorithm, achieving a highly diverse initial population and a balanced global exploration and local exploitation capability. The aforementioned literature represents improvements to OOA, enhancing its performance, yet further research is needed to better improve convergence accuracy and speed, as well as to achieve a balance between global search and local exploitation capabilities.
The aforementioned literature represents improvements to the Osprey Optimization Algorithm (OOA), which enhances its performance. However, further research is required to better improve convergence accuracy and speed, as well as to balance the global search and local exploitation abilities. Given this, the primary contributions of this paper are as follows:
Initialization: The population is initialized by combining logistic chaos mapping and the good point set method, ensuring a diverse initial population. The four-color theory is then applied to divide the population into four color groups (blue, red, green, and orange), aiming to balance the diversity of the population with the algorithm’s global search and local development capabilities.
Two-Color Complementary Mechanism: A two-color complementary mechanism is introduced in the algorithm design. The blue population maintains the original exploration strategy (OOA) to preserve fundamental effectiveness and stability. The red population adopts a heuristic strategy inspired by the behavior of the Harris Hawk to enhance the avoidance of local minima. The green population implements a strolling strategy to broaden the search boundaries and maintain population diversity. Lastly, the orange population combines the spiral search strategy and the firefly perturbation strategy to deepen the exploration of the solution space and perturb local optimal solutions, respectively, promoting the algorithm’s ability to escape from local optima and sustain long-term exploration vitality.
Comprehensive Validation: This paper conducts extensive validation through simulations of classic benchmark functions, CEC2020 and CEC2022 test sets, and statistical tests (ANOVA, Wilcoxon, and Friedman). The results show that the convergence accuracy and speed of IOOA are significantly improved while verifying its practical value and advantages in engineering optimization applications.
The article is structured as follows: Section 2 outlines the basic principles of OOA. Section 3 describes the details of the improved IOOA algorithm. Section 4 shows the results of testing and analyzing the new algorithm in comparison with the existing algorithms. Section 5 applies to the optimization examples of real engineering problems. Section 6 provides a detailed performance analysis of IOOA, summarizes the research results, and looks at future research directions.

2. OOA

Dehghani et al. were inspired by the fishing activities of fish eagles from lakes and proposed a novel heuristic optimization algorithm, Osprey Optimization Algorithm (OOA), in 2023. The algorithm simulates the process of fishing by fish eagles from lakes: the fish eagles capture their prey after detecting its location and then bring it to a suitable location to eat it. Therefore, in OOA, the location change process of the fish eagle population consists of two phases: the first phase is for the fish eagles to recognize the location of the fish and catch it (global exploration), and the second phase is for the fish eagles to bring the fish to a suitable location (local exploitation) after the first phase of the fish eagles’ recognition of the location of the fish and catching it (global exploration). This population intelligence algorithm achieves efficient exploration of the search space by simulating the fishing behavior of the fish eagle, which helps solve complex optimization problems.

2.1. Population Initialization

OOA is an intelligent population-based optimization algorithm inspired by the hunting behavior of ospreys in nature. Similar to other intelligent optimization algorithms, it performs random initialization of population operations in the search space by means of a population initialization formula:
x i , j = d j + r i , j w j d j , i = 1,2 , , N ,       j = 1,2 , , m
where x i , j is the individual, d j is the lower search limit, w j is the upper search limit, and r i , j is a random number between [0, 1].

2.2. Positioning and Fishing

In the OOA design, for each osprey, the locations of other ospreys in the search space with better objective function values are considered underwater fish. The first stage of population updating in OOA is modeled based on the simulation of the natural behavior of osprey hunting. The modeling of ospreys attacking fish results in significant changes in the positions of ospreys in the search space, which increases the exploratory power of the Osprey Optimization Algorithm in identifying optimal regions and escaping from local optima. In the design of OOA, for each osprey, the locations of other osprey in the search space with better objective function values are considered as underwater fish. The equation for the position of the fish corresponding to each osprey is shown below:
Q i = X k k 1,2 , , N F k < F i X b e s t
where Qi, is the set of locations of the ith fish eagle, Xbest is the best candidate solution, N is the number of fish eagle populations, F and F are the objective function values corresponding to the kth and ith fish eagle, respectively.
The fish eagle randomly detects the position of one of the fish and attacks it. Based on the simulated movement of the fish eagle toward the fish, the new position of the corresponding fish eagle is updated using the following Equations (3) and (4).
x i , j P 1 = x i , j + r i , j H i , j I i , j x i , j
X i , j t + 1 = x i , j P 1                         d j x i , j P 1 w j , d j                                         x i , j P 1 < d j , w j                                         x i , j P 1 < w j ,
X i = Q x i P 1 ,               F i p 1 < F i X i ,         o t h e r w i s e
where x i P 1 is the new position of the ith osprey in the first stage of the osprey-based optimization algorithm, x i , j P 1 is its jth dimension, H i the fish selected by the jth osprey in the ith iteration, H i , j is its jth dimension and is a random number in the interval [0, 1], and I i , j is a random number from the set {1, 2}. If this new position is better, the previous position of the osprey is replaced according to Equation (5). where F i P 1 is its objective function value.

2.3. Hunting and Feeding

After hunting a fish, the osprey takes it to a safe location to eat it. Phase 2 of the population update of OOA is based on the simulation of the natural behavior of the osprey to eat fish. In the design of the OOA, this behavior of the osprey is modeled. For each member of the population, a new random position is calculated as the “suitable position for eating fish” using Equations (6) and (7).
x i , j P 2 = x i , j + d j + r w j d j t , i = 1,2 , , N , j = 1,2 , , m , t = 1,2 , , T
X i , j P 2 = x i , j P 2                 d j x i , j P 2 w j , d j                             x i , j P 2 < d j , w j                             x i , j P 2 < w j ,
X i = Q x i P 2 , F i P 2 < F i X i , o t h e r w i s e
where X i P 2 is the new position of the ith osprey based on OOA, x i , j P 2 is its jth dimension, x i P 2 is the optimal value of its objective function, t is the number of iterations, and T is the maximum total number of iterations. Then, if the objective function value is improved at this new position, it will replace the previous position of osprey according to Equation (8).

3. IOOA

In this section, IOOA will be introduced in detail, which is an efficient search algorithm that incorporates multiple optimization strategies based on the OOA. In the initialization phase of the IOOA, the initial populations generated by the strategy of fusing the logistic mapping with the set of good points in order to produce a diverse and uniformly distributed initial populations; secondly, drawing on the well-known conclusions of the Graph Theory, the four-color theorem divides the populations into four parts, which are the red group, orange group, blue group, and green group, and the focus of the search methods of different groups is different, and the four groups of populations will use the two-color complementary mechanism to improve the position update, and these strategies together constitute the core search mechanism of the IOOA algorithm.

3.1. L-Good Point Set Initialization Strategy

In the initialization phase of the OOA, the population is constructed using the L-good point set method, which is an initialization strategy that combines the logistic chaotic mapping [23] and the good point set method [24]. The L-good point set method carefully selects or constructs a series of “good points” in the solution space through specific mathematical constructs (e.g., based on prime numbers, cosine function, etc.), which usually have good spatial distribution characteristics and can cover the key regions of the solution space, avoiding the problem of concentrated or sparse distribution of points that may be caused by random initialization. Logistic mapping, as a kind of chaotic mapping, is capable of generating sequences with a high degree of randomness and long-term unpredictability, which helps expand the search range of feasible solutions, balancing the trade-off between local exploitation performance and the global search capability of the algorithm.

3.1.1. Good Point Set Method

Good point set theory is a branch of mathematics that focuses on the set formed by all points that are a certain distance away from a point in a Euclidean space. This theory has a wide range of applications in geometry, topology, and number theory. An intuitive way to choose n points from X to maximize the likelihood that it contains an optimal point for a thesis X when nothing is known about its traits is to take n points and make it the most uniformly distributed in X. Because of the most uniform distribution, the probability of taking the optimal point is maximized. To be uniform, minimizing deviation in number theory is uniformity, and this problem of minimizing deviation can be explained by using good points in number theory.
In practice, the following three methods are usually used:
Exponential sequence methods: r k = { e k , 1 k m } ;
Square root sequence method: r k = p k , 1 k m , r k are mutually unequal prime numbers;
Split-circle domain method: r k = 2 cos 2 π k p , 1 k m , p is the smallest prime number satisfying the condition (p − 3)/2 ≥ m.
The principle of good point sets used in this paper is as follows:
Let G s be a unit cube in s-dimensional Euclidean space if r G s , of the form.
P n k = r 1 n k , r 2 n k , , r s n k , 1 k n
Its deviation φ ( n ) satisfies φ ( n ) = C r , ε n 1 + ε , where C r , ε is a constant related only to r and ε (ε is an arbitrary positive number), then P n k is said to be the set of good points, and r is the good point. { r s ( n ) · k } stands for taking the fractional part, n denotes the number of points, and r = {2 cos(2πk/p), 1 ≤ ks} (p is the smallest prime number satisfying (p − 3)/2 ≥ s). Map this to the search space as follows:
x i j = u b j l b j r j i k + l b j
u b j and l b j denote the upper and lower bounds of the jth dimension.
The initialization problem of IOOA is essentially how to search a larger space using a finite number of individual ospreys. If the initial population is randomly generated, it cannot traverse the various conditions in the solution space. Utilizing the construction method of the theory of good point sets, its computational accuracy is independent of the number of dimensions, so drawing on the method of good point sets to uniformly design the initial osprey population can overcome the shortcomings of the above methods and can produce an initial population with better diversity.

3.1.2. Logistic Chaotic Map

Logistic chaotic mapping demonstrates significant benefits in population initialization by virtue of its ability to generate highly stochastic and long-term unpredictable sequences. This study integrates logistic chaotic mapping with the preferred point set technique with the aim of optimizing the quality of the algorithm’s starting solution. The mapping formula is as follows:
x i t + 1 = a x i t 1 x i t
Here, x i t and x i t + 1 denotes the current and next iteration states of the ith variable within the interval [0, 1], respectively; the parameter a, governing the degree of chaos, ranges within (0, 4], with a = 4 marking full chaotic behavior. The chaotic trajectory spans the interval (0, 1).

3.1.3. L-Good Point Set Method

In optimization algorithms, the quality and diversity of the initial population are crucial to the algorithm’s optimization search ability and convergence performance. In order to enhance the global search ability and convergence performance of the algorithm, logistic mapping is fused with the good point set method—L-good point set method. The main principle of this strategy is to combine the chaotic sequences generated by logistic mapping with the good point set strategy during the initialization process, implying that each element of the chaotic sequences is mapped onto the unit sphere to form a good point set. This combination produces an initial population that is both stochastic and maintains structure by combining the randomness of the chaotic dynamical system with the orderliness of the other structures to improve search efficiency and global search capability. This strategy not only enhances the optimization ability and search efficiency of the whole population but also speeds up the convergence of the algorithm. In this way, the optimization algorithm is provided with a stronger search ability in the initialization stage and better solves the actual optimization problem.
This strategy can effectively avoid the algorithm’s blind search, reduce the solution efficiency, and lay the foundation for improving the solution accuracy and accelerating the convergence speed. Figure 1 clearly shows the comparison between the fish eagle population before and after initialization.

3.2. A Two-Color Complementary Strategy Based on the Four-Color Theorem

OOA, as an emerging bio-heuristic optimization algorithm, has a strong ability to find the optimal and certain search accuracy, but any algorithm has its limitations, and the OOA is no exception. In some complex optimization problems, different regions of the solution space may require different search strategies. The population of OOA is not classified, and all individuals may follow the same search rules and step sizes, which may limit the algorithm’s adaptability to different regions and exploration efficiency. A single search strategy may perform poorly in dealing with problems with multiple local optimal solutions, which may easily cause the algorithm to prematurely converge to a non-global optimal solution, especially when there is no effective mechanism to promote population diversity or escape from local optimality. To address these shortcomings of OOA, this paper adopts a two-color complementary strategy based on the four-color theorem [25] for fish eagle populations.

3.2.1. Stock Grouping Adjustment Mechanism

In OOA, the execution of the same search strategy by all populations may bring certain limitations. When encountering problems with more solution spaces, it is difficult to fully search the global solution space due to fewer population searching methods, and when multiple local optimal solution problems are encountered, it is difficult to jump out of the local optimal solution due to the single searching method. In order to overcome this problem, this paper draws on the four-color theorem in number theory to introduce a grouping adjustment mechanism, which divides the population into four-color populations composed of two pairs of complementary colors, which are the red group and the green group as well as the blue group and the orange group, respectively.
Initially, set the mathematical expression for the ratio of the red group to the green group:
P r e d = 0.4
P g r e e n = 0.4
As the main populations in the IOOA algorithm, the red group and the green group act as the main exploring group, whose main task is to carry out extensive and in-depth exploration in the solution space and work on discovering new solution regions. Through extensive exploration, the exploration group helps prevent the algorithm from falling into local optimality too early, provides fresh solution ideas for the whole algorithm, and increases the probability of finding the global optimal solution. Initially, set the mathematical expression for the ratio of the blue group to the orange group:
P b l u e = 0.1
P o r a n g e = 0.1
The blue group and the orange group are in development in the IOOA algorithm, whose main responsibility is to perform fine search and local optimization around the found better solution regions. By focusing on the refinement of local regions, the development group helps improve the quality of the solutions and promotes the algorithm to converge to higher quality solutions, especially for those complex optimization problems that contain multiple local minima. Implementing such a population grouping tuning strategy performs different strategies by grouping the overall population while improving the overall performance and optimization search efficiency. Figure 2 illustrates the grouping based on the four-color theorem in the population.

3.2.2. Two-Color Complementary Mechanism

In OOA, a certain global search capability is maintained at the initial stage by randomly selecting “prey” and a large exploration step size, which helps cover a wide search space. However, if the initial setup is poor or the adaptation is complex over a wide area, the exploration phase may not be sufficient to fully traverse all critical regions, and as the iteration progresses, especially in the second phase, the algorithm shifts to rely more on searching in the vicinity of the current optimal solution, and the local search capability increases. While this helps refine the quality of the solution, it may also fall into a local optimum by shrinking the search range too quickly. To solve the above problem, a two-color complementary mechanism is introduced, where the current population is divided into red and green groups, blue group, and orange group according to the above grouping based on the four-color theorem, and different update strategies are adopted for the populations of different color groups. Specifically, for the red and green groups, as the main exploratory groups, a larger variation step or a more aggressive search strategy can be adopted to skip the neighborhood of existing solutions and find the potential global optimal solution. For the blue and orange groups, a smaller variation step size or a local search strategy can usually be adopted to gradually improve the existing solutions. At the same time, this paper introduces a two-color complementary mechanism to strengthen the correlation between populations by adopting the method of mutual exchange between the exploring and exploiting groups within the group, such that the different updating strategies of multiple populations can help to balance the exploration and utilization of individuals in the search space, improve the global search performance of the algorithm, and thus search for the optimal solution more efficiently.
Determine the size of the red, green, blue, and orange group populations:
N r e d = N · P r e d : red group population size;
N g r e e n = N · P g r e e n : green group population size;
N b l u e = N · P b l u e : blue group population size;
N o r a n g e = N · P o r a n g e : orange group population size.
After randomly sorting according to fitness, the population is divided into four parts, and the population exchange within the exploration group and within the exploitation group is realized at the same time. The following is the design of the specific exchange method:
p 0 = 0.5 : the initial intra-group exchange probability;
δ = 0.02 : the decay rate of the exchange probability in each generation;
p m i n = 0.2 : the minimum exchange probability;
p t = max p 0 × ( 1 δ ) ^ t , p m i n : the formula for the decay of the exchange probability, t is the number of iterations. The specific two-color complementary exchange idea is shown in Figure 3.
IOOA employs a two-phase strategy framework that aims to balance global exploration with local exploitation capabilities through the two-color complementary mechanism. This framework is divided into two core phases: the initial broad exploration (exploration phase) aims to explore the vast potential of the solution space, while the subsequent deep optimization (exploitation phase) focuses on escaping the local optimum and refining the solution. Based on this framework, this paper implements a differentiation strategy for each group of populations through the color population division method to enhance the global search efficiency and convergence accuracy of the algorithm.
Specifically, the blue population follows the OOA framework and retains the original exploration and exploitation strategy, in which the exploration phase encourages a broad search of the solution space, while the exploitation phase focuses on the detailed optimization in the vicinity of the existing better solutions, which embodies the core idea of the traditional bio-inspired algorithm.
The red population, on the other hand, has made strategic innovations in the framework of OOA, especially in the exploitation phase, where the HHO is introduced. This strategy demonstrates the optimization and extension of the traditional algorithm by dynamically adjusting the exploration behavior and development behavior of individuals based on the current iterative progress and stochastic factors, which effectively facilitates the escape from the local optimum.
The strategy of the green population achieves an enhancement of the flexibility of the search strategy by dynamically adjusting the intensity of the random walks while maintaining the iterative exploration. Specifically, depending on the different stages of the algorithm operation, the green population is able to flexibly switch between global exploration and local enhancement, which enhances the efficient coverage of the solution space by exploiting the time-dependent neighborhood search radius variation.
As for the orange population, the optimized spiral search strategy is adopted in the exploration phase, which utilizes the time-decay parameter and randomized spiral motion to not only broaden the search path but also increase the randomness of the exploration process. As for the exploitation stage, the orange population borrows the firefly perturbation strategy, which achieves the fine adjustment and optimization of the local optimal solution by simulating the attraction and repulsion behaviors between fireflies, demonstrating a high degree of strategy innovation and optimization ability.
In summary, by implementing strategy customization for each color population separately, IOOA not only embodies the diversity and flexibility of the algorithm design but also effectively improves the global exploration efficiency and local convergence accuracy of the algorithm in solving complex optimization problems by integrating a variety of bio-inspired strategies.
The mathematical expressions and search logic of these various population adoption strategies are explained in detail below.
Harris Hawk-inspired strategy:
The strategy of the red population employs the escape energy mechanism in the Harris Hawk optimization algorithm [26] in the second phase to facilitate further optimization of the solution, especially in solving the local optimum trap problem. The core of the strategy in this phase is to simulate the collaborative hunting behaviors exhibited by Harris Hawks in nature and their dynamic flight strategies when escaping from predators as a way to enhance the algorithm’s exploratory capability and ability to jump out of local optima.
The formula for the energy decay factor E1 is shown below:
E 1 = 2 1 t T
denotes the gradual decay of the Harris Hawk’s “escape energy” as the number of iterations, t, increases, simulating a gradual shift from extensive exploration to more focused development of the algorithm over time.
The escape energy E formula is shown below:
E = E 1 2 r 1
where r is a random number between [0, 1]. The escape energy E determines the strength of the escape, which can be either positive (exploration) or negative (exploitation), enhancing the randomness and flexibility of the strategy.
Depending on the random number q between the escape energy E and [0, 1], the Harris Hawk’s attack is determined, one encouraging exploratory behavior away from the current solution and the other exploitative behavior with more fine-tuning toward the globally optimal solution.
When the escape energy E 1 1 , the Harris Hawk performs the exploration behavior, and the update rule formula is shown below:
The nonlinear inertia weights ω 2 are calculated as shown below:
X n e w = X c u r r e n t E X r a n d o m 2 r · X c u r r e n t , q < 0.5 X B e s t X p o p u l a t i o n ¯ E X B e s t X c u r r e n t , q 0.5
When the escape energy E 1 < 1 , the Harris Hawk performs the exploitation behavior, updating the rule equation as shown below:
X n e w = X c u r r e n t + E X B e s t X c u r r e n t
The second-stage strategy of the red population achieves a smooth transition from extensive exploration to local fine exploitation by incorporating the biobehavioral characteristics of the Harris Hawk, especially in dealing with the local optimum problem, which demonstrates a unique strategic advantage. Through the dynamically adjusted escape energy and flexible exploration and exploitation mechanism, the algorithm is able to effectively jump out of the local optimum trap and continuously explore and optimize the solution space, thus improving the overall optimization performance.
Random walk strategy:
The strategy design of green population aims to balance the exploration and exploitation capabilities of the algorithm through two phases—the global exploration phase and the local exploitation phase—in order to efficiently search the solution space and eventually converge to a high-quality solution during the optimization process, which is exactly what the random walk strategy [26] can satisfy. The following is a detailed analysis of these two phases of the strategy and the interpretation of the relevant mathematical formulas:
The formulas for the global exploration phase are described below: In the early stage of IOOA, the search range is wide, and an exponential decay function is used to control the rapid narrowing of the exploration range with the increase in the number of iterations t, encouraging a fast coverage of the solution space, the exponential decay function whose formula is shown below:
N F ( t ) = 0.2 e t T × 0.3
In IOOA, a random vector is used in this paper for randomly selecting the search direction to increase the randomness and diversity of the search, and the random perturbation function whose formula is shown below:
M D = 2 U 0,1 0.5
where U (0,1) represents a uniformly distributed random variable.
The position update formula for the global exploration phase is as follows:
X n e w g l o b a l = X B e s t + N F ( t ) · M D
When t M × 0.3 , the algorithm starts to enter a localized development phase based on the increase in the number of iterations t. As the iterations go deeper, the perturbations get smaller and smaller, which helps make fine adjustments rather than large jumps. The formulas for the local development phase are introduced as follows:
X n e w = X B e s t + u b l b r a n d 1 , d i m t 0.3 M
This formulation focuses on optimizing in the vicinity of the discovered better solutions by reducing the magnitude of the perturbations. The strategy of green populations combines the advantages of global and local search by dynamically adjusting the exploration range and the magnitude of the perturbations, which both achieves a broad exploration of the solution space and ensures efficient local optimization after discovering the potentially optimal region.

3.3. Optimize Spiral Search Strategy

The strategy of orange population is designed around two core phases, the exploration phase and the exploitation phase, each of which employs a unique bio-heuristic strategy to balance the exploration and exploitation capabilities of the algorithm, aiming to improve the quality of search efficiency and solution. The search strategy [27] used in the first phase of the orange population is an optimized spiral search strategy, which extensively explores the solution space by simulating a spiral motion with the aim of jumping out of the local optimum and discovering new solution regions.
The formulation of the global exploration phase with respect to the orange population is presented as follows:
α = 2 ( 1 t T )
α is a linear decay parameter that decreases with the number of iterations t, reflecting the gradual reduction in the exploration range over time.
l = 2 π 1 t T · r a n d
l is the spiral angle, which combines the current iteration ratio t T and a random number to introduce a randomization factor to increase the diversity of the search. Rand represents a random number between 0 and 1.
To further increase the randomness of the search, a deflation factor A is introduced by combining the linear decay parameter α and another random number with the following equation:
A = 2 α · r a n d
Combining spiral motion and random perturbations to form a new positional candidate solution. The formula is as follows:
X n e w = X O r a n g e + A X B e s t X O r a n g e e 2 l cos l + 0.1 u b l b r a n d 1 , d i m 0.5

3.4. Firefly Spoiler Strategy

Based on the exploration phase of the orange population, local optimization is performed to further improve the quality of the solution by simulating firefly interactions. The normalization of distance, an important part of the firefly perturbation algorithm [28], calculates the normalized distance r i , j between the current individual X O r a n g e and the hypothetical optimal solution X B e s t for different sizes of the solution space as follows:
r i , j = n o r m X O r a n g e X B e s t d m a x
where d m a x = u b l b d i m .
Based on the exponential decay of the distance r i , j , an attraction coefficient β is introduced to control the strength of the inter-individual interaction with the following equation:
β = β 0 e γ · r i , j m
where β 0 = 2 ,   γ = 1 ,   m = 2 are preset parameters.
Its position update formula for the development phase is as follows:
X n e w = X O r a n g e + β · X O r a n g e X B e s t + α r a n d 1 , d i m 0.5
In summary, the strategy of orange populations achieves a wide coverage of the search space in the exploration phase by optimizing the spiral search strategy and then uses the firefly perturbation strategy to finely optimize the potential high-quality solution regions in the exploitation phase, and this combination of strategies effectively balances the algorithm’s global exploration capability and local exploitation capability.

3.5. OOA and IOOA Basic Process

The pseudo-code of the OOA is shown in Algorithm 1 and Figure 4.
Algorithm 1. OOA steps can be summarized as follows.
Input parameters: Osprey population size N and maximum number of iterations i t e r m a x .
Output parameters: best position X P o s , best fitness value X B e s t
(1) Initialization phase
Initial population of fish eagles initialized using Equation (1).
(2) Algorithm flow:
The iteration begins:
while (t < i t e r m a x )
  Initialize the array of fitness values.
  for i = 1: N
Phase 1: Position identification and hunting the fish
    Update fish positions set for the ith OOA member using Equation (2).
    Determine the selected fish by the ith osprey at random.
    Calculate new position of the ith OOA member based on the first phase of OOA using Equation (3).
    Check the boundary conditions for the new position of OOA members using Equation (4).
    Update the ith OOA member using Equation (5).
Phase 2: Carrying the fish to the suitable position
    Calculate new position of the ith OOA member based on the second phase of OOA using Equation (6).
    Check the boundary conditions for the new position of OOA members using Equation (7).
    Update the ith OOA member using Equation (8).
  end
  for i = 1: N
    Get the current new location;
    If the new location is better than before, update it;
  End for
  t = t + l;
  Get the current new location;
  If the new location is better than before, update it;
End while
IOOA is evolved from OOA, which enhances the performance of the algorithm through a series of strategic adjustments. First, in the initial stage, logistic chaos mapping and the good point set method are used to generate the initial population, and the population is divided into four different color groups according to the four-color theory as a way to improve the diversity of the population and balance the ability of global search and local exploitation. Then, a two-color complementary mechanism is introduced into the algorithm design, and a differentiated strategy is adopted for the populations of different color groups. For the blue population, the focus is on maintaining the original OOA exploration strategy to ensure the basic performance and stability of the algorithm. For the red population, a heuristic strategy based on the behavior of the Harris Hawk is introduced, which mimics the efficient search and localization of the Harris Hawk and improves the algorithm’s ability to avoid falling into local optimal solutions. The green population employs a strolling wandering strategy, which extends the search range by mimicking random wandering behavior while maintaining population diversity. Finally, the orange population adopts an optimized spiral search strategy to explore the solution space in depth and employs the firefly perturbation strategy to effectively perturb the local optimal solution, which facilitates the algorithm to jump out of the local optimum and continue to explore. The combination of these strategies enables IOOA to further improve the global search capability and the ability to avoid local optimums by inheriting the advantages of the original OOA. The flow charts of OOA and IOOA are shown in Figure 4 and Figure 5.
The pseudo-code of the OOA is shown in Algorithm 2.
Algorithm 2. IOOA steps can be summarized as follows.
Input parameters: Osprey population size N, number of populations in the red group N R e d , number of populations in the green group N G r e e n , number of populations in the blue group N B l u e , number of populations in the orange group N O r a n g e , and maximum number of iterations i t e r m a x .
Output parameters: best position X P o s , best fitness value X B e s t
(1) Initialization phase
The initial population of the fish eagle was initialized with the L-good point set using Equations (11) and (12), and then its initialized population was randomly grouped, and the red group of populations, the green group of populations, the blue group of populations, and the orange group of populations were selected according to Equations (13)–(16).
(2) Algorithm flow:
The iteration begins:
while (t < i t e r m a x )
  Initialize the array of fitness values.
  The exchange probability is calculated dynamically for each generation. According to equation ( p t = max p 0 × ( 1 δ ) ^ t , p m i n
  Setting the minimum exchange probability threshold to prevent the probability from being too small.
  If p t p m i n
     p t = p m i n
  end
  Blue stock location update
  for i = 1: N B l u e
      Using Equation (5) update the blue population’s location;
      Using Equation (8) update the blue population’s location;
  end
   R e d stock location update
  for i = 1: N R e d
      Using Equation (5) update the blue population’s location;
      Using Equations (18) and (19) update the blue population’s location;
  end
   G r e e n stock location update
  for i = 1: N G r e e n
      Using Equation (23) update the blue population’s location;
      Using Equation (8) update the blue population’s location;
  end
   O r a n g e stock location update
  for i = 1: N O r a n g e
      Using Equation (27) update the blue population’s location;
      Using Equation (30) update the blue population’s location;
  end
  for i = 1: N
    Get the current new location;
    If the new location is better than before, update it;
  End for
  t = t + l;
End while

3.6. Complexity Analysis

In the standard OOA, let the osprey population size be N and the dimensionality of the solution space be D. OOA performs parameter initialization with a time complexity of O(N × D), individual fitness with O(N), and population complexity with O(N × D). Since these operations are performed for each individual of each population and involve fitness function calls (assuming O(D) time complexity for fobj), the total time complexity is O(M × N × D). So, the overall complexity of OOA is as follows:
O ( O O A ) = O ( N × D ) + O ( N ) + O ( M × N × D ) = O ( M × N × D )
The IOOA algorithm first generates an N × D random matrix using a logistic mapping in the initialization phase, and next, two loops are performed: an outer loop i from 1 to N and an inner loop j from 2 to D. The logistic mapping is computed each time the inner loop is performed, which is a constant time operation. Therefore, the time complexity of this part is O(N × D).
The good point set strategy mainly uses a prime-based operation, which has a time complexity of O(N × D) because it involves computing an expression for each dimension and copying it to all individuals, which is O(N × D) overall.
Combined with the above analysis, the most time-consuming parts of the initialization process are the computation of the logistic mapping, the combination operation with the good point set strategy, and the generation of random permutations. Among them, the generation of random permutations is O(N × D), and all other major steps are O(N × D). Therefore, considering the worst-case dominant term, the time complexity of the whole initialization function is roughly O(N × D). During the operation of IOOA, there is mainly the exchange probability calculation, which is a constant time complexity O(1). The individual exchange part between populations is performed separately for red and green populations and blue and orange populations. The time complexity of the exchange operation for each color population is O(N × P t ), but the actual number of operations decreases as P t decays over time. Overall, the time complexity of this part is close to O ( N ) if the average exchange probability is considered. The time complexity calculations for each color group population separately are O ( N R e d × D), O ( N G r e e n × D ), O ( N B l u e × D ), and O ( N O r a n g e × D ), respectively. Different update rules are applied to each population, including the calculation of new positions and fitness values. Since these operations are performed for each individual of each population and involve fitness function calls (assuming O(D) time complexity for fobj and O(M × ( N R e d + N G r e e n + N B l u e + N O r a n g e ) × D) for the total time complexity). Recording the optimal adaptation per generation is a typical linear operation with a time complexity of O(M). The total time complexity of this IOOA is as follows:
O I O O A = O N × D + O 1 + O N + O M × N R e d + N G r e e n + N B l u e + N O r a n g e × D = O M × N × D
In summary, IOOA does not show a significant increase in time complexity compared to OOA.

4. Experimental Simulation and Result Analysis

In this chapter, a simulation study will be conducted to evaluate the effectiveness of IOOA in optimization. The experiments were conducted on MATLAB R2021b with an 11th-generation Intel Core i7 processor, 16 GB of dual-channel RAM, and a 512 GB SSD. The graphics card is a GeForce RTX 3050Ti.

4.1. Selection of Benchmark Function and Experimental Setup

This paper evaluates the performance of the IOOA algorithm in handling various objective functions by using 30 test functions, including 8 benchmark functions, the CEC2020 test set, and the CEC2022 test set [29]. The study compares IOOA with ROA, SFO, CDO, BA, GSA, WWPA, OOA, and two improved SOOA and ADSOOA to assess the quality of the best solution provided by IOOA. The control parameters of all the algorithms are set according to the recommendations of the algorithm proposers and are described in detail in Table 1.
The above is a selection of the static parameters of the chosen algorithm, and the following will focus on the three main dynamic parameters in IOOA: P, a, and A.
P is a parameter in the exploration phase used to control the probability of individual exchange between populations. The initial value of P is set to 0.5, and this parameter decreases over time, allowing the algorithm to move from extensive exploration in the early stages to refined exploitation in the later stages. As the number of iterations proceeds to a certain level, the value of P will be maintained to minimize the exchange probability, and in this paper, the minimum exchange probability threshold pmin is set to 0.2. The variation in P with the number of iterations is shown in Figure 6.
A is the deflation factor in the black population, which is used to control the search radius in the spiral search strategy. a is a parameter that decays linearly with the number of iterations, which can control the search radius of A well with the increase in the number of iterations, so A will decrease linearly with the increase in the number of iterations, which will gradually narrow down the search range and improve the development of the algorithm. Refer to Equations (25) and (27) for details. At the beginning of the iteration, A is larger, which is favorable for the algorithm to conduct a comprehensive search of the problem space; as the number of iterations increases, A gradually decreases, which helps the algorithm to transition from the exploration stage to the development stage, and thus more accurately approximate the global optimal solution. The variation in A and a with the number of iterations is shown in Figure 7.
In the experimental design, the performance test of the algorithm is gradually transitioned from the benchmark test set to the CEC2020 test set and then to the CEC2022 test set. Its details are shown in Table 2, Table 3 and Table 4 in turn. This gradual transition aims to evaluate the performance of the algorithms on different test sets in depth and to provide a more accurate basis for performance evaluation. The number of populations N, the maximum number of iterations T, and the dimension Dim of each algorithm on different test sets in the experiment are shown in Table 5. To reduce the chance of the experiment and increase the persuasiveness of the experimental results, each algorithm is run independently 10 times on each benchmark test function. To analyze the performance of the algorithms, convergence analysis, stability analysis, and nonparametric tests are performed. Convergence analysis includes analyzing performance using convergence graphs and convergence tables, where the convergence table contains four evaluation functions such as best, average, standard deviation (Std), and number of iterations. These analyses help to understand the convergence speed and stability of each algorithm and its performance in solving problems. Stability analysis was performed using the ANOVA test [30] for variance and box line plots to indicate the stability of the algorithms. Nonparametric tests such as the Wilcoxon signed rank test [31] and the Friedman rank sum test [32] were also used to delve into the statistical significance of the differences in the performance of the algorithms on different problem instances.
The experiments and analyses in this study aim to comprehensively evaluate the performance of the IOOA algorithm in dealing with various objective functions and provide strong comparative and statistical evidence for further research and applications.

4.2. Convergence Analysis

4.2.1. Convergence Analysis of IOOA on Benchmark Test Set

From the data in Table 6, it can be observed that the IOOA shows excellent performance on several test functions. Specifically, on the test functions F1 to F3, F5, F6, and F8, the minimum, mean, and standard deviation obtained by the IOOA algorithm are the smallest among all the algorithms, which fully demonstrates that its optimization on these test functions is the best. Meanwhile, in the F3 test function, apart from the IOOA algorithm, the ROA, GSA, WWPA, OOA, SOOA, and ADSOOA algorithms also found optimal values. Although in the F4 test function, the IOOA algorithm does not have as good a minimum, mean, and standard deviation as the ADSOOA algorithm, the answers found by the IOOA algorithm are extremely close to the results of the ADSOOA algorithm, which indicates that both algorithms perform quite well on this test function. In the F7 test function, although the mean and standard deviation of the IOOA algorithm are not the smallest among all the algorithms, the minimum value found by it is still the smallest among all the algorithms, which also highlights the significant advantage of the IOOA algorithm in finding the optimal solution. In summary, the IOOA algorithm shows excellent optimization performance on most of the tested functions.
It can be clearly observed from Figure 8 that IOOA shows excellent convergence accuracy and accelerated convergence in all test functions. Especially on the test functions F1, F2, and F6 to F8, IOOA has a clear advantage in convergence accuracy compared to other algorithms. In addition, on the three test functions from F3 to F5, although the convergence accuracies achieved by multiple algorithms do not differ much, IOOA also shows good performance. Especially noteworthy is that in the F3 test function, IOOA’s convergence speed is significantly faster than the other algorithms, which further reflects its optimization ability in this test function. In summary, IOOA shows excellent convergence performance and accuracy on several test functions.

4.2.2. Convergence Analysis of IOOA on CEC2020 Test Set

From the data in Table 7, it can be observed that the IOOA algorithm exhibits excellent performance on several test functions. Specifically, on four test functions, F9, F12, F16, and F18, IOOA obtained the smallest minimum, mean, and standard deviation, indicating that it optimized best on these four functions. For the three test functions F10, F11, F14, and F15, although the standard deviation of IOOA is not the smallest, its minimum and mean values are at the optimal level, which further proves the optimization ability of IOOA. In addition, when comparing IOOA with CDO in test function F13 and IOOA with ADSOOA in test function F17, although the mean and standard deviation of IOOA are not the smallest, the minimum value it finds is still the smallest among all the algorithms, which likewise demonstrates the advantage of IOOA in finding the optimal solution. In summary, the IOOA algorithm shows excellent optimization performance on most of the tested functions.
It is clear from Figure 9 that IOOA demonstrates excellent convergence accuracy as well as the ability to accelerate convergence in all the tested functions. In particular, for the functions F9, F10, F12, F13, F15, and F16, the convergence accuracy of IOOA has a clear advantage over other algorithms. In addition, on the test functions F11 and F17, ADSOOA follows IOOA and also shows good performance.

4.2.3. Convergence Analysis of IOOA on CEC2022 Test Set

As can be seen from Table 8, the IOOA algorithm demonstrates excellent performance on several test functions, especially on the four functions F19, F24, F27, and F29, which are particularly optimized. IOOA not only achieves the minimum on these four functions but also outperforms the other algorithms in terms of the mean and stability of the results (measured in terms of standard deviation), which highlights its excellent performance on these tests. For the F20, F26, and F28 test functions, although IOOA is not optimal in terms of stability (standard deviation) of the results, it is able to consistently achieve and maintain the lowest minimum and mean values, which further validates the power and stability of the IOOA algorithm in complex optimization problems. In the F22 and F23 test functions, although IOOA is not optimal in terms of the mean and standard deviation metrics, it successfully mines the smallest minima among all participating algorithms, which again emphasizes the unique advantages of IOOA in deep searching and discovering globally optimal solutions. In summary, the IOOA algorithm generally demonstrated excellent optimization performance on a wide set of test functions, not only achieving the best overall performance on multiple functions but also highlighting its efficiency and accuracy in finding optimal solutions on specific functions. These results fully demonstrate the potential and value of the IOOA algorithm in the field of optimization.
Figure 10 visualizes the excellent performance of IOOA in all the tested functions, not only in its high convergence accuracy but also in its remarkable ability to accelerate convergence. In particular, on the functions F19 to F22, F24, and F29, the convergence accuracy of IOOA far exceeds that of other algorithms, showing a clear advantage. It is worth noting that on the F23 function, although IOOA still takes the lead, ADSOOA follows closely and also performs well, showing its unique optimization ability. However, on these functions from F25 to F27, IOOA’s performance is comparable to that of the other algorithms, indicating that different algorithms have their own adaptability to different types of problems. In summary, IOOA shows excellent optimization performance on most of the tested functions.

4.3. Stability Analysis

A powerful data visualization tool, box-and-line charts excel at demonstrating the distributional properties of continuous data sets, especially concentrated trends, dispersion, and possible extreme data points. The chart is built on top of several key statistics: lower quartile, upper quartile, median, and a body consisting of boxes and whiskers, with occasional outliers labeled. The box encloses 50% of the overall data, bounded by the lower and upper quartiles, while the median falls inside the box, reflecting the center of the data. The whiskers, on the other hand, extend outward to the regular boundaries of the data and are generally defined within 1.5 times the expansion of the interquartile spacing, with points beyond this boundary considered outliers.
Box-and-line plots are particularly important when evaluating algorithm performance, as they visualize the similarities and differences in how algorithms perform when faced with the same task or in different contexts. The compactness of the box directly reflects the concentration of data and the consistency of the algorithm’s results, while lower box positions imply the algorithm’s ability to find high-quality solutions. Comparing the boxplots of multiple algorithms, algorithms with more robust performance on specific tasks, higher solution accuracy, and lower sensitivity to external interference can be quickly identified, providing a strong visual basis for algorithm selection and tuning.

4.3.1. IOOA Stability Analysis on Benchmark Test Sets

As can be observed in Figure 11, IOOA shows its excellent performance stability and advantages through box-and-line plots on this test function. Specifically, for the test functions F6 and F8, IOOA’s boxplot shows a lower median value, which indicates that its performance is relatively better on these two functions. In addition, for the test functions F3, F4, and F5, the boxplots of IOOA show excellent stability, indicating that its performance fluctuates less on these functions. On the other hand, for the test functions F1, F2, and F7, IOOA is able to find even smaller values, further demonstrating its ability to optimize these functions. In summary, IOOA shows excellent performance and stability on several test functions.

4.3.2. IOOA Stability Analysis on CEC2020 Test Sets

As can be observed from Figure 12, IOOA shows excellent performance stability and benefits through box-and-line plots for the set of test functions of CEC2020. For most of the test functions, the boxplots of IOOA show lower median values, especially more prominent for the functions F9, F10, F12, F14, F16, and F18. In addition to this, IOOA shows excellent stability for the functions F9, F12, F16, and F18. This intuitively shows that the algorithm is able to produce the smallest variance during the optimization of these types of functions, which means that it is able to provide the most stable solutions.

4.3.3. IOOA Stability Analysis on CEC2022 Test Sets

In the presentation in Figure 13, IOOA’s performance on the CEC2022 test function set is particularly impressive, with extremely significant performance stability and advantages reflected in its box-and-line plots. Specifically, IOOA maintains low median values on most of the test functions, and this advantage is particularly strong for functions F19 through F24, F26, and F30, showing the algorithm’s superior performance on these functions. For the functions F19, F20, F26, and F30, IOOA not only performs well but also shows great stability. The extremely small variance range in the boxplots visually demonstrates this, implying that when optimizing these functions, IOOA is able to generate highly consistent found optimal solutions, reducing the volatility and uncertainty of the results. In summary, IOOA’s performance on the CEC2022 test function set proves its excellent performance stability and significant advantages on specific function types, providing a powerful tool for solving complex optimization problems.

4.4. Nonparametric Test

Nonparametric tests play an indispensable role in the statistical evaluation of algorithm performance, especially when dealing with situations where the data distribution is unknown or deviates from normality. Wilcoxon rank sum test and Friedman’s test, as two widely recognized nonparametric analytical tools, are particularly useful in the comparative study of algorithm performance, which not only overcame the stringent requirements of the distribution pattern of the data but also deeply explored the statistical significance of performance differences between algorithms.
Specifically, the Wilcoxon signed rank test is good at analyzing paired data or single-sample situations and reveals the significance of the internal differences of algorithms by comparing their performance changes in different conditions or points in time. On the other hand, Friedman’s test, from a broader perspective, focuses on the overall performance distribution of all algorithms under multiple levels (e.g., multiple test functions) and detects whether there is a significant difference between algorithms by calculating the average ranking of each algorithm, which is especially adapted to experimental data of round-robin or paired designs.

4.4.1. IOOA Nonparametric Analysis on Benchmark Test Sets

As can be seen from Table 9, IOOA excels in performance. It significantly outperforms SFO, WWPA, and SOOA on all eight test functions; significantly outperforms CDO, BA, and ADSOOA on seven test functions when comparing with CDO, BA, and ADSOOA; significantly outperforms ROA and GSA on six test functions when comparing with ROA and GSA, but there is no significant difference in F3 and F7; and significantly outperforms OOA on half of the test functions when comparing with OOA and has no significant difference with OOA on the other half of the no significant difference with OOA on the other half of the test functions. In summary, IOOA demonstrates excellent performance on several test functions and shows significant advantages over multiple algorithms.
Figure 14 shows that IOOA has the highest ranking on this test function, indicating the best performance followed by GSA. OOA and ADSOOA are third and fourth. ROA, CDO, BA, SOOA, SFO, and WWPA, the average scores of these algorithms are in increasing order. This indicates that IOOA shows a significant advantage in this test function.

4.4.2. IOOA Nonparametric Analysis on CEC2020 Test Sets

As can be seen from Table 10, IOOA is significantly better than ROA on 9 test functions and is not significantly different from ROA on F15.IOOA is significantly better than SFO, BA, GSA, WWPA, and OOA on all 10 test functions. IOOA is significantly better than CDO on 9 test functions and is not significantly different from CDO on F13.IOOA is significantly better than BA on 9 test functions and is not significantly different from CDO on F13, significantly better than BA, and is not significantly different from BA on F11. IOOA is significantly better than SOOA on 9 test functions and is not significantly different from SOOA on F15. IOOA is significantly better than ADSOOA on 5 test functions and is not significantly different from ADSOOA on F5. In summary, IOOA significantly outperforms all other algorithms except ADSOOA on most of the test functions.
Figure 15 shows that IOOA is the top performer in all functions. IOOA and ADSOOA are the best performers, ranked first and second, respectively. ROA, SFO, GSA, and CDO are close to each other and ranked in the middle of the list. BA and WWPA are worse performers, ranked ninth and tenth, respectively.

4.4.3. IOOA Nonparametric Analysis on CEC2022 Test Sets

As can be seen from Table 11, IOOA significantly differs in performance compared to the other algorithms in most cases. As can be seen from the +/=/− columns, IOOA significantly outperforms the other algorithms in the vast majority of comparisons and is only significantly inferior to the other algorithms in very few cases. In particular, in the comparisons with SFO, GSA, and SOOA, IOOA significantly outperforms these algorithms on all functions (F19 through F30). For some functions (e.g., F21), there is no significant difference in performance between IOOA and some algorithms (e.g., ROA), and it is even significantly inferior to CDO in some cases (vs. CDO), which may indicate a limitation of IOOA for some specific problems or function types. For other functions (e.g., F25), there is no significant difference in performance between IOOA and multiple algorithms (e.g., ROA, OOA, ADSOOA), which may be related to the complexity of the problem or the characteristics of the algorithm. When compared to BA, on F22 and F24, there is no significant difference, and it is inferior to BA on F30. In conclusion, IOOA shows a significant performance advantage in most of the comparisons with multiple algorithms on CEC2022.
Figure 16 clearly demonstrates the excellent performance of IOOA on all the tested functions, placing it firmly in the first place and highlighting its best performance, followed by ADSOOA, which performs well, and BA, ROA, CDO, and GSA, which perform equally well and are located in the middle of the ranking. In contrast, SOOA, OOA, SFO, and WWPA are gradually moving down the rankings. This result fully proves that IOOA has a significant performance advantage on the current set of test functions.

5. Engineering Applications of the IOOA

In this chapter, the potential of IOOA in solving typical engineering design challenges is explored, and its powerful optimization capabilities and practical benefits are demonstrated through three examples. These three cases are the design of tension and compression springs, the optimization of a welded beam problem, and the design and optimization of a gear transmission system, all of which are common and complex problems in the engineering field and place high demands on the performance of the algorithms. The breadth and depth of the application of IOOA in real engineering problems are comprehensively examined. First, in the design of tension and compression springs, faced with the comprehensive consideration of material utilization, strength, and elasticity properties, IOOA successfully finds the optimal design solution to minimize the spring mass under the given constraints. Secondly, for the welded beam problem, IOOA demonstrated its excellent ability to optimize the structural parameters under multiple constraints, effectively reducing the manufacturing cost while ensuring the safety and reliability of the structure. Finally, in gear train optimization, IOOA further proves its efficiency and accuracy in improving the efficiency of mechanical systems by minimizing the transmission ratio.

5.1. Extension/Compression Spring Design

When performing an optimized configuration of an extension/compression spring, the core design elements involve the spring wire diameter d (denoted as ( x 1 )), the average diameter of the spring coils (D) (labeled ( x 2 )) and the number of coils (P) (denoted ( x 3 )). The design aims to minimize the overall mass of the spring while ensuring that key performance indicators such as deflection, shear stress, and vibration frequency meet preset criteria. The structural layout of the spring is shown graphically in Figure 17.
The problem of tension spring design is modeled as the following mathematical optimization formulation:
m i n f x = x 1 2 x 2 x 3 + 2 x 1 2 x 2
s . t . g 1 ( x ) = 1 x 2 3 x 3 71785 x 2 4 0 g 2 ( x ) = 4 x 2 2 x 1 x 2 12566 ( x 1 3 x 2 x 1 4 ) + 1 5108 x 1 2 1 0 g 3 ( x ) = 1 140.45 x 1 x 2 2 x 3 0 g 4 ( x ) = x 1 + x 2 1.5 1 0 0.05 x 1 2 ,      0.25 x 2 1.3 2 x 3 15
The IOOA algorithm is employed to address the above spring design challenges, and its performance is compared with seven other algorithms. The comparison results are summarized in Table 12 and Figure 18, which clearly show that the IOOA algorithm is finding the best set of design parameters ( x 1 , x 2 , x 3 ) . and the lowest quality fmin(X) with a specific value of 1.27 × 10−02. This remarkable result validates the efficiency and leading edge of the IOOA algorithm in solving the design challenges of tension springs.

5.2. Optimized Design Problems for Welded Beams

The optimal design challenge for welded beams is a complex problem that combines structural mechanics and mathematical optimization, aiming to ensure that a set of engineering constraints are satisfied by tuning the design parameters X = (h, l, t, b), i.e., the height, length, thickness, and width of the beams while ensuring that the premise of minimizing the manufacturing cost. Specifically, this process considers the shear stress τ, the bending stress σ, the bending load Pc imposed on the beam, the maximum deflection δ at the end of the beam, and the geometric boundary constraints, all of which need to be controlled to stay within the safety margins. The model is shown in Figure 19.
The design objective function is expressed as minimizing cost:
m i n f x = 1.10471 x 1 2 x 2 + 0.04811 x 3 x 4 14.0 + x 2
s . t . g 1 X = τ x τ m a x 0 g 2 X = σ x σ m a x 0 g 3 X = δ x δ m a x 0 g 4 X = x 1 x 4 0 g 5 X = P P c x 0 g 6 X = 0.125 x 1 0 g 7 X = 1.10471 x 1 2 + 0.04811 x 3 x 4 14.0 + x 2 5.0
where
τ x = τ 2 + 2 τ τ x 2 2 R + τ 2 τ = p 2 x 1 x 2 τ = M R J M = P L + x 2 2 R = x 2 2 4 + x 1 + x 3 2 2 J = 2 2 x 1 x 2 x 2 2 12 + x 1 + x 3 2 2 σ x = 6 P L x 4 x 3 2 δ x = 4 P L 3 E x 3 3 x 4 P c x = 4.013 E x 3 2 x 4 6 36 L 2 1 x 3 2 L E 4 G P = 6000   l b L = 14   i n δ m a x = 0.25   i n E = 30 × 10 6   p s i G = 12 × 10 6   p s i τ m a x = 13600   p s i δ m a x = 30000   p s i
With bounds,
0 x 1 , x 4 2 ,
0.1 x 2 , x 3 10 .
To explore the solution of this optimization problem, several optimization algorithms were used for comparative evaluation. Figure 20 and Table 13 show that the IOOA algorithm stands out in this problem, and the optimal combination of design variables it finds achieves the minimization of the objective function value of 1.71 × 10+00, which is a more superior solution compared with the other algorithms, indicating that IOOA has higher efficiency and accuracy in solving this kind of complex structural optimization problems. This finding is not only of direct guidance to the engineering practice of welded beams but also provides a valuable reference for the application of optimization algorithms in the field of structural design.

5.3. Gear Train Optimization Design

As a core component of mechanical engineering, the design optimization of gearing mechanisms is essential to improve equipment performance, reduce weight and size, and thus effectively control operating costs. This design challenge focuses on the construction of a precision transmission system with four gears, aiming to enhance the overall system efficiency by minimizing the transmission ratios, as shown in the simplified model in Figure 21.
The transmission ratio (gr) is quantified in this scenario as follows:
m i n f x = ( 1 6.931 T D T B T A T F ) 2
where TA, TB, TD, and TF represent the number of teeth of the four gears, respectively, and the range of their values is bounded:
s . t . 12 T A 60 12 T B 60 12 T D 60 12 T F 60
To explore the solution of this optimization problem, a series of optimization algorithms are employed, and their performances are compared. The results of the fitness iteration curves for each algorithm are shown in Figure 22, which shows that IOOA has a faster convergence speed and higher convergence accuracy than the other algorithms. Table 14 summarizes the optimal design variables and their corresponding minimum values of the objective function fmin(X) obtained by each algorithm, and the results show that IOOA stands out with a significant advantage, with its computed objective function value as low as (2.70 × 10−12), which reveals that in the pursuit of a very small transmission ratio to improve the system efficiency, the IOOA algorithm is able to most accurately approximate the ideal solution, outperforming other algorithms such as ROA, SFO, CDO, BA, GSA, WWPA, OOA, SOOA, and ADSOOA. This finding not only emphasizes the efficiency and accuracy of the IOOA algorithm in solving complex gearing design problems but also provides a powerful tool for finding better gearing solutions in the field of mechanical design, which further promotes the development of high-performance gearing systems.

6. IOOA Performance Analysis and Conclusions

6.1. IOOA Performance Analysis

IOOA demonstrates excellent performance on several benchmark functions, especially showing fast convergence on the general test functions, CEC2020 and CEC2022 test sets. Specifically, IOOA significantly outperforms other algorithms in terms of convergence speed and accuracy on functions F1, F2, F6 to F8, F9, F12, F16, F18, F19, F24, F27, and F29, while on functions F3, F10, F11, F14, F15, F13, and F17, it maintains high convergence, despite convergence speeds that are similar to those of other algorithms’ accuracy. Statistical analysis of multiple independent runs of several test functions shows that IOOA performs well in terms of stability, with its minimum, mean, and standard deviation remaining low on most of the test functions, especially on functions such as F6, F8, F19, F26, and F30, indicating that it maintains good consistency of results while maintaining efficient optimization search. Boxplot analysis further confirms that IOOA shows stability with low median values and narrow distribution on most of the tested functions. By introducing the two-color complementary mechanism, IOOA effectively improves the balance in the search process and enables the algorithm to achieve a better balance between global search and local exploitation, which speeds up the convergence and improves the efficiency of resource utilization. Especially when dealing with test functions with high complexity, such as some functions in CEC2020 and CEC2022, IOOA is able to find near-optimal solutions in a shorter period of time, which reflects high efficiency. IOOA is able to find the optimal solution or near-optimal solutions on several test functions, especially on the functions F19, F24, F27, and F29, with the minimum value, mean, and standard deviation better than other algorithms, indicating its strong effectiveness in dealing with complex optimization problems. Through the Wilcoxon signed rank test and the Friedman test, IOOA shows significant performance advantages over other algorithms on most of the tested functions, which further validates its effectiveness. By introducing different policy adjustment mechanisms, IOOA ensures that the algorithm can accurately locate the global optimal solution or a near-optimal solution, especially on the CEC2020 and CEC2022 test sets. IOOA can find the minimum or near-minimum solution on most test functions, which fully proves its accuracy.
In summary, IOOA is able to quickly converge to the optimal solution or near-optimal solution on most of the test functions, demonstrating excellent convergence performance and, at the same time, demonstrating a high degree of stability and the ability to maintain consistent result quality. On the test functions with higher complexity, IOOA performs well and can effectively solve practical engineering problems. Although IOOA performs well on most of the test functions, it slightly underperforms other algorithms on some specific types of functions (e.g., F4, F11, F22, and F23), which may be due to the special nature of these functions. The convergence speed of IOOA may suffer when dealing with certain complex problems that are highly nonlinear or have multiple local optimal solutions. Overall, IOOA demonstrates strong capabilities in solving complex optimization problems, especially in terms of convergence speed, stability, and accuracy.

6.2. Conclusions

In this paper, in order to solve the shortcomings of OOA, that is, the search accuracy is not enough, and it is easy to fall into the local optimal solution, an innovative improvement strategy is proposed, and IOOA based on the two-color complementary mechanism is designed and implemented. By drawing on the four-color theory and multi-strategy fusion, the algorithm has achieved more superior performance in the field of global optimization. In the initialization stage of the algorithm, the good point set strategy of logistic chaotic mapping is used to construct an initial population with high diversity, and the four-color classification method is introduced to reconstruct the population, which effectively improves the global search ability of the algorithm. Subsequently, the core lies in the proposed two-color complementary mechanism, which distinguishes the roles and functions of the population through refinement. The blue and red populations, respectively, assume the roles of robust exploration and local obstacle avoidance, while the green and orange populations further enhance the randomness of the algorithm and the ability to effectively escape from the local optimal solution by introducing the following strategies: roaming and optimal spiral search. The embedding of the firefly perturbation strategy provides a new idea for solving local optimization problems. In the experimental part, this paper selects a series of classic benchmark test functions and the standard test sets of CEC2020 and CEC2022 to systematically evaluate the performance of IOOA and compare it in detail with various existing state-of-the-art optimization algorithms. Through rigorous verification of statistical tests (ANOVA, Wilcoxon test, and Friedman test), it was confirmed that IOOA has significantly improved convergence accuracy and convergence speed. At the same time, IOOA was applied to three engineering optimization design problems, indicating that the algorithm has a wider application potential and competitive advantage in solving complex optimization problems. In future research, we can further explore the integration of more diverse bio-inspired strategies with IOOA in order to achieve a deeper optimization effect in optimization problems in specific fields.

Author Contributions

Conceptualization, F.W. and X.S.; methodology, F.W.; software, X.S.; validation, F.W., X.S. and Y.F.; formal analysis, X.S.; investigation, Y.F.; resources, F.W.; data curation, F.W.; writing—original draft preparation, Y.F.; writing—review and editing, X.S.; visualization, X.S.; supervision, F.W.; project administration, F.W.; funding acquisition, F.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Shaanxi Province Natural Science Basic Research Program (Grant No. 2024JC-YBMS-329) and the Science and Technology Plan Project of Yulin City (Grant No. 2023-CXY-206).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

This paper does not contain any studies with human participants or animals performed by any of the authors.

Data Availability Statement

The data used in the study can be accessed upon your request to the author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shadhik, M.; Senthil Kumaran, M. Meta Heuristic Optimization Algorithm Based Rectifier Control in Switched Reluctance Generator. IETE J. Res. 2023, 1–23. [Google Scholar] [CrossRef]
  2. Fogel Lawrence, J.; Owens Alvin, J.; Walsh Michael, J. Artificial Intelligence through Simulated Evolution; Wiley-IEEE Press: Hoboken, NJ, USA, 1998; pp. 227–296. [Google Scholar]
  3. Holland, J.H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence; MIT Press: Cambridge, MA, USA, 1992; p. 211. [Google Scholar] [CrossRef]
  4. Storn, R.; Price, K. Differential Evolution–A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  5. Kirkpatrick, S.; Gelatt, C.D.; Vecchi, M.P. Optimization by Simulated Annealing. Science 1983, 220, 671–680. [Google Scholar] [CrossRef] [PubMed]
  6. Tanyildizi, E.; Demir, G. Golden Sine Algorithm: A Novel Math-Inspired Algorithm. Adv. Electr. Comput. Eng. 2017, 35, 71–78. [Google Scholar] [CrossRef]
  7. Shehadeh, H.A. Chernobyl Disaster Optimizer (CDO): A Novel Meta-Heuristic Method for Global Optimization. Neural Comput. Appl. 2023, 35, 10733–10749. [Google Scholar] [CrossRef]
  8. Xue, J.; Shen, B. A Novel Swarm Intelligence Optimization Approach: Sparrow Search Algorithm. Syst. Sci. Control. Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  9. Geem, Z.W.; Kim, J.H.; Loganathan, G.V. A New Heuristic Optimization Algorithm: Harmony Search. Simulation 2001, 76, 60–68. [Google Scholar] [CrossRef]
  10. He, S.; Wu, Q.H.; Saunders, J.R. A Novel Group Search Optimizer Inspired by Animal Behavioural Ecology. In Proceedings of the 2006 IEEE International Conference on Evolutionary Computation, Vancouver, BC, Canada, 16–21 September 2006. [Google Scholar] [CrossRef]
  11. Das, B.; Mukherjee, V.; Das, D. Student Psychology Based Optimization Algorithm: A New Population Based Optimization Algorithm for Solving Optimization Problems. Adv. Eng. Softw. 2020, 146, 102804. [Google Scholar] [CrossRef]
  12. Faridmehr, I.; Nehdi Moncef, L.; Davoudkhani, I.; Poolad, A. Mountaineering Team-Based Optimization: A Novel Human-Based Metaheuristic Algorithm. Mathematics 2023, 11, 1273. [Google Scholar] [CrossRef]
  13. Eberhart, R.; Kennedy, J. A New Optimizer Using Particle Swarm Theory. In Proceedings of the Sixth International Symposium on Micro Machine and Human Science, MHS’95, Nagoya, Japan, 4–6 November 2002. [Google Scholar] [CrossRef]
  14. Dorigo, M.; Maniezzo, V.; Colorni, A. Ant System: Optimization by a Colony of Cooperating Agents. IEEE Trans. Syst. Man Cybern. Part B Cybern. 1996, 26, 29–41. [Google Scholar] [CrossRef]
  15. Yang, X.-S. A New Metaheuristic Bat-Inspired Algorithm. arXiv 2010, arXiv:1004.4170. [Google Scholar]
  16. Reza Naji, H.; Shadravan, S.; Mousa Jafarabadi, H.; Momeni, H. Accelerating Sailfish Optimization Applied to Unconstrained Optimization Problems on Graphical Processing Unit. Engineering Science and Technology. Int. J. 2022, 32, 101077. [Google Scholar] [CrossRef]
  17. Jia, H.; Peng, X.; Lang, C. Remora Optimization Algorithm. Expert Syst. Appl. 2021, 185, 115665. [Google Scholar] [CrossRef]
  18. Abdelhamid, A.A.; Towfek, S.K.; Khodadadi, N.; Alhussan, A.A.; Khafaga, D.S.; Eid, M.M.; Ibrahim, A. Waterwheel Plant Algorithm: A Novel Metaheuristic Optimization Method. Processes 2023, 11, 1502. [Google Scholar] [CrossRef]
  19. Dehghani, M.; Trojovský, P. Osprey Optimization Algorithm: A New Bio-Inspired Metaheuristic Algorithm for Solving Engineering Optimization Problems. Front. Mech. Eng. 2023, 8, 1126450. [Google Scholar] [CrossRef]
  20. Yuan, Y.; Yang, Q.; Ren, J.; Mu, X.; Wang, Z.; Shen, Q.; Zhao, W. Attack-Defense Strategy Assisted Osprey Optimization Algorithm for PEMFC Parameters Identification. Renew. Energy 2024, 225, 120211. [Google Scholar] [CrossRef]
  21. Wen, X.; Liu, X.; Yu, C.; Gao, H.; Wang, J.; Liang, Y.; Yu, J.; Bai, Y.; Wen, X.; Liu, X.; et al. IOOA: A Multi-Strategy Fusion Improved Osprey Optimization Algorithm for Global Optimization. Electron. Res. Arch. 2024, 32, 2033–2074. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Liu, P. Research on Reactive Power Optimization Based on Hybrid Osprey Optimization Algorithm. Energies 2023, 16, 7101. [Google Scholar] [CrossRef]
  23. Li, L.; Huang, X.; Qian, S.; Li, Z.; Li, S.; Mansour, R. Fuzzy Hybrid Coyote Optimization Algorithm for Image Thresholding. Comput. Mater. Contin. 2022, 72, 3073–3090. [Google Scholar] [CrossRef]
  24. Liu, D.; Zhang, S.; Wang, B.; Li, Z. Seagull Algorithm Based on Good Point Set and Dual Hybrid Strategy. In Proceedings of the International Conference on Cloud Computing, Performance Computing, and Deep Learning (CCPCDL 2023), Huzhou, China, 17–19 February 2023; Volume 12712, pp. 79–84. [Google Scholar]
  25. Kurapov, S.; Davidovsky, M. Algorithmic Methods of Finite Discrete Structures. The Four Color Theorem. Theory, Methods, Algorithms. arXiv 2024, arXiv:2405.05270. [Google Scholar]
  26. Javaheri, D.; Gorgin, S.; Lee, J.A.; Masdari, M. An Improved Discrete Harris Hawk Optimization Algorithm for Efficient Workflow Scheduling in Multi-Fog Computing. Sustain. Comput. Inform. Syst. 2022, 36, 100787. [Google Scholar] [CrossRef]
  27. Wei, X.; Zhang, Y.; Song, H.; Qin, H.; Zhao, G. Research on Evacuation Path Planning Based on Improved Sparrow Search Algorithm. CMES-Comput. Model. Eng. Sci. 2024, 139, 1295–1316. [Google Scholar] [CrossRef]
  28. Ouyang, C.; Qiu, Y.; Zhu, D. Adaptive Spiral Flying Sparrow Search Algorithm. Sci. Program. 2021, 2021, e6505253. [Google Scholar] [CrossRef]
  29. Khare, P.; Raju, M.; Gupta, S.C.; Shukla, H. Liberalized Automatic Generation Control of Interconnected Thermal-Hydro-Gas System Using Firefly Algorithm Optimized PID Controller. Int. J. Numer. Model. Electron. Netw. Devices Fields 2024, 37, e3136. [Google Scholar] [CrossRef]
  30. Langenberg, B.; Janczyk, M.; Koob, V.; Kliegl, R.; Mayer, A. A Tutorial on Using the Paired t Test for Power Calculations in Repeated Measures ANOVA with Interactions. Behav. Res. 2023, 55, 2467–2484. [Google Scholar] [CrossRef] [PubMed]
  31. Hernanda, Y.; Kaporina, A.; Nurlaily, D. AnalisisTingkat Pengangguran Provinsi Kalimantan Timur Menggunakan Sign Test, Wilcoxon Test Dan Paired Sample t-Test. Semin. Nas. Teknol. Inf. Dan Mat. Semiot. 2023, 2, 94–102. [Google Scholar]
  32. Vigya; Raj, S.; Shiva, C.K.; Vedik, B.; Mahapatra, S.; Mukherjee, V. A Novel Chaotic Chimp Sine Cosine Algorithm Part-I: For Solving Optimization Problem. Chaos Solitons Fractals 2023, 173, 113672. [Google Scholar] [CrossRef]
Figure 1. Comparison of population initialization.
Figure 1. Comparison of population initialization.
Biomimetics 09 00486 g001
Figure 2. Grouping based on the four-color theorem.
Figure 2. Grouping based on the four-color theorem.
Biomimetics 09 00486 g002
Figure 3. Two-color complementary mechanism.
Figure 3. Two-color complementary mechanism.
Biomimetics 09 00486 g003
Figure 4. The flow chart of the OOA.
Figure 4. The flow chart of the OOA.
Biomimetics 09 00486 g004
Figure 5. The flow chart of the IOOA.
Figure 5. The flow chart of the IOOA.
Biomimetics 09 00486 g005
Figure 6. Variation in P-value with number of iterations.
Figure 6. Variation in P-value with number of iterations.
Biomimetics 09 00486 g006
Figure 7. Variation in A and a with number of iterations.
Figure 7. Variation in A and a with number of iterations.
Biomimetics 09 00486 g007
Figure 8. Convergence analysis on benchmark functions.
Figure 8. Convergence analysis on benchmark functions.
Biomimetics 09 00486 g008
Figure 9. Convergence analysis on CEC2020.
Figure 9. Convergence analysis on CEC2020.
Biomimetics 09 00486 g009aBiomimetics 09 00486 g009b
Figure 10. Convergence analysis on CEC2022.
Figure 10. Convergence analysis on CEC2022.
Biomimetics 09 00486 g010aBiomimetics 09 00486 g010b
Figure 11. Stability analysis on benchmark functions.
Figure 11. Stability analysis on benchmark functions.
Biomimetics 09 00486 g011
Figure 12. Stability analysis on CEC2020.
Figure 12. Stability analysis on CEC2020.
Biomimetics 09 00486 g012aBiomimetics 09 00486 g012b
Figure 13. Stability analysis on CEC2022.
Figure 13. Stability analysis on CEC2022.
Biomimetics 09 00486 g013aBiomimetics 09 00486 g013b
Figure 14. The Friedman test results on benchmark functions.
Figure 14. The Friedman test results on benchmark functions.
Biomimetics 09 00486 g014
Figure 15. The Friedman test results on CEC2020.
Figure 15. The Friedman test results on CEC2020.
Biomimetics 09 00486 g015
Figure 16. The Friedman test results on CEC2022.
Figure 16. The Friedman test results on CEC2022.
Biomimetics 09 00486 g016
Figure 17. An illustration of the schematic structural diagram of a tension/compression spring.
Figure 17. An illustration of the schematic structural diagram of a tension/compression spring.
Biomimetics 09 00486 g017
Figure 18. Iterative curves of various algorithms for a tension/compression spring.
Figure 18. Iterative curves of various algorithms for a tension/compression spring.
Biomimetics 09 00486 g018
Figure 19. Schematic diagram of three-rod truss structure.
Figure 19. Schematic diagram of three-rod truss structure.
Biomimetics 09 00486 g019
Figure 20. Iterative curves of various algorithms for the welded beams design.
Figure 20. Iterative curves of various algorithms for the welded beams design.
Biomimetics 09 00486 g020
Figure 21. Simplified model of the gear train.
Figure 21. Simplified model of the gear train.
Biomimetics 09 00486 g021
Figure 22. Iterative curves of various algorithms for gear transmission.
Figure 22. Iterative curves of various algorithms for gear transmission.
Biomimetics 09 00486 g022
Table 1. The list of internal parameters of algorithms.
Table 1. The list of internal parameters of algorithms.
AlgorithmParameter
ROA C = 0.1 ;
SFO P D = 2 3 ;
CDO S β = log 10 r a n d 1 : 270,000 ;
BA A = 0.5 , r = 0.5 ;
GSA a = p i , b = p i ;
WWPA C = 5 R a n d 1 , F = 5 R a n d 1 ;
OOA P = 0.5 ;
SOOA α = 0 ;
ADSOOA α = 0.2 , β = 100 ;
IOOA   p 0 = 0.5 , δ = 0.02 ,   p m i n = 0.2 ;
Table 2. The information of benchmark test function.
Table 2. The information of benchmark test function.
No.FunctiondimRangeFmin
1Rotated Ackley’s Function100[−100, 100]0
2Schwefel’s Problem 1.2 with Noise100[−50, 50]0
3Rotated Hybrid Composition Function with Noise2[−65.536, 65.536]0
4Hybrid Composition Function4[−5, 5]0
5Composition Function No. 46[0, 1]−3.32
6BOOTH FUNCTION2[−10, 10]0
7ROSENBROCK FUNCTION100[−5, 10]0
8COLVILLE FUNCTION4[−10, 10]0
Table 3. The information of CEC2020.
Table 3. The information of CEC2020.
No.Functionsfopt
Unimodal function9Shifted and Rotated Bent Cigar Function100
Basic functions10Shifted and Rotated Schwefel’s Function1100
11Shifted and Rotated Lunacek bi-Rastrigin Function700
12Expanded Rosenbrock’s plus Griewangk’s Function1900
Hybrid functions13Hybrid Function 1 (N = 3)1700
14Hybrid Function 2 (N = 4)1600
15Hybrid Function 3 (N = 5)2100
Composition functions16Composition Function 1 (N = 3)2200
17Composition Function 2 (N = 4)2400
18Composition Function 3 (N = 5)2500
Search range: [−100, 100]
Table 4. The information of CEC2022.
Table 4. The information of CEC2022.
No.Functionsfopt
Unimodal function19Shifted and full-rotated Zakharov function300
Basic functions20Shifted and full-rotated Rosenbrock’s function400
21Shifted and full-rotated expanded Schaffer’s f6 function600
22Shifted and full-rotated noncontinuous Rastrigin’s function800
23Shifted and full-rotated levy function900
Hybrid functions24Hybrid Function 1 (N = 3)1800
25Hybrid Function 2 (N = 6)2000
26Hybrid Function 3 (N = 5)2200
Composition functions27Composition Function 1 (N = 5) 2300
28Composition Function 2 (N = 4)2400
29Composition Function 3 (N = 5)2600
30Composition Function 4 (N = 6)2700
Search range: [−100, 100]
Table 5. Experimental parameter setting.
Table 5. Experimental parameter setting.
FunctionNTDim
Benchmark functions501000[0, 100]
CEC202050200020
CEC20225010,00020
Table 6. Experimental results on benchmark functions.
Table 6. Experimental results on benchmark functions.
FunctionROASFOCDOBAGSAWWPAOOASOOAADSOOAIOOA
F1Best2.35 × 10−064.45 × 10+002.50 × 10+013.41 × 10+041.04 × 10−061.57 × 10−010.00 × 10+001.35 × 10−021.63 × 10+007.70 × 10−30
Mean1.68 × 10−041.50 × 10+012.50 × 10+015.49 × 10+041.57 × 10−045.77 × 10+001.45 × 10−087.94 × 10−024.75 × 10+003.68 × 10−10
Std1.40 × 10−045.85 × 10+000.00 × 10+001.34 × 10+043.92 × 10−045.79 × 10+004.57 × 10−083.25 × 10−024.58 × 10+001.16 × 10−09
Iter86610005571000916181851000475976
F2Best1.00 × 10−093.10 × 10−031.16 × 10+003.19 × 10+066.67 × 10−096.74 × 10−044.71 × 10−337.90 × 10−052.10 × 10−024.71 × 10−33
Mean1.87 × 10−062.04 × 10−011.27 × 10+002.68 × 10+074.98 × 10−079.42 × 10−021.35 × 10−136.62 × 10−043.13 × 10−027.32 × 10−15
Std2.98 × 10−062.74 × 10−018.16 × 10−021.63 × 10+077.00 × 10−071.59 × 10−014.18 × 10−134.25 × 10−041.51 × 10−021.94 × 10−14
Iter91898584810007063921000249731
F3Best9.98 × 10−011.99 × 10+009.90 × 10+002.98 × 10+009.98 × 10−019.98 × 10−019.98 × 10−019.98 × 10−019.98 × 10−019.98 × 10−01
Mean9.98 × 10−018.13 × 10+001.23 × 10+019.34 × 10+009.98 × 10−013.51 × 10+001.69 × 10+002.88 × 10+001.61 × 10+009.98 × 10−01
Std2.03 × 10−114.55 × 10+001.21 × 10+006.43 × 10+005.20 × 10−102.98 × 10+008.17 × 10−011.99 × 10+005.01 × 10−011.30 × 10−15
Iter5729451183504220342326871906
F4Best3.11 × 10−043.31 × 10−043.12 × 10−041.02 × 10−033.12 × 10−041.82 × 10−033.12 × 10−044.01 × 10−043.07 × 10−043.08 × 10−04
Mean4.50 × 10−048.99 × 10−033.21 × 10−044.85 × 10−033.58 × 10−041.25 × 10−024.74 × 10−042.64 × 10−033.08 × 10−043.10 × 10−04
Std2.08 × 10−041.03 × 10−026.43 × 10−065.49 × 10−036.83 × 10−059.24 × 10−032.56 × 10−043.16 × 10−036.70 × 10−071.81 × 10−06
Iter65061950499991830345610009491000
F5Best−3.20 × 10+00−3.23 × 10+00−3.29 × 10+00−3.32 × 10+00−3.27 × 10+00−1.83 × 10+00−2.96 × 10+00−3.09 × 10+00−3.32 × 10+00−3.32 × 10+00
Mean−2.95 × 10+00−3.04 × 10+00−3.23 × 10+00−3.24 × 10+00−3.11 × 10+00−1.41 × 10+00−2.52 × 10+00−2.73 × 10+00−3.25 × 10+00−3.32 × 10+00
Std3.42 × 10−011.44 × 10−015.55 × 10−025.74 × 10−021.41 × 10−012.43 × 10−013.08 × 10−012.94 × 10−017.35 × 10−023.74 × 10−06
Iter68598216882492223415567996997
F6Best4.31 × 10−063.05 × 10−103.43 × 10−051.41 × 10−117.95 × 10−061.20 × 10−016.25 × 10−142.35 × 10−263.83 × 10−140.00 × 10+00
Mean2.00 × 10−019.43 × 10−092.31 × 10−041.28 × 10−106.16 × 10−051.21 × 10+003.54 × 10−034.94 × 10−042.10 × 10−062.78 × 10−16
Std6.32 × 10−019.06 × 10−091.65 × 10−041.51 × 10−104.57 × 10−057.25 × 10−011.07 × 10−021.56 × 10−035.01 × 10−068.81 × 10−16
Iter59878552432178967149340912765
F7Best1.61 × 10−064.85 × 10+014.69 × 10+014.24 × 10+011.62 × 10−069.33 × 10−010.00 × 10+004.85 × 10+014.84 × 10+011.23 × 10−13
Mean1.25 × 10−044.85 × 10+014.71 × 10+014.47 × 10+011.20 × 10−046.94 × 10+012.43 × 10+014.85 × 10+014.85 × 10+013.09 × 10+01
Std1.75 × 10−049.51 × 10−049.41 × 10−021.35 × 10+001.58 × 10−045.70 × 10+012.56 × 10+012.50 × 10−033.28 × 10−022.32 × 10+01
Iter6525783081000530661851000935982
F8Best2.03 × 10−069.18 × 10−052.27 × 10−024.54 × 10−061.97 × 10−072.13 × 10−010.00 × 10+006.94 × 10−032.84 × 10−070.00 × 10+00
Mean2.50 × 10−044.20 × 10+001.35 × 10−016.27 × 10−011.64 × 10−056.08 × 10+004.36 × 10−052.14 × 10+006.98 × 10−031.17 × 10−13
Std3.40 × 10−043.27 × 10+001.52 × 10−011.27 × 10+003.52 × 10−057.71 × 10+001.38 × 10−042.31 × 10+001.23 × 10−023.42 × 10−13
Iter648464218992705137661000967895
Table 7. Experimental results on CEC2020.
Table 7. Experimental results on CEC2020.
FunctionROASFOCDOBAGSAWWPAOOASOOAADSOOAIOOA
F9Best7.39 × 10+099.26 × 10+092.11 × 10+101.39 × 10+101.08 × 10+104.05 × 10+101.71 × 10+101.68 × 10+101.01 × 10+102.63 × 10+05
Mean1.93 × 10+101.88 × 10+102.12 × 10+102.37 × 10+101.75 × 10+104.28 × 10+102.60 × 10+102.40 × 10+101.47 × 10+105.58 × 10+07
Std4.87 × 10+096.44 × 10+091.40 × 10+087.38 × 10+095.85 × 10+091.20 × 10+096.91 × 10+094.24 × 10+092.50 × 10+091.16 × 10+08
Iter86119436972000195391146257219861999
F10Best4.10 × 10+034.79 × 10+034.35 × 10+033.46 × 10+034.63 × 10+035.83 × 10+034.54 × 10+034.70 × 10+033.79 × 10+032.80 × 10+03
Mean4.95 × 10+035.39 × 10+034.84 × 10+034.30 × 10+035.14 × 10+036.16 × 10+035.00 × 10+035.31 × 10+034.26 × 10+033.23 × 10+03
Std6.11 × 10+024.12 × 10+022.73 × 10+024.38 × 10+024.44 × 10+022.63 × 10+023.42 × 10+024.22 × 10+024.00 × 10+023.45 × 10+02
Iter7666104931367152922613965716311998
F11Best9.44 × 10+021.00 × 10+039.77 × 10+021.07 × 10+031.01 × 10+031.11 × 10+039.81 × 10+029.50 × 10+028.99 × 10+028.81 × 10+02
Mean9.95 × 10+021.03 × 10+039.85 × 10+021.23 × 10+031.04 × 10+031.13 × 10+031.02 × 10+031.03 × 10+039.44 × 10+029.42 × 10+02
Std2.79 × 10+011.97 × 10+011.26 × 10+011.11 × 10+021.66 × 10+011.11 × 10+012.15 × 10+013.42 × 10+013.28 × 10+013.15 × 10+01
Iter921181476113942000553338731512000
F12Best1.70 × 10+043.28 × 10+043.87 × 10+052.42 × 10+045.68 × 10+031.71 × 10+069.21 × 10+042.48 × 10+041.57 × 10+041.91 × 10+03
Mean3.37 × 10+051.04 × 10+053.88 × 10+055.37 × 10+051.23 × 10+052.86 × 10+065.15 × 10+051.60 × 10+051.09 × 10+051.93 × 10+03
Std8.02 × 10+056.25 × 10+047.95 × 10+026.04 × 10+058.43 × 10+049.13 × 10+053.46 × 10+051.11 × 10+051.21 × 10+052.02 × 10+01
Iter84016644822000197883116745617711699
F13Best9.72 × 10+055.99 × 10+054.16 × 10+052.39 × 10+051.20 × 10+066.53 × 10+063.33 × 10+061.60 × 10+062.86 × 10+051.64 × 10+05
Mean2.75 × 10+063.58 × 10+066.69 × 10+051.03 × 10+073.41 × 10+062.69 × 10+076.30 × 10+066.35 × 10+069.83 × 10+057.77 × 10+05
Std1.37 × 10+062.32 × 10+062.72 × 10+051.50 × 10+071.51 × 10+069.89 × 10+063.04 × 10+065.56 × 10+066.25 × 10+054.19 × 10+05
Iter127010827872000178870189826019511986
F14Best2.39 × 10+032.36 × 10+032.76 × 10+032.64 × 10+032.61 × 10+033.49 × 10+032.81 × 10+032.67 × 10+032.16 × 10+032.11 × 10+03
Mean2.74 × 10+033.02 × 10+033.09 × 10+033.05 × 10+033.02 × 10+034.17 × 10+033.04 × 10+033.18 × 10+032.57 × 10+032.37 × 10+03
Std2.88 × 10+023.59 × 10+021.82 × 10+022.94 × 10+023.20 × 10+023.49 × 10+021.55 × 10+022.37 × 10+022.57 × 10+022.34 × 10+02
Iter9101993443200018651811990188819761727
F15Best1.28 × 10+056.66 × 10+043.45 × 10+054.24 × 10+061.09 × 10+055.72 × 10+061.91 × 10+053.19 × 10+051.40 × 10+055.45 × 10+04
Mean1.25 × 10+063.54 × 10+062.42 × 10+077.99 × 10+062.21 × 10+062.86 × 10+071.86 × 10+061.72 × 10+064.30 × 10+053.51 × 10+05
Std1.74 × 10+062.85 × 10+062.59 × 10+075.86 × 10+062.31 × 10+062.24 × 10+072.35 × 10+061.29 × 10+062.57 × 10+053.32 × 10+05
Iter1331173591620001971281603199818832000
F16Best3.78 × 10+033.42 × 10+034.22 × 10+034.19 × 10+033.66 × 10+036.71 × 10+033.88 × 10+034.43 × 10+033.37 × 10+032.31 × 10+03
Mean5.24 × 10+035.51 × 10+035.22 × 10+035.82 × 10+035.67 × 10+037.41 × 10+035.33 × 10+035.55 × 10+033.90 × 10+032.32 × 10+03
Std1.16 × 10+031.61 × 10+031.31 × 10+038.62 × 10+029.65 × 10+022.78 × 10+029.29 × 10+026.49 × 10+023.38 × 10+021.22 × 10+01
Iter61315271370200017112221012199419761986
F17Best3.08 × 10+032.95 × 10+033.24 × 10+033.14 × 10+033.05 × 10+033.20 × 10+033.28 × 10+033.34 × 10+032.97 × 10+032.90 × 10+03
Mean3.19 × 10+033.17 × 10+033.34 × 10+033.27 × 10+033.25 × 10+033.48 × 10+033.55 × 10+033.53 × 10+033.01 × 10+033.03 × 10+03
Std7.99 × 10+011.06 × 10+024.93 × 10+018.12 × 10+011.87 × 10+022.07 × 10+021.78 × 10+021.20 × 10+022.09 × 10+011.13 × 10+02
Iter716199344320001997801709719662000
F18Best3.44 × 10+033.63 × 10+035.91 × 10+034.24 × 10+033.28 × 10+036.85 × 10+034.41 × 10+033.95 × 10+033.34 × 10+032.96 × 10+03
Mean4.20 × 10+034.31 × 10+035.92 × 10+035.46 × 10+034.51 × 10+038.72 × 10+035.59 × 10+034.90 × 10+034.01 × 10+033.00 × 10+03
Std7.69 × 10+026.30 × 10+028.74 × 10+001.48 × 10+036.59 × 10+021.34 × 10+039.61 × 10+026.74 × 10+023.16 × 10+022.43 × 10+01
Iter977199359120001920943587176732000
Table 8. Experimental results on CEC2022.
Table 8. Experimental results on CEC2022.
FunctionROASFOCDOBAGSAWWPAOOASOOAADSOOAIOOA
F19Best1.52 × 10+042.74 × 10+042.25 × 10+042.37 × 10+032.08 × 10+049.28 × 10+042.11 × 10+042.65 × 10+041.13 × 10+043.00 × 10+02
Mean3.58 × 10+044.75 × 10+042.26 × 10+041.13 × 10+054.82 × 10+041.56 × 10+054.17 × 10+045.16 × 10+041.76 × 10+043.01 × 10+02
Std1.35 × 10+041.31 × 10+046.16 × 10+011.26 × 10+054.00 × 10+043.65 × 10+048.34 × 10+032.61 × 10+043.93 × 10+032.46 × 10+00
Iter57689929285310,0007392630596486788410,000
F20Best8.51 × 10+021.14 × 10+031.96 × 10+036.44 × 10+028.63 × 10+022.11 × 10+031.59 × 10+031.04 × 10+031.11 × 10+034.03 × 10+02
Mean1.42 × 10+031.71 × 10+031.97 × 10+038.34 × 10+021.46 × 10+034.11 × 10+032.50 × 10+032.06 × 10+031.47 × 10+034.33 × 10+02
Std5.07 × 10+024.22 × 10+025.07 × 10+001.35 × 10+024.72 × 10+021.11 × 10+036.13 × 10+026.50 × 10+023.07 × 10+021.58 × 10+01
Iter51379966395510,0008320861382589353879975
F21Best6.47 × 10+026.63 × 10+026.47 × 10+026.59 × 10+026.64 × 10+026.98 × 10+026.53 × 10+026.59 × 10+026.58 × 10+026.41 × 10+02
Mean6.68 × 10+026.81 × 10+026.49 × 10+026.67 × 10+026.78 × 10+027.10 × 10+026.71 × 10+026.72 × 10+026.66 × 10+026.57 × 10+02
Std1.24 × 10+011.02 × 10+011.93 × 10+009.12 × 10+001.08 × 10+016.71 × 10+001.12 × 10+018.48 × 10+005.38 × 10+006.86 × 10+00
Iter42579900282410,00093764556241701761459883
F22Best9.07 × 10+029.63 × 10+028.99 × 10+028.62 × 10+029.16 × 10+029.96 × 10+029.34 × 10+029.43 × 10+029.19 × 10+028.72 × 10+02
Mean9.43 × 10+029.86 × 10+029.19 × 10+028.91 × 10+029.42 × 10+021.02 × 10+039.62 × 10+029.65 × 10+029.37 × 10+028.86 × 10+02
Std1.70 × 10+011.71 × 10+011.37 × 10+012.37 × 10+011.71 × 10+011.14 × 10+011.30 × 10+011.72 × 10+011.48 × 10+017.08 × 10+00
Iter4598375338683229873036821316244794189779
F23Best2.59 × 10+032.76 × 10+032.76 × 10+032.43 × 10+032.94 × 10+035.09 × 10+032.80 × 10+032.53 × 10+032.36 × 10+032.39 × 10+03
Mean3.27 × 10+034.28 × 10+032.89 × 10+033.26 × 10+033.31 × 10+035.47 × 10+033.37 × 10+033.08 × 10+032.76 × 10+032.43 × 10+03
Std4.41 × 10+027.97 × 10+027.30 × 10+016.84 × 10+021.94 × 10+022.28 × 10+024.23 × 10+024.61 × 10+022.61 × 10+021.41 × 10+01
Iter83049688407395149373799892554751359994
F24Best4.95 × 10+067.87 × 10+077.28 × 10+082.18 × 10+032.95 × 10+072.95 × 10+091.57 × 10+085.38 × 10+086.15 × 10+072.01 × 10+03
Mean4.00 × 10+082.25 × 10+083.34 × 10+098.65 × 10+036.43 × 10+086.09 × 10+091.56 × 10+091.16 × 10+093.22 × 10+083.24 × 10+03
Std5.73 × 10+081.13 × 10+082.75 × 10+099.68 × 10+037.00 × 10+081.71 × 10+091.03 × 10+097.46 × 10+082.36 × 10+086.82 × 10+02
Iter464699604142981898031121119156597006264
F25Best2.11 × 10+032.17 × 10+032.26 × 10+032.16 × 10+032.16 × 10+032.35 × 10+032.15 × 10+032.16 × 10+032.12 × 10+032.07 × 10+03
Mean2.18 × 10+032.31 × 10+032.30 × 10+032.28 × 10+032.22 × 10+032.45 × 10+032.18 × 10+032.21 × 10+032.15 × 10+032.16 × 10+03
Std4.55 × 10+011.04 × 10+023.57 × 10+019.88 × 10+014.21 × 10+015.81 × 10+011.84 × 10+013.65 × 10+012.93 × 10+013.23 × 10+01
Iter45218979374594789733177090910,00097669309
F26Best2.24 × 10+032.26 × 10+032.24 × 10+032.61 × 10+032.23 × 10+032.34 × 10+032.25 × 10+032.24 × 10+032.23 × 10+032.23 × 10+03
Mean2.31 × 10+032.51 × 10+032.24 × 10+032.98 × 10+032.28 × 10+032.69 × 10+032.31 × 10+032.38 × 10+032.32 × 10+032.24 × 10+03
Std2.06 × 10+022.08 × 10+024.97 × 10+004.28 × 10+026.85 × 10+012.38 × 10+026.44 × 10+011.47 × 10+028.91 × 10+015.57 × 10+00
Iter5671989434699476937215599413908948610,000
F27Best2.73 × 10+032.77 × 10+033.08 × 10+032.64 × 10+032.65 × 10+033.81 × 10+032.96 × 10+033.03 × 10+032.64 × 10+032.47 × 10+03
Mean2.79 × 10+032.91 × 10+033.43 × 10+032.75 × 10+032.89 × 10+034.26 × 10+033.41 × 10+033.28 × 10+032.76 × 10+032.47 × 10+03
Std6.51 × 10+011.02 × 10+021.22 × 10+021.09 × 10+021.73 × 10+022.62 × 10+024.15 × 10+022.33 × 10+021.03 × 10+021.62 × 10+00
Iter77359993344410,0009468262211717776949000
F28Best5.30 × 10+036.32 × 10+035.02 × 10+034.53 × 10+032.60 × 10+036.68 × 10+034.88 × 10+032.70 × 10+032.74 × 10+032.50 × 10+03
Mean5.88 × 10+037.15 × 10+035.55 × 10+035.24 × 10+035.56 × 10+037.17 × 10+036.40 × 10+035.27 × 10+034.23 × 10+032.77 × 10+03
Std4.26 × 10+024.54 × 10+023.47 × 10+025.16 × 10+021.11 × 10+032.83 × 10+026.32 × 10+021.73 × 10+031.20 × 10+035.77 × 10+02
Iter6587996471410,000968840155790981891249548
F29Best5.58 × 10+037.78 × 10+038.42 × 10+033.42 × 10+046.65 × 10+039.19 × 10+038.01 × 10+036.91 × 10+036.45 × 10+032.61 × 10+03
Mean7.16 × 10+038.61 × 10+038.44 × 10+035.33 × 10+047.72 × 10+039.76 × 10+038.92 × 10+038.34 × 10+037.41 × 10+032.77 × 10+03
Std9.07 × 10+026.37 × 10+021.85 × 10+011.31 × 10+045.50 × 10+023.60 × 10+025.51 × 10+026.47 × 10+026.89 × 10+021.68 × 10+02
Iter47644097162010,0009456620548411050229995
F30Best3.04 × 10+033.07 × 10+033.44 × 10+032.90 × 10+033.15 × 10+032.90 × 10+033.65 × 10+033.64 × 10+033.11 × 10+032.98 × 10+03
Mean3.21 × 10+033.26 × 10+033.47 × 10+032.90 × 10+033.41 × 10+032.90 × 10+033.86 × 10+033.90 × 10+033.17 × 10+033.04 × 10+03
Std1.75 × 10+021.14 × 10+021.72 × 10+015.07 × 10−052.58 × 10+023.23 × 10−051.28 × 10+023.05 × 10+027.55 × 10+013.94 × 10+01
Iter426183653121841855797426402998999439742
Table 9. The Wilcoxon test results on benchmark functions.
Table 9. The Wilcoxon test results on benchmark functions.
IOOA vs.ROASFOCDOBAGSAWWPAOOASOOAADSOOA
F10.0002 (+)0.0002 (+)0.0001 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.5204 (=)0.0002 (+)0.0002 (+)
F20.0002 (+)0.0002 (+)0.0001 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.4272 (=)0.0002 (+)0.0002 (+)
F3NaN (=)0.0001 (+)0.0000 (+)0.0001 (+)NaN (=)0.0008 (+)0.0146 (+)0.0007 (+)0.0002 (+)
F40.0004 (+)0.0002 (+)0.0003 (+)0.0002 (+)0.0003 (+)0.0002 (+)0.0003 (+)0.0002 (+)0.0234 (−)
F50.0001 (+)0.0001 (+)0.0001 (+)0.0016 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0148 (+)
F60.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0009 (+)0.0001 (+)
F70.1390 (=)0.0060 (+)0.4649 (=)0.4708 (=)0.1390 (=)0.0252 (+)0.4988 (=)0.0060 (+)0.0362 (+)
F80.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.5199 (=)0.0002 (+)0.0002 (+)
+/=/−6/2/08/0/07/1/07/1/06/2/08/0/04/4/08/0/07/1/0
Table 10. The Wilcoxon test results on CEC2020.
Table 10. The Wilcoxon test results on CEC2020.
IOOA vs.ROASFOCDOBAGSAWWPAOOASOOAADSOOA
F90.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F100.0002 (+)0.0002 (+)0.0002 (+)0.0006 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F110.0028 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0010 (+)0.8501 (=)
F120.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F130.0010 (+)0.0046 (+)0.4960 (=)0.0022 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.5964 (=)
F140.0081 (+)0.0008 (+)0.0002 (+)0.0004 (+)0.0010 (+)0.0002 (+)0.0002 (+)0.0003 (+)0.0694 (=)
F150.1041 (=)0.0046 (+)0.0013 (+)0.0002 (+)0.0091 (+)0.0002 (+)0.0073 (+)0.0010 (+)0.3075 (=)
F160.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F170.0091 (+)0.0140 (+)0.0002 (+)0.0008 (+)0.0058 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.8501 (=)
F180.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
+/=/−9/1/010/0/09/1/010/0/010/0/010/0/010/0/010/0/05/5/0
Table 11. The Wilcoxon test results on CEC2022.
Table 11. The Wilcoxon test results on CEC2022.
IOOA vs.ROASFOCDOBAGSAWWPAOOASOOAADSOOA
F190.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F200.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)0.0001 (+)
F210.0757 (=)0.0002 (+)0.0073 (-)0.0140 (+)0.0002 (+)0.0002 (+)0.0113 (+)0.0017 (+)0.0046 (+)
F220.0002 (+)0.0002 (+)0.0002 (+)0.6232 (=)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F230.0002 (+)0.0002 (+)0.0002 (+)0.0010 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0211 (+)
F240.0002 (+)0.0002 (+)0.0002 (+)0.0539 (=)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F250.4727 (=)0.0008 (+)0.0002 (+)0.0140 (+)0.0036 (+)0.0002 (+)0.0757 (=)0.0046 (+)0.1405 (=)
F260.0073 (+)0.0002 (+)0.0091 (+)0.0002 (+)0.0073 (+)0.0002 (+)0.0002 (+)0.0003 (+)0.0022 (+)
F270.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F280.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0003 (+)0.0002 (+)0.0002 (+)0.0010 (+)0.0022 (+)
F290.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)0.0002 (+)
F300.0004 (+)0.0003 (+)0.0002 (+)0.0001 (-)0.0002 (+)0.0002 (−)0.0002 (+)0.0002 (+)0.0004 (+)
+/=/−10/2/012/0/011/0/19/2/112/0/011/0/111/1/012/0/011/1/0
Table 12. Comparison results of the tension spring design.
Table 12. Comparison results of the tension spring design.
MethodsOptimal Solution of Design Variablesfmin (X)Iter
x 1 x 2 x 3
ROA5.89 × 10−025.54 × 10−015.09 × 10+001.36 × 10−021353
SFO5.70 × 10−024.97 × 10−016.15 × 10+001.31 × 10−021418
CDO5.00 × 10−023.17 × 10−011.42 × 10+011.28 × 10−02904
BA4.96 × 10−023.09 × 10−011.47 × 10+011.27 × 10−021638
GSA5.85 × 10−025.44 × 10−015.30 × 10+001.36 × 10−022000
WWPA1.44 × 10−011.30 × 10+001.50 × 10+013.99 × 10+99185
OOA5.37 × 10−024.61 × 10−014.20 × 10+001.1 × 10+100276
SOOA5.90 × 10−025.59 × 10−014.97 × 10+001.36 × 10−021442
ADSOOA5.00 × 10−023.17 × 10−011.40 × 10+011.27 × 10−021870
IOOA5.18 × 10−023.60 × 10−011.11 × 10+011.27 × 10−021247
Table 13. Comparison results of the welded beams design.
Table 13. Comparison results of the welded beams design.
MethodsOptimal Solution of Design Variablesfmin (X)Iter
x 1 x 2 x 3 x 4
ROA1.85 × 10−013.86 × 10+008.84 × 10+002.15 × 10−011.78 × 10+001334
SFO1.39 × 10−017.69 × 10+007.14 × 10+003.29 × 10−012.62 × 10+001954
CDO2.00 × 10−013.36 × 10+009.32 × 10+002.07 × 10−011.76 × 10+00457
BA−9.26 × 10−029.08 × 10+006.94 × 10+00−3.28 × 10−01−1.39 × 10+1084
GSA2.04 × 10−013.40 × 10+009.08 × 10+002.06 × 10−011.72 × 10+002000
WWPA7.75 × 10−011.72 × 10+001.00 × 10+016.40 × 10−019.86 × 10+100661
OOA2.45 × 10−017.14 × 10+006.73 × 10+003.70 × 10−013.01 × 10+00693
SOOA1.73 × 10+005.62 × 10−013.32 × 10+001.52 × 10+002.08 × 10+100501
ADSOOA2.54 × 10−013.05 × 10+007.59 × 10+002.91 × 10−012.03 × 10+001974
IOOA1.98 × 10−013.43 × 10+009.03 × 10+002.06 × 10−011.71 × 10+001997
Table 14. Comparison results of design problems of gear transmission.
Table 14. Comparison results of design problems of gear transmission.
MethodsOptimal Solution of Design Variablesfmin (X)Iter
x 1 x 2 x 3 x 4
ROA6.00 × 10+012.02 × 10+012.62 × 10+016.00 × 10+012.73 × 10−08915
SFO4.66 × 10+011.31 × 10+011.21 × 10+012.30 × 10+019.92 × 10−1020
CDO3.73 × 10+012.11 × 10+011.45 × 10+015.86 × 10+013.07 × 10−10200
BA2.32 × 10+02−2.05 × 10+02−7.33 × 10+014.44 × 10+021.00 × 10−061173
GSA6.00 × 10+011.35 × 10+014.03 × 10+016.00 × 10+012.73 × 10−082000
WWPA2.92 × 10+161.23 × 10+171.21 × 10+163.53 × 10+172.17 × 10−07879
OOA4.12 × 10+011.63 × 10+011.67 × 10+014.56 × 10+013.45 × 10−0916
SOOA4.57 × 10+013.75 × 10+011.24 × 10+015.28 × 10+011.43 × 10−0315
ADSOOA2.33 × 10+011.33 × 10+011.20 × 10+014.66 × 10+019.92 × 10−1027
IOOA4.27 × 10+011.60 × 10+011.91 × 10+014.90 × 10+012.70 × 10−121429
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, F.; Shi, X.; Feng, Y. Improved Osprey Optimization Algorithm Based on Two-Color Complementary Mechanism for Global Optimization and Engineering Problems. Biomimetics 2024, 9, 486. https://doi.org/10.3390/biomimetics9080486

AMA Style

Wei F, Shi X, Feng Y. Improved Osprey Optimization Algorithm Based on Two-Color Complementary Mechanism for Global Optimization and Engineering Problems. Biomimetics. 2024; 9(8):486. https://doi.org/10.3390/biomimetics9080486

Chicago/Turabian Style

Wei, Fengtao, Xin Shi, and Yue Feng. 2024. "Improved Osprey Optimization Algorithm Based on Two-Color Complementary Mechanism for Global Optimization and Engineering Problems" Biomimetics 9, no. 8: 486. https://doi.org/10.3390/biomimetics9080486

APA Style

Wei, F., Shi, X., & Feng, Y. (2024). Improved Osprey Optimization Algorithm Based on Two-Color Complementary Mechanism for Global Optimization and Engineering Problems. Biomimetics, 9(8), 486. https://doi.org/10.3390/biomimetics9080486

Article Metrics

Back to TopTop