Next Article in Journal
Toward a Wong–Zakai Approximation for Big Order Generators
Previous Article in Journal
Symmetry and Asymmetry in the Thermo-Magnetic Convection of Silver Nanofluid
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Modified Whale Optimization Algorithm with Single-Dimensional Swimming for Global Optimization Problems

1
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
2
Nanjing Research Institute of Electronic Engineering, Nanjing 210007, China
3
Nanjing Customs, Nanjing 210001, China
*
Author to whom correspondence should be addressed.
Symmetry 2020, 12(11), 1892; https://doi.org/10.3390/sym12111892
Submission received: 1 October 2020 / Revised: 12 November 2020 / Accepted: 16 November 2020 / Published: 18 November 2020
(This article belongs to the Section Computer)

Abstract

:
As a novel meta-heuristic algorithm, the Whale Optimization Algorithm (WOA) has well performance in solving optimization problems. However, WOA usually tends to trap in local optimal and it suffers slow convergence speed for large-scale and high-dimension optimization problems. A modified whale optimization algorithm with single-dimensional swimming (abbreviated as SWWOA) is proposed in order to overcome the shortcoming. First, tent map is applied to generate the initialize population for maximize search ability. Second, quasi-opposition learning is adopted after every iteration for further improving the search ability. Third, a novel nonlinearly control parameter factor that is based on logarithm function is presented in order to balance exploration and exploitation. Additionally, the last, single-dimensional swimming is proposed in order to replace the prey behaviour in standard WOA for tuning. The simulation experiments were conducted on 20 well-known benchmark functions. The results show that the proposed SWWOA has better performance in solution precision and higher convergence speed than the comparison methods.

1. Introduction

With the development of technology, increasing global optimization problems have to be solved in various fields, such as economic scheduling, aerospace, signal processing, artificial intelligence, mechanical design, chemical engineering [1,2,3], etc. In general, optimization problems with typical mathematical characteristics can be solved by traditional algorithms and the optimal solution is guaranteed in this case. However, many problems in modern applications have the characteristics of large-scale, high-dimensional, and lack of typical mathematical characteristics, and they cannot be solved by traditional optimization algorithms or the solution is too complex to be feasible. For this, many scholars have conducted research on the meta-heuristic algorithm and have remarkable results.
A meta-heuristic algorithm is an implementation on a specific problem guided by a set of guidelines or strategies, which adopts the “trial-and-error” mechanism [4]. The “trial-and-error” mechanism is a method for obtaining a feasible solution at first, and then gradually improve it by comparing the fitness of feasible solutions, finally approach or obtain the optimal solution. The meta-heuristic algorithms cannot guarantee to obtain the optimal, but it can obtain a satisfactory solution within a certain amount of time. The “trial-and-error” mechanism adopted by the meta-heuristic algorithms ensure that it does not require the problem to have precise mathematical characteristics, and it is very adaptable. Motivated by the diversity of engineering applications, many meta-heuristic algorithms have been proposed, such as the Genetic Algorithm (GA) [5,6], Simulated Annealing (SA) [7,8], and the subsequently, Ant Colony Optimization (ACO) [9,10], Differential Evolution (DE) [11], Particle Swarm Optimization (PSO) [12,13,14], Gravitational Search Algorithm (GSA) [15], and so on. Some of these meta-heuristic algorithms are called Swarm Intelligent algorithm (SI), which iteratively solve problems by simulating group collaboration. In recent years, a large number of novel swarm intelligent optimization algorithms have been proposed in order to meet the challenge of global optimization problems, such as artificial bee colony optimization (ABC) [16,17,18], Whale Optimization Algorithm, (WOA) [19,20,21,22], Glowworm Swarm Optimization (GSO) [23,24], Grey Wolf Optimization (GWO) [25], and Symbiotic Organisms Search (SOS) [26,27], etc.
The meta-heuristic algorithm can generally obtain good results when solving small-scale optimization problems, but, when targeting some high-dimensional large-scale optimization problems, there are often two problems: (1) convergence speed slow, which leads to long computation time; and, (2) it is easy to fall into the local optimum. These two problems are related to each other, the former is that the approximation speed is too low, which can be improved by improving some parameters or introducing some mechanisms, while the latter is mainly due to the lack of population diversity, which is directly related to the final solution quality, and it can be improved by enhancing population diversity. Many scholars have conducted a lot of work on the shortcomings of meta-heuristic algorithms, and have obtained good results by improving the standard algorithm or mixing the standard algorithm with various mechanisms.
Whale Optimization Algorithm is a new meta-heuristic optimization algorithm for simulating humpback whale hunting behavior proposed by Mirjalili [28] in 2016, and it has been shown that WOA has better optimization performance when compared to PSO, ABC, and DE algorithms [29], but it still suffers from slow convergence and low solution accuracy when solving high-dimensional large-scale optimization problems. In view of this, a lot of work has been done on WOA for improving the WOA algorithm in order to obtain better performance. In the literature [30], Adel introduces the concept of leader to guide the population into the optimal solution region, which enhances the convergence speed of WOA. In [31], Mohamed proposed that chaotic sequences and Opposition-Based Learning (OBL) are the two most effective ways to improve WOA. The adaptive chaotic sequence selection method is proposed in the paper in order to improve the diversity of the initial population and, at the same time, a part of the population is selected to execute the DE algorithm, which finally results in an improved DEWCO algorithm that is a mixture of WOA and DE. A nonlinear dynamic control parameter update strategy that is based on a cosine function is proposed in [20] in order to balance the exploration and tuning ability. Additioanlly, the Lévy flight strategy is used to make the algorithm jump out of the local optimum and avoid stagnation [22]. In the literature [19], a WOA based on quadratic interpolation is proposed. The algorithm mainly introduces new parameters, improves the search process, and balances the convergence speed and solution accuracy. At the same time, quadratic interpolation is used to search the optimal agent, which improves the solution accuracy. The literature [21] employs a logistic chaos map in order to improve the distribution of the initial population in the solution space and a quasi-opposition learning mechanism, in which both the standard algorithm and the quasi-opposition learning generate and evaluate agents during predation, and the better agent is retained by comparing the adaptation values of the two agents to improve the convergence speed. In literature [32], two strategies are used to improve the standard WOA: (1) random replacement of poorer agents with better ones, which is used to improve the algorithm’s solution convergence speed, and (2) adaptive double weights, which is used to balance the algorithm’s early spatial search ability and later local spatial tuning ability. There are three main improvements in the literature [33]; firstly, a chaotic sequence is introduced for optimizing the initial population. Subsequently, Gaussian variation is used to maintain the diversity level of the population. Finally, a “reduced” strategy is used to search near the optimal solution. The literature [34] introduced quantum behavior in the standard WOA to simulate the hunting process of humpback whales in order to enhance the search capability of the algorithm, which is used for feature selection. Although these studies have improved the standard WOA algorithm to some extent, there are still problems of slow convergence and low solution accuracy, especially for high-dimensional large-scale optimization problems.
In this paper, a modified WOA algorithm SWWOA based on single-dimensional swimming is proposed, with four main improvements: (1) the use of tent chaotic sequences to optimize the quality of the initial population; (2) the introduction of quasi-opposition learning mechanism, any agent updated position will be learned by quasi-opposition learning, retain the better agent by fitness; (3) the use of logarithmic function to dynamically update the weights, instead of the original linear weights. Balancing the between convergence speed and solution quality; and, (4) the single-dimensional swimming improvement is borrowed from the single-dimensional update of position in the ABC algorithm, which replaces the full dimensional update in the standard WOA with single-dimensional improvement in order to further improve the algorithm’s ability.
The rest of paper is organized, as follows. Section 2 introduces the standard whale optimization algorithm. Section 3 describes the proposed SWWOA algorithm. Simulations and the discussion of results are shown in Section 4. Finally, Section 5 gives the conclusions.

2. Standard WOA Algorithm

WOA is a novel meta-heuristic optimization algorithm by imitating the hunting mechanism of humpback whales, which consists of three phases: encircling prey, spiral bubble-net feeding maneuver, and search for prey [28].

2.1. Encircling Prey (Exploitation Phase)

At this stage, the algorithm sets the current optimal position in the population as the global optimal, which is, the prey. All of the whales in the population move towards the prey and gradually shrink to surround the prey, as follows:
D = C · X * t X t
X t + 1 = X * t A · D
where t represents the current time, X  is the position vector, which represents a feasible solution. X *  represents the optimal solution at the current time and · represents the absolute value. A and C are two control parameters, which are calculated, as follows:
A = 2 a r a
C = 2 r
where a is linearly factor from 2 to 0 over the whole iterations (both exploration and exploitation) and r is a random number in [0, 1].

2.2. Bubble-Net Attacking Method (Exploitation Phase)

At this stage, WOA imitate humpback whales attacking prey with bubble net, which are essentially spiral search space, whose mathematical model is as follows:
D = X * t X t
X t + 1 = D · e b l · cos 2 π l + X * t
where D represents the distance between current whale and the best solution of population. l is a random number in [−1, 1] and b is a constant used in order to define the shape of the spiral, normally is 1. Exploitation phase has bubble-net attacking and encircling prey two methods, each is implemented with 50% probability. Accordingly, combining the two method, the following formula can be obtained:
X t + 1 = X * t A · D p < 0.5 X * t + D · e b l · cos 2 π l p 0.5

2.3. Search for Prey (Exploration Phase)

In a standard WOA algorithm, this stage is the mainly phase for exploration, where the mathematical model is similar to Equations (1) and (2), and the only difference is the use of a random agent instead of the optimal agent. The formula is as follows:
D = C · X r a n d t X t
X t + 1 = X r a n d t A · D
X r a n d denotes a random agent in population, other meanings same to the Section 2.1. It is worth noting that the scheduling between encircling (exploitation) and search (exploration) is done by the value of A . When A < 1 , exploitation is selected. When A 1 , exploration is selected.

2.4. The Pseudo Code of WOA

When compared with other meta-heuristic algorithms, in addition to the necessary parameters, such as population size and max iteration times, the WOA algorithm only has one parameter a for balanced exploitation and exploration, which is a very good advantage. However, from another aspect, WOA has very few adjustable parameters and it is slightly lacking in flexibility. The pseudo code of the standard WOA Algorithm 1 is as follows:
Algorithm 1 WOA
01 initialize maxIteration, popsize and parameter b
02 initialize the population and calculate fitness
03 obtain the optimal agent
04 WHILE t<maxIteration DO
05  update a, A, C by Equations (3) and (4)
06  WHILE i<popsize DO
07   generate random number p [ 0 , 1 ]
08   IF p < 0.5 THEN
09    IF A < 1 THEN
10     update position of agent i by Equation (2)
11    ELSE
12     generate random agent rand
13     update position of agent i by Equation (9)
14    ENDIF
15   ELSE
16     update position of agent i by Equation (6)
17   ENDIF
18   i = i + 1
19  ENDWHILE
20  update optimal agent if there is a better solution
21   t = t + 1
22 ENDWHILE
23 RETURN optimal agent

3. Whale Optimization Algorithm with Single-Dimensional Swimming (SWWOA)

In this paper, an improved algorithm is proposed for the features of the standard WOA algorithm. The main improvements are divided into four interrelated aspects. First, the quality of the initial population is directly related to the convergence speed and, if there are agents located in the optimal solution region at the beginning, it will save a lot of useless calculations. Subsequently, quasi-opposition learning is introduced in order to improve search capability in all directions. On the basis of greatly improving the search ability, this paper introduces a logarithm-based nonlinear parameter, which is used in order to apply more computation to tuning and improve the solution accuracy. Finally, borrowing from the ABC algorithm for finding food, the full-dimensional encircling prey is replaced by a single-dimensional swimming, which is essentially a more fine-grained approach to finding excellence and, moreover, gives the algorithm the ability to jump out of the local optimum. Before describing the improvements in this paper, the test functions are first described.

3.1. Chaotic Sequence Based on Tent Map

A chaotic system is a deterministic system, in which there is seemingly random irregular movement that behaves in an indeterminate, unrepeatable, and unpredictable manner [21]. Chaos is an inherent property of nonlinear system and it is a common phenomenon in nonlinear systems. There are many map functions for chaotic system, the most commonly used of which are logistic map and tent map [21,31,33]. In this paper, we use tent map with the following formula. Presently, most of the improved algorithms use logistic map. However, logistic map has uneven traversal characteristics, while tent map has better uniform characteristics [31]. Figure 1 shows the specific comparison of logistic map and tent map.
In this paper, the tent map is selected, whose formula is as follows:
s k + 1 = 10 s k / 7 s k < 0.7 10 ( 1 s k ) / 3 s k 0.7
when the initial value s 1 0 , 1 is randomly given, through Equation (10) iteration n 1 times, a chaotic sequence s 1 , s 2 , , s n can be generated. The initial agent can be generated by mapping the sequence to the solution, as follows:
x i = x i , min + ( x i , max x i , min ) s i
where x i , min represents the lower boundary of x in the i-th dimension, x i , max represents the upper boundary of x in the i-th dimension. At the time of initialization, all of the agents in the population are mapping according to formula Equations (10) and (11) in order to generate chaotic population.

3.2. Quasi-Opposition Learning

In the meta-heuristic algorithm, feasible solutions are obtained first, and then the solution space is searched for the optimal based on the fitness of the feasible solution. Various different meta-heuristic algorithms search in different ways, but all of them essentially have a random factor. In recent years, Opposition-Based Learning [35] has been widely used in meta-heuristic algorithms [21,27,29], which is essentially a method of replacing random search with symmetric search, which can greatly improve the search ability of the algorithm.
Although the opposite-learning method improves the search capability of the algorithm considerably, Opposition-Based Learning is too fixed and not very effective for tuning in a small space. The literature [31,36] proposes quasi-opposite learning, which adds a random factor to Opposition-Based Learning, and the resulting position is not a fixed symmetric position, but a random position between the central position and the symmetric position. Assuming x [ a , b ] , the expression for quasi-opposite learning is as follows:
x o = a + b 2 + r a + b 2 x
where r [ 0 , 1 ] is a random variable. If r [ 0 , 2 ] , quasi-opposite point theoretically has 50% chance falling within the symmetry point and the center, and 50% chance of falling outside the symmetry point. If it falls within the symmetry point, it is good for algorithm tuning to improve solution quality, while falling outside the symmetry point is more good for spatial search to improve convergence speed. If the variables x are multi-dimensional, then each dimension of x needs to separately execute Equation (12), i.e., where a and b are vectors, the multi-dimensional expression is as follows:
x o = a + b 2 + r a + b 2 x

3.3. Logarithm-Based Nonlinear Control Parameter

There is only one control parameter a in the standard WOA, as mentioned in Section 2.3 and Section 2.4. This parameter controls the proportion of exploitation and exploration, as shown in Equation (3). In the standard WOA, the parameter a varies linearly from 2 to 0, while the direct control of exploration and exploitation is A . From Equation (3) alone, exploration and exploitation each account for about 50% weight. At the beginning of WOA, exploration accounts for a greater proportion and, after beginning, the proportion of exploration gradually decreases, while that of exploitation gradually increases.
The proposed SWWOA algorithm, employs a chaos mechanism Section 3.1 and quasi-opposite learning Section 3.2 mechanism, which has already improved exploration more substantially, so nonlinear control mechanism based on logarithm with a greater proportion of exploitation is employed, which improves the tuning ability of the algorithm and ultimately improves the quality of the solution. The expression of parameter is as follows, and the comparison of two parameter mechanism is shown in Figure 2:
a = 2 log 10 1 + 99 t t max
where t denotes t-th iteration, t max denotes the max iteration times. As can be seen in Figure 2, the logarithm-based control parameter curve starts out very steep and declines rapidly, while it flattens out later. Overall, the tuning is performed with greater probability.

3.4. Single-Dimensional Swimming

It has been documented in [16], that the ABC algorithm has better optimization performance when compared with algorithms, such as PSO and GA. Most of the above improvements in this paper are dedicated to improving the search capability, in order to allow for the population to perform a fine-grained search in a narrow space near the optimal; this paper introduces the employed bee position update method in the ABC algorithm for single-dimensional swimming.
D d = C · X d * t X d t
X d t + 1 = X d * t A · D d
where d denotes a dimension, randomly generated for each agent, otherwise refer to Equations (1) and (2).

3.5. The Pseudo Code of SWWOA

SWWOA introduces four improvements to the standard WOA. These four improvements are interrelated. Chaos mechanisms and opposition-based learning are used in order to improve the spatial search capability; on top of this, nonlinear control parameter factors are introduced, which are used to apply more computations to tuning, and finally single-dimensional swimming are used inn order to search in narrow space and improve the solution quality. In Section 4, the ablation experiment is designed to discuss the impact of each improvement on the algorithm. The pseudo code of the standard SWWOA Algorithm 2 is as follows:
Algorithm 2 SWWOA
01 initialize maxIteration, popsize and parameter b
02 initialize chaos population and calculate fitness by Equations (10) and (11)
03 obtain the optimal agent
04 WHILE t<maxIteration DO
05  update a by Equation (14)
06  update A, C by Equations (3) and (4)
07  WHILE i<popsize DO
08   quasi-opposition learning x i o on agent i by Equation (13)
09   generate random number p [ 0 , 1 ]
10   IF p < 0.5 THEN
11    IF A < 1 THEN
12     generate random dimension d
13     update position of agent i by Equations (15) and (16)
14    ELSE
15     generate random agent rand
16     update position of agent i by Equation (9)
17    ENDIF
18   ELSE
19    update position of agent i by Equation (6)
20   ENDIF
21   compare x i o and x i , retaining the better agent
22   i = i + 1
23  ENDWHILE
24  update optimal agent if there is a better solution
25  t = t + 1
26 ENDWHILE
27 RETURN optimal agent

4. Experimental Results and Analysis

In order to verify the effectiveness of the proposed algorithm, we select other four algorithms (WOA [28], ABC [16], PSO [12], and OBCWOA [21]) to conduct comparative experiments on 20 well-known test functions. The language used for the implementation is C/C++, the compiler is gcc-4.8.5, the computer CPU is i3-9100, the memory is 16GB, and the CentOS-7.5 amd64 with kernel 3.10 operating system is used.
The PSO algorithm is a very famous meta-heuristic algorithm, which has a very stable performance and it is often used as a benchmark for meta-heuristic algorithm. The ABC algorithm is also a much studied algorithm with high solution quality, which is selected in this paper, because SWWOA draws on its scout bee update mechanism. SWWOA is a proposed algorithm based on standard WOA, and the standard WOA is also selected. In addition, the literature [21], which is almost the latest improvement algorithm of the standard WOA, it uses a chaos mechanism and quasi-opposition learning mechanism, which is similar to SWWOA, so the OBCWOA algorithm proposed in [21] is selected. Among the above four algorithms, ABC is special, in that it is mainly updated in single-dimensional way. For fairness, more iterations and a bigger population size are set for ABC. Table 1 shows the specific parameters of algorithm.

4.1. Test Functions

A series of large-scale and high-dimensional test functions presented in Table 2 are utilized to test algorithms’ performance. f 1 f 6 are unimodal-separable functions (abbreviated as US), mainly to check the convergence speed of the algorithm. f 7 f 12 are unimodal-nonseparable functions (abbreviated as UN), when compared with the US functions, more able to detect the search capability of the algorithm. f 13 f 15 are multimodal-separable functions (abbreviated as MS), more test algorithm out of the local optimum ability. f 16 f 20 are multimodal-nonseparable functions (abbreviated as MN); these types of functions are more complex and more reflective of the overall performance.

4.2. Numerical Analysis

In the comparative experiments, the problem dimensions are set to 20 (Table 3), 100 (Table 4), 200 (Table 5), 500 (Table 6), and 1000 (Table 7). Each algorithm is run independently on each function 20 times, and the optimal, average, and standard deviation are taken in tables.
The data in Table 3, Table 4, Table 5, Table 6 and Table 7 are the results of experiments conducted on different functional dimensions. The data avg denotes the mean value, which is mainly an indicator of the performance of the algorithm, while the best denotes the optimal result, which mainly reflects the accuracy of the algorithm, and std denotes the standard deviation of the data, which reflects the stability of the algorithm. The following section relies mainly on the avg value for evaluation. It is noteworthy that ABC and PSO are not available in Table 7, because, from the data presented in Table 3, Table 4, Table 5 and Table 6, these two algorithms have not achieved good results, and the gap between the results is too large, which loses the meaning of the analysis.
The data presented in Table 3 are obtained with the function dimension is set to 20. From the data presented in the table, we can see that f 1 f 6 are functions of type “US”, and SWWOA has obtained all of the optimal solutions on these functions. OBCWOA obtains the optimal solution on five functions except f 3 . Standard WOA obtains the optimal solution only on f 6 . ABC and PSO have not obtained the optimal solution. The average value of PSO is lower than ABC except f 17 , and the standard deviation is also smaller than ABC. From these results, SWWOA can obtain the optimal solution stably, which shows that SWWOA has better spatial search capability. f 7 f 12 are functions of type “UN”, which are more difficult to optimize on the basis of “US” functions. SWWOA obtains the optimal solutions on four functions ( f 9 f 12 ). OBCWOA obtains the optimal solution on three functions ( f 9 , f 11 , and f 12 ). Standard WOA, ABC, and PSO did not obtain the optimal solution. The overall ABC has the worst performance and SWWOA has the best performance. f 13 f 20 are multimodal functions. This type functions test the algorithm’s ability to jump out of the local optimum based on the space search. When compared with the unimodal function, its optimization is more difficult. SWWOA obtained the optimal solution on seven functions ( f 13 f 17 , f 19 , f 20 ), OBCWOA obtained the optimal solution on five functions ( f 14 f 17 , f 20 ), and the standard WOA obtained the optimal solution on four functions ( f 14 f 17 ), ABC and PSO have not obtained the optimal solution. In terms of performance, SWWOA does not obtain the optimal solution on f 7 , f 8 , f 18 and, in comparison, SWWOA does not obtain the optimal result only on f 8 . The overall situation is similar to that of the unimodal function.
The above analysis focuses on algorithm performance (avg), which is also the basis for other analyses. If the avg of an algorithm is poor, but the std is good, then it can only mean that the algorithm is stuck in local optimum and stagnant. Additionally, an algorithm best is good, but the avg is very bad. From the side, it can show that std must be bad, indicating that the algorithm lacks stability. On the basis of performance analysis, analyze the functions f 7 , f 8 , and f 18 , respectively.
Function f 7 : the “avg” order is SWWOA>PSO>OBCWOA>WOA>ABC, the “best” order is SWWOA>PSO>WOA>OBCWOA>ABC, the “std” order is PSO>SWWOA>ABC>OBCWOA>WOA. From the order view, the “avg” of ABC is the worst, but the “std” of ABC is approximately the best, which can indicate that ABC stably obtains a poor solution on this function, and it has actually fallen into a local optimum. For PSO, “avg” is approximately the best, “best” is also good, which indicates that the solution accuracy and performance of PSO is good. The “avg” and “best” of SWWOA are both best, “std” is also good, indicating that SWWOA can stably obtain better solutions.
Function f 8 : the “avg” order is WOA>PSO>OBCWOA>SWWOA>ABC, the “best” order is WOA>OBCWOA>PSO>SWWOA>ABC, the “std” order is PSO>ABC>WOA>SWWOA>OBCWOA. With the analysis of f 7 , ABC is still the worst. The “avg” and “best” of WOA are good and the best. The proposed SWWOA does not achieve the desired effect on this function. This is mainly because the optimal solution area of the function is a narrow and approximately flat area. When SWWOA is close to the optimal solution, it searches near the approximate optimal solution with a very small step, see Equation (14), which is biased towards tuning, instead of searching, so it fails to search the solution space more effectively.
Function f 18 : the “avg” order is SWWOA=OBCWOA>WOA>PSO>ABC, the “best” order is SWWOA=OBCWOA=WOA>PSO>ABC, and the “std” order is SWWOA=OBCWOA=PSO>WOA>ABC. This function is a multimodal function. Both SWWOA and OBCWOA fall into the local optimum, but, overall, SWWOA and OBCWOA should be the two algorithms with the best performance on this function.
Based on the above data, when considering “avg”, “best”, and “std”, the overall performance ranking of the algorithms is roughly SWWOA>OBCWOA>WOA>PSO>ABC.
From the perspective of algorithm comparison, Table 4, Table 5, Table 6 and Table 7 are similar to Table 3, so this paper does not analyze them in detail. The following is to select a function from “US”, “UN”, “MS”, and “MN”, respectively ( f 3 , f 7 , f 13 , f 19 ), and from Table 8 to perform algorithm performance between different dimensions.
The meta-heuristic algorithm has a great advantage, which is, it is not sensitive to the problem dimension. This feature is also an important indicator to measure the algorithm. The data presented in Table 8 are “avg” data. From the table, it can be seen that the average values of WOA, OBCWOA, and SWWOA almost all deteriorate with the dimensionality, indicating that the algorithms all have good robustness. Among them, WOA and OBCWOA still have some fluctuations, while the SWWOA steadily become worse with the increase of dimensions, correspondingly the best.
In order to more comprehensively compare the performance of algorithms, Table 9 compares algorithms according to different function types and dimensions. The data in the table are the number of times the algorithm won.
First of all, it can be seen from Table 9 that SWWOA is the algorithm with the most wins in various function types and dimensions, which shows that SWWOA has better adaptability. In addition, it can be seen from Table 9 that the effectiveness of the algorithm has little to do with the dimensionality of the function, but it has a greater relationship with the characteristics of the function itself. In general, if an algorithm wins on a function with lower dimension, it is highly likely that the algorithm can win on a higher dimension. When combining the data in Table 8, it can be concluded that the function dimension affects the accuracy of the solution, but whether an algorithm is effective on the function depends more on the characteristics of the function itself.

4.2.1. Test on Shifted Rotated Functions

Most of the benchmark functions shown in the Table 2 obtain the optimal value when x * = [ 0 , 0 , 0 , , 0 ] . This situation can easily lead to better performance of the algorithm while using the “average mechanism”. In view of this, all of the algorithms have been tested on eight shifted/rotated functions in CEC 2014 benchmark functions [37]. The algorithms use the same settings as above, and the test results are shown in Table 10 and Table 11.
From the data presented in Table 10, it can be seen that the proposed algorithm is still in the leading position on shifted rotated f 8 , rotated f 11 , rotated f 12 , and shifted f 14 , but, when compared with the above results, the advantage is not big. On Shifted Rotated f 14 , Shifted Rotated f 16 , Shifted Rotated f 17 , and Shifted Rotated f 18 , the results of all algorithms are exactly the same, and no optimal solution is obtained, which shows that all algorithms fall into local optima on these functions. The data presented in Table 11 and the data in Table 10 show almost the same results. The only difference is that the quality of the obtained solution decreases due to the increase of the function dimension. Combining the data from the two tables, it can be seen that SWWOA has certain advantages, but the advantages are not obvious, especially in the case of complex shifted rotated functions, the ability to jump out of the local optimum needs to be strengthened.

4.3. Wilcoxon’S Rank Sum Test Analysis

Almost all meta-heuristic algorithms include certain random factors. Wilcoxon’s rank sum test [38] is adopted in order to statistically reflect the superiority of the proposed algorithm. The significant differences between SWWOA and comparison algorithms are indicated by the p-values that were obtained from the Wilcoxon’s rank sum test, and the significance level is set at 0.05. When p-value < 0.05, it means that SWWOA has statistical advantages in solving problems as compared with the comparison algorithms. Table 12 shows the test results. It can be seen from Table 12 that most of the p-values are less than 0.001, which shows that SWWOA can solve the problem more effectively in most cases. In addition, there are some p-values that are equal to 1 in Table 12. This situation is because the two compared algorithms have obtained the optimal solution. There are only three p-values that are greater than 0.05 in Table 12. They are (1) SWWOA vs. OBCWOA on f 8 with n = 10 , (2) SWWOA vs. OBCWOA on f 8 with n = 100 , and (3) SWWOA vs. WOA on f 18 with n = 500 . This shows that, in these three cases, SWWOA does not have statistical advantages, which is consistent with the situation in Section 4.2.

4.4. Convergence Speed Comparison

Because different algorithms have different optimization mechanisms, for example, PSO has high global search capability. Therefore, only standard WOA (black), OBCWOA (blue) and SWWOA (green) are selected in this experiment, and the function dimension is set to 200. In Figure 3, the abscissa is the number of iterations, and the ordinate is the logarithm of the function value.
It can be clearly seen from Figure 3 that SWWOA converges to a better solution at a very high speed in all other functions, except for the two functions f 7 and f 8 . In the f 7 test, the convergence speed of SWWOA is not superior to that of WOA and OBCWOA, but it finally converges to the better solution. In the f 8 test, SWWOA and OBCWOA have a slight advantage in the convergence speed, but both have stagnated, and the final solution quality is not as good as WOA. Generally speaking, when compared with WOA and OBCWOA, SWWOA has a very high convergence speed.

4.5. Ablation Experiment

As mentioned above, this section selects f 3 , f 7 , f 13 , and f 19 for experiments, and the selected function’s dimension are set to 20. For the validity of the experiment, we run the algorithm independently on four functions 20 times, and count the results. The SWWOA has four improvements on the standard WOA, corresponding to the design algorithm WOA+Chaos (A1), A1+Quasi-Opposition-Learning (A2), A2+nonlinear-control-parameter (A3), A3+Single-dimensional swimming (SWWOA), and algorithms’ parameters refer to Table 1.
Table 13 shows the results of the ablation experiment. From the results of A1, the chaotic sequence alone did not improve the algorithm performance, but it decreased. This is because chaos is super random, and the initial population needs a stronger search mechanism to obtain better solution. From the results of A2, on the basis of the chaos initial population, quasi-opposition learning completely improves the algorithm’s spatial search ability. Except for the function f 7 , the other functions have obtained the optimal solutions. From the perspective of A3, the performance of the algorithm is once again reduced. This is because more calculations are used for tuning, but there is a lack of a tuning mechanism. Finally, single-dimensional swimming is added. From the data of f 7 , the accuracy of the algorithm is greatly improved. In summary, the chaotic sequence strengthens the chaos of the initial population, coupled with quasi-opposition learning, greatly improves the algorithm’s spatial search ability, even a little excessive. As a result, we put forward nonlinear control parameters, use excess calculations for tuning, and add unique single-dimensional swimming to achieve better results.

5. Conclusions

Based on the study of WOA, this paper proposes a modified WOA algorithm that is based on single-dimensional swimming (abbreviated as SWWOA). By proposing a chaotic sequence to generate the initial population, and using quasi-opposition learning on these foundations, the global search capability of SWWOA is greatly improved. At the same time, the original linear control parameter of WOA is improved, and nonlinear control parameter based on logarithm is used to balance The relationship between search and tuning, finally a single-dimensional swimming mechanism is proposed, which maximizes the tuning capability. The comparative experiments of 20 test functions in different dimensions show that the proposed algorithm can obtain high-quality solutions in few iterations and, at the same time, has strong stability and robustness. However, Section 4.2.1 presents a test conducted on complex shifted rotated functions. When compared with the comparison algorithm, although SWWOA has certain advantages, the advantages are not obvious, and there is obviously stagnation. Solving the stagnation problem and enhancing the ability of the algorithm to jump out of the local optimum are the directions of future work.

Author Contributions

The contributions of the authors are mentioned in this part. Conceptualization—P.D.; Methodology—P.D., H.Z.; Writing—original draft—P.D.; Writing—review and editing—P.D., W.C., N.L.; Funding acquisition—H.Z.; Formal analysis—H.Z.; Software—P.D., W.C.; Resources—N.L., J.L.; Project administration—J.L.; Investigation—N.L., W.C.; Data curation—W.C.; Supervision—J.L.; Visualization—W.C.; Validation—J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (NSFC) under grant number 61872187, 61371040.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yildiz, Y.E.; Topal, A.O. Large scale continuous global optimization based on micro differential evolution with local directional search. Inf. Sci. 2019, 477, 533–544. [Google Scholar] [CrossRef]
  2. Deng, H.; Peng, L.; Zhang, H.; Yang, B.; Chen, Z. Ranking-based biased learning swarm optimizer for large-scale optimization. Inf. Sci. 2019, 493, 120–137. [Google Scholar] [CrossRef]
  3. Han, F.; Jiang, J.; Ling, Q.H.; Su, B.Y. A survey on metaheuristic optimization for random single-hidden layer feedforward neural network. Neurocomputing 2019, 335, 261–273. [Google Scholar] [CrossRef]
  4. Boussaïd, I.; Lepagnot, J.; Siarry, P. A survey on optimization metaheuristics. Inf. Sci. 2013, 237, 82–117. [Google Scholar] [CrossRef]
  5. López-Campos, J.A.; Segade, A.; Casarejos, E.; Fernández, J.R.; Días, G.R. Hyperelastic characterization oriented to finite element applications using genetic algorithms. Adv. Eng. Softw. 2019, 133, 52–59. [Google Scholar] [CrossRef]
  6. Hussain, S.F.; Iqbal, S. CCGA: Co-similarity based Co-clustering using genetic algorithm. Appl. Soft Comput. 2018, 72, 30–42. [Google Scholar] [CrossRef]
  7. Assad, A.; Deep, K. A Hybrid Harmony search and Simulated Annealing algorithm for continuous optimization. Inf. Sci. 2018, 450, 246–266. [Google Scholar] [CrossRef]
  8. Morales-Castañeda, B.; Zaldívar, D.; Cuevas, E.; Maciel-Castillo, O.; Aranguren, I.; Fausto, F. An improved Simulated Annealing algorithm based on ancient metallurgy techniques. Appl. Soft Comput. 2019, 84, 105761. [Google Scholar] [CrossRef]
  9. Devi Priya, R.; Sivaraj, R.; Sasi Priyaa, N. Heuristically repopulated Bayesian ant colony optimization for treating missing values in large databases. Knowl. Based Syst. 2017, 133, 107–121. [Google Scholar] [CrossRef]
  10. Silva, B.N.; Han, K. Mutation operator integrated ant colony optimization based domestic appliance scheduling for lucrative demand side management. Future Gener. Comput. Syst. 2019, 100, 557–568. [Google Scholar] [CrossRef]
  11. Pedroso, D.M.; Bonyadi, M.R.; Gallagher, M. Parallel evolutionary algorithm for single and multi-objective optimisation: Differential evolution and constraints handling. Appl. Soft Comput. 2017, 61, 995–1012. [Google Scholar] [CrossRef]
  12. Kohler, M.; Vellasco, M.M.; Tanscheit, R. PSO+: A new particle swarm optimization algorithm for constrained problems. Appl. Soft Comput. 2019, 85, 105865. [Google Scholar] [CrossRef]
  13. Lynn, N.; Suganthan, P.N. Ensemble particle swarm optimizer. Appl. Soft Comput. 2017, 55, 533–548. [Google Scholar] [CrossRef]
  14. Zhang, X.; Liu, H.; Tu, L. A modified particle swarm optimization for multimodal multi-objective optimization. Eng. Appl. Artif. Intell. 2020, 95, 103905. [Google Scholar] [CrossRef]
  15. Wang, Y.; Yu, Y.; Gao, S.; Pan, H.; Yang, G. A hierarchical gravitational search algorithm with an effective gravitational constant. Swarm Evol. Comput. 2019, 46, 118–139. [Google Scholar] [CrossRef]
  16. Kong, D.; Chang, T.; Dai, W.; Wang, Q.; Sun, H. An improved artificial bee colony algorithm based on elite group guidance and combined breadth-depth search strategy. Inf. Sci. 2018, 442–443, 54–71. [Google Scholar] [CrossRef]
  17. Kumar, D.; Mishra, K. Co-variance guided Artificial Bee Colony. Appl. Soft Comput. 2018, 70, 86–107. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Cheng, S.; Shi, Y.; Gong, D.W.; Zhao, X. Cost-sensitive feature selection using two-archive multi-objective artificial bee colony algorithm. Expert Syst. Appl. 2019, 137, 46–58. [Google Scholar] [CrossRef]
  19. Sun, Y.; Yang, T.; Liu, Z. A whale optimization algorithm based on quadratic interpolation for high-dimensional global optimization problems. Appl. Soft Comput. 2019, 85, 105744. [Google Scholar] [CrossRef]
  20. Sun, Y.; Wang, X.; Chen, Y.; Liu, Z. A modified whale optimization algorithm for large-scale global optimization problems. Expert Syst. Appl. 2018, 114, 563–577. [Google Scholar] [CrossRef]
  21. Chen, H.; Li, W.; Yang, X. A whale optimization algorithm with chaos mechanism based on quasi-opposition for global optimization problems. Expert Syst. Appl. 2020, 158, 113612. [Google Scholar] [CrossRef]
  22. Ling, Y.; Zhou, Y.; Luo, Q. Lévy Flight Trajectory-Based Whale Optimization Algorithm for Global Optimization. IEEE Access 2017, 5, 6168–6186. [Google Scholar] [CrossRef]
  23. Ding, S.; An, Y.; Zhang, X.; Wu, F.; Xue, Y. Wavelet twin support vector machines based on glowworm swarm optimization. Neurocomputing 2017, 225, 157–163. [Google Scholar] [CrossRef]
  24. Chen, X.; Zhou, Y.; Tang, Z.; Luo, Q. A hybrid algorithm combining glowworm swarm optimization and complete 2-opt algorithm for spherical travelling salesman problems. Appl. Soft Comput. 2017, 58, 104–114. [Google Scholar] [CrossRef]
  25. Luo, K.; Zhao, Q. A binary grey wolf optimizer for the multidimensional knapsack problem. Appl. Soft Comput. 2019, 83, 105645. [Google Scholar] [CrossRef]
  26. Ezugwu, A.E.; Prayogo, D. Symbiotic organisms search algorithm: Theory, recent advances and applications. Expert Syst. Appl. 2019, 119, 184–209. [Google Scholar] [CrossRef]
  27. Truong, K.H.; Nallagownden, P.; Baharudin, Z.; Vo, D.N. A Quasi-Oppositional-Chaotic Symbiotic Organisms Search algorithm for global optimization problems. Appl. Soft Comput. 2019, 77, 567–583. [Google Scholar] [CrossRef]
  28. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  29. Gharehchopogh, F.S.; Gholizadeh, H. A comprehensive survey: Whale Optimization Algorithm and its applications. Swarm Evol. Comput. 2019, 48, 1–24. [Google Scholar] [CrossRef]
  30. Got, A.; Moussaoui, A.; Zouache, D. A guided population archive whale optimization algorithm for solving multiobjective optimization problems. Expert Syst. Appl. 2020, 141, 112972. [Google Scholar] [CrossRef]
  31. Elaziz, M.A.; Mirjalili, S. A hyper-heuristic for improving the initial population of whale optimization algorithm. Knowl. Based Syst. 2019, 172, 42–63. [Google Scholar] [CrossRef]
  32. Chen, H.; Yang, C.; Heidari, A.A.; Zhao, X. An efficient double adaptive random spare reinforced whale optimization algorithm. Expert Syst. Appl. 2020, 154, 113018. [Google Scholar] [CrossRef]
  33. Luo, J.; Chen, H.; Heidari, A.A.; Xu, Y.; Zhang, Q.; Li, C. Multi-strategy boosted mutative whale-inspired optimization approaches. Appl. Math. Model. 2019, 73, 109–123. [Google Scholar] [CrossRef]
  34. Agrawal, R.; Kaur, B.; Sharma, S. Quantum based Whale Optimization Algorithm for wrapper feature selection. Appl. Soft Comput. 2020, 89, 106092. [Google Scholar] [CrossRef]
  35. Tizhoosh, H.R. Opposition-Based Learning: A New Scheme for Machine Intelligence. In Proceedings of the International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (CIMCA-IAWTIC’06), Vienna, Austria, 28–30 November 2005; Volume 1, pp. 695–701. [Google Scholar] [CrossRef]
  36. Rahnamayan, S.; Tizhoosh, H.R.; Salama, M.M.A. Quasi-oppositional Differential Evolution. In Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007; pp. 2229–2236. [Google Scholar] [CrossRef]
  37. Garg, V.; Deep, K. Performance of Laplacian Biogeography-Based Optimization Algorithm on CEC 2014 continuous optimization benchmarks and camera calibration problem. Swarm Evol. Comput. 2016, 27, 132–144. [Google Scholar] [CrossRef]
  38. Drueta, S.; Ivi, S. Examination of benefits of personal fitness improvement dependent inertia for Particle Swarm Optimization. Soft Comput. 2017, 21, 3387–3400. [Google Scholar] [CrossRef]
Figure 1. Tent map and Logistic map.
Figure 1. Tent map and Logistic map.
Symmetry 12 01892 g001
Figure 2. Two parameter mechanism with t m a x = 100 .
Figure 2. Two parameter mechanism with t m a x = 100 .
Symmetry 12 01892 g002
Figure 3. Convergence speed comparison with n = 200.
Figure 3. Convergence speed comparison with n = 200.
Symmetry 12 01892 g003
Table 1. Parameter setting.
Table 1. Parameter setting.
AlgorithmParameter Settings
ABC p o p s i z e = 60 , t max = 2000 , t r i a l = 20
PSO p o p s i z e = 30 , t max = 1000 , w = 0.9 ( 0.9 0.2 ) t / t max ,
c 1 = 1.5 , c 2 = 1.5 , V [ 0.5 , 0.5 ]
WOA p o p s i z e = 30 , t max = 1000 , a = 2 1 t / t max ,
OBCWOA p o p s i z e = 30 , t max = 1000 , a = 2 1 t / t max , a log i s t i c = 4 , b = 1
SWWOA p o p s i z e = 30 , t max = 1000 , a = 2 log 10 1 + 99 t / t max , b = 1
Table 2. Test functions.
Table 2. Test functions.
NameEquationRangeType
Sphere f 1 ( x ) = i = 1 n x i 2 [ 100 , 100 ] n US
Sum Squares f 2 ( x ) = i = 1 n i x i 2 [ 10 , 10 ] n US
Schwefel 2.21 f 3 ( x ) = max | x i | , 1 i n [ 100 , 100 ] n US
Powell Sum f 4 ( x ) = i = 1 n | x i | i + 1 [ 1 , 1 ] n US
Quartic f 5 ( x ) = i = 1 n i x i 4 [ 1.28 , 1.28 ] n US
Step f 6 ( x ) = i = 1 n x i + 0.5 2 [ 100 , 100 ] n US
Zakharov f 7 ( x ) = i = 1 n x i 2 + ( i = 1 n 0.5 i x i ) 2 + ( i = 1 n 0.5 i x ) 4 [ 5 , 10 ] n UN
Rosenbrock f 8 ( x ) = i = 1 n 1 [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] [ 30 , 30 ] n UN
Schwefel 1.2 f 9 ( x ) = i = 1 n ( j = 1 i x j ) 2 [ 100 , 100 ] n UN
Schwefel 2.22 f 10 ( x ) = i = 1 n | x i | + i = 1 n | x i | [ 10 , 10 ] n UN
Discus f 11 ( x ) = 10 6 x 1 2 + i = 2 n x i 6 [ 1 , 1 ] n UN
Cigar f 12 ( x ) = x 1 2 + 10 6 i = 2 n x i 6 [ 100 , 100 ] n UN
Alpine f 13 ( x ) = x i sin x i + 0.1 x i [ 10 , 10 ] n MS
Rastrigin f 14 ( x ) = i = 1 n ( x i 2 10 cos 2 π x i + 10 ) [ 5.12 , 5.12 ] n MS
Bohachevsky f 15 ( x ) = i = 1 n 1 ( x i 2 + 2 x i + 1 2 0.3 cos 3 π x i 0.4 cos 4 π x i + 1 + 0.7 ) [ 50 , 50 ] n MS
Griewank f 16 ( x ) = i = 1 n x i 2 / 4000 i = 1 n cos ( x i / i ) + 1 [ 60 , 60 ] n MN
Weierstrass f 17 ( x ) = i = 1 n k = 0 20 { [ cos ( 2 π 3 k ( x i + 0.5 ) ) cos ( 2 π 3 k · 0.5 ) ] / 2 k } [ 0.5 , 0.5 ] n MN
Ackley f 18 ( x ) = 20 exp ( 0.2 i = 1 n x i 2 n ) exp i = 1 n cos 2 π x i n + 20 + e [ 32 , 32 ] n MN
Schaffer f 19 ( x ) = 0.5 + [ sin 2 ( i = 1 n x i 2 ) 0.5 0.5 ] / ( 1 + i = 1 n x i 2 / 1000 ) 2 [ 100 , 100 ] n MN
Salomon f 20 ( x ) = 1 cos ( 2 π i = 1 n x i 2 ) + i = 1 n x i 2 / 10 [ 100 , 100 ] n MN
Table 3. Comparison results for 20 test functions with n = 20.
Table 3. Comparison results for 20 test functions with n = 20.
Function ABCPSOWOAOBCWOASWWOA
best8.38 × 10 07 1.29 × 10 39 1.86 × 10 185 0.00 × 10 + 00 0.00 × 10 + 00
f 1 avg8.38 × 10 07 1.29 × 10 39 1.15 × 10 162 0.00 × 10 + 00 0.00 × 10 + 00
std4.74 × 10 22 1.46 × 10 54 2.19 × 10 161 0.00 × 10 + 00 0.00 × 10 + 00
best1.32 × 10 08 2.30 × 10 40 4.19 × 10 191 0.00 × 10 + 00 0.00 × 10 + 00
f 2 avg1.46 × 10 04 9.61 × 10 39 1.81 × 10 167 0.00 × 10 + 00 0.00 × 10 + 00
std4.80 × 10 04 3.80 × 10 38 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best7.24 × 10 + 01 4.18 × 10 05 1.34 × 10 30 1.96 × 10 218 0.00 × 10 + 00
f 3 avg7.24 × 10 + 01 4.18 × 10 05 1.58 × 10 21 2.87 × 10 200 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 3.00 × 10 20 0.00 × 10 + 00 0.00 × 10 + 00
best1.10 × 10 05 4.33 × 10 90 1.49 × 10 267 0.00 × 10 + 00 0.00 × 10 + 00
f 4 avg1.10 × 10 05 4.33 × 10 90 1.81 × 10 226 0.00 × 10 + 00 0.00 × 10 + 00
std1.52 × 10 20 7.80 × 10 105 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.05 × 10 12 1.22 × 10 72 7.63 × 10 303 0.00 × 10 + 00 0.00 × 10 + 00
f 5 avg5.76 × 10 09 1.22 × 10 72 5.52 × 10 263 0.00 × 10 + 00 0.00 × 10 + 00
std4.46 × 10 08 2.25 × 10 87 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.97 × 10 + 02 4.20 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 6 avg1.97 × 10 + 02 4.20 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best3.07 × 10 + 02 1.06 × 10 09 4.91 × 10 08 3.38 × 10 02 6.94 × 10 19
f 7 avg3.07 × 10 + 02 1.06 × 10 09 6.87 × 10 + 01 3.15 × 10 + 01 2.48 × 10 15
std5.08 × 10 13 1.85 × 10 24 3.03 × 10 + 02 1.31 × 10 + 02 2.68 × 10 14
best3.79 × 10 + 02 1.14 × 10 + 00 3.37 × 10 03 1.44 × 10 02 1.17 × 10 + 01
f 8 avg3.79 × 10 + 02 1.14 × 10 + 00 3.59 × 10 01 7.16 × 10 + 00 1.31 × 10 + 01
std5.08 × 10 13 0.00 × 10 + 00 2.31 × 10 + 00 3.84 × 10 + 01 5.46 × 10 + 00
best1.45 × 10 + 04 4.83 × 10 05 2.23 × 10 43 0.00 × 10 + 00 0.00 × 10 + 00
f 9 avg1.47 × 10 + 04 4.83 × 10 05 8.77 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00
std1.02 × 10 + 03 0.00 × 10 + 00 1.53 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00
best1.54 × 10 02 2.52 × 10 10 3.54 × 10 120 2.10 × 10 239 0.00 × 10 + 00
f 10 avg1.54 × 10 02 2.52 × 10 10 1.68 × 10 112 3.03 × 10 225 0.00 × 10 + 00
std3.10 × 10 17 0.00 × 10 + 00 3.21 × 10 111 0.00 × 10 + 00 0.00 × 10 + 00
best1.12 × 10 02 2.36 × 10 85 1.70 × 10 304 0.00 × 10 + 00 0.00 × 10 + 00
f 11 avg1.12 × 10 02 2.36 × 10 85 2.04 × 10 233 0.00 × 10 + 00 0.00 × 10 + 00
std1.55 × 10 17 3.83 × 10 100 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best3.12 × 10 + 07 1.24 × 10 72 3.60 × 10 303 0.00 × 10 + 00 0.00 × 10 + 00
f 12 avg3.12 × 10 + 07 1.16 × 10 66 3.12 × 10 251 0.00 × 10 + 00 0.00 × 10 + 00
std1.67 × 10 08 1.73 × 10 66 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.64 × 10 + 00 4.26 × 10 09 3.48 × 10 122 6.94 × 10 241 0.00 × 10 + 00
f 13 avg2.75 × 10 + 00 4.26 × 10 09 3.80 × 10 105 1.47 × 10 231 0.00 × 10 + 00
std3.19 × 10 01 7.40 × 10 24 7.35 × 10 104 0.00 × 10 + 00 0.00 × 10 + 00
best1.60 × 10 + 01 2.49 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 14 avg2.81 × 10 + 01 2.49 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std3.12 × 10 + 01 3.18 × 10 14 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best3.85 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 15 avg1.99 × 10 01 1.39 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std2.01 × 10 + 00 1.43 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best3.13 × 10 02 3.33 × 10 16 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 16 avg1.23 × 10 01 3.33 × 10 16 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std3.71 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.02 × 10 02 5.34 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 17 avg7.10 × 10 01 1.23 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std4.93 × 10 + 00 5.42 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best6.85 × 10 01 4.31 × 10 14 4.44 × 10 16 4.44 × 10 16 4.44 × 10 16
f 18 avg2.02 × 10 + 00 4.31 × 10 14 2.40 × 10 15 4.44 × 10 16 4.44 × 10 16
std9.08 × 10 + 00 0.00 × 10 + 00 7.90 × 10 15 0.00 × 10 + 00 0.00 × 10 + 00
best4.99 × 10 01 4.15 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 19 avg4.99 × 10 01 4.28 × 10 01 8.26 × 10 03 4.86 × 10 04 0.00 × 10 + 00
std4.97 × 10 16 9.25 × 10 02 1.55 × 10 02 9.47 × 10 03 0.00 × 10 + 00
best1.21 × 10 + 01 4.50 × 10 + 00 4.25 × 10 86 0.00 × 10 + 00 0.00 × 10 + 00
f 20 avg1.21 × 10 + 01 4.50 × 10 + 00 8.49 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00
std1.59 × 10 14 3.97 × 10 15 1.59 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00
Table 4. Comparison results for 20 test functions with n = 100.
Table 4. Comparison results for 20 test functions with n = 100.
Function ABCPSOWOAOBCWOASWWOA
best4.13 × 10 + 02 7.69 × 10 02 3.61 × 10 181 0.00 × 10 + 00 0.00 × 10 + 00
f 1 avg7.82 × 10 + 02 8.34 × 10 02 2.15 × 10 166 0.00 × 10 + 00 0.00 × 10 + 00
std4.96 × 10 + 03 2.88 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best3.16 × 10 + 03 8.59 × 10 01 2.13 × 10 194 0.00 × 10 + 00 0.00 × 10 + 00
f 2 avg3.16 × 10 + 03 8.76 × 10 01 1.02 × 10 162 0.00 × 10 + 00 0.00 × 10 + 00
std4.07 × 10 12 7.57 × 10 02 1.47 × 10 161 0.00 × 10 + 00 0.00 × 10 + 00
best9.45 × 10 + 01 5.55 × 10 + 00 3.63 × 10 36 4.78 × 10 196 0.00 × 10 + 00
f 3 avg9.46 × 10 + 01 5.80 × 10 + 00 9.67 × 10 19 4.43 × 10 185 0.00 × 10 + 00
std1.17 × 10 + 00 1.23 × 10 + 00 1.88 × 10 17 0.00 × 10 + 00 0.00 × 10 + 00
best1.00 × 10 02 1.55 × 10 26 3.57 × 10 266 0.00 × 10 + 00 0.00 × 10 + 00
f 4 avg1.23 × 10 01 9.91 × 10 22 4.61 × 10 221 0.00 × 10 + 00 0.00 × 10 + 00
std5.45 × 10 01 7.22 × 10 21 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.20 × 10 02 7.38 × 10 04 8.34 × 10 296 0.00 × 10 + 00 0.00 × 10 + 00
f 5 avg3.30 × 10 02 1.73 × 10 03 4.91 × 10 262 0.00 × 10 + 00 0.00 × 10 + 00
std1.57 × 10 01 3.24 × 10 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.01 × 10 + 04 2.90 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 6 avg2.01 × 10 + 04 3.05 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 6.23 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.48 × 10 + 03 1.64 × 10 + 03 6.41 × 10 + 02 3.61 × 10 + 02 1.15 × 10 11
f 7 avg1.39 × 10 + 08 2.11 × 10 + 03 1.56 × 10 + 03 2.07 × 10 + 03 3.08 × 10 05
std2.08 × 10 + 08 2.35 × 10 + 03 1.27 × 10 + 03 4.08 × 10 + 03 3.56 × 10 04
best8.24 × 10 + 06 1.00 × 10 + 02 5.59 × 10 02 3.17 × 10 01 9.50 × 10 + 01
f 8 avg8.24 × 10 + 06 1.78 × 10 + 02 1.71 × 10 + 00 5.70 × 10 + 01 9.75 × 10 + 01
std0.00 × 10 + 00 2.57 × 10 + 02 1.31 × 10 + 01 2.05 × 10 + 02 5.28 × 10 + 00
best2.79 × 10 + 05 1.63 × 10 + 03 7.36 × 10 04 0.00 × 10 + 00 0.00 × 10 + 00
f 9 avg3.18 × 10 + 05 2.05 × 10 + 03 1.83 × 10 + 05 0.00 × 10 + 00 0.00 × 10 + 00
std2.37 × 10 + 05 1.53 × 10 + 03 8.87 × 10 + 05 0.00 × 10 + 00 0.00 × 10 + 00
best3.55 × 10 + 01 1.21 × 10 + 00 2.42 × 10 120 3.33 × 10 227 0.00 × 10 + 00
f 10 avg4.21 × 10 + 01 1.50 × 10 + 00 5.95 × 10 106 1.07 × 10 213 0.00 × 10 + 00
std1.22 × 10 + 01 1.17 × 10 + 00 1.16 × 10 104 0.00 × 10 + 00 0.00 × 10 + 00
best5.09 × 10 02 1.00 × 10 + 00 3.20 × 10 316 0.00 × 10 + 00 0.00 × 10 + 00
f 11 avg5.09 × 10 02 1.60 × 10 + 00 7.45 × 10 253 0.00 × 10 + 00 0.00 × 10 + 00
std6.21 × 10 17 2.19 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.00 × 10 + 11 1.07 × 10 + 02 1.75 × 10 315 0.00 × 10 + 00 0.00 × 10 + 00
f 12 avg1.84 × 10 + 13 1.17 × 10 + 02 1.78 × 10 245 0.00 × 10 + 00 0.00 × 10 + 00
std1.26 × 10 + 13 3.28 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.63 × 10 + 01 7.77 × 10 02 1.16 × 10 121 2.16 × 10 232 0.00 × 10 + 00
f 13 avg3.56 × 10 + 01 2.08 × 10 01 2.25 × 10 110 3.78 × 10 214 0.00 × 10 + 00
std2.78 × 10 + 01 4.83 × 10 01 2.49 × 10 109 0.00 × 10 + 00 0.00 × 10 + 00
best3.36 × 10 + 02 1.02 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 14 avg4.44 × 10 + 02 1.21 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std2.78 × 10 + 02 6.84 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.66 × 10 + 02 2.68 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 15 avg4.07 × 10 + 02 3.26 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.11 × 10 + 03 1.84 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.05 × 10 + 00 4.12 × 10 04 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 16 avg1.80 × 10 + 00 5.91 × 10 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std4.11 × 10 + 00 2.53 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best8.03 × 10 + 00 7.44 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 17 avg2.27 × 10 + 01 9.07 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std4.53 × 10 + 01 4.14 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best7.75 × 10 + 00 3.46 × 10 + 00 4.44 × 10 16 4.44 × 10 16 4.44 × 10 16
f 18 avg9.89 × 10 + 00 5.07 × 10 + 00 1.69 × 10 15 4.44 × 10 16 4.44 × 10 16
std6.97 × 10 + 00 7.40 × 10 + 00 7.58 × 10 15 0.00 × 10 + 00 0.00 × 10 + 00
best5.00 × 10 01 4.96 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 19 avg5.00 × 10 01 4.96 × 10 01 9.15 × 10 03 4.37 × 10 03 0.00 × 10 + 00
std3.93 × 10 07 1.80 × 10 03 3.36 × 10 02 2.16 × 10 02 0.00 × 10 + 00
best4.87 × 10 + 01 7.50 × 10 + 00 4.90 × 10 89 0.00 × 10 + 00 0.00 × 10 + 00
f 20 avg4.87 × 10 + 01 8.38 × 10 + 00 5.49 × 10 02 2.00 × 10 02 0.00 × 10 + 00
std0.00 × 10 + 00 4.82 × 10 + 00 2.22 × 10 01 1.79 × 10 01 0.00 × 10 + 00
Table 5. Comparison results for 20 test functions with n = 200.
Table 5. Comparison results for 20 test functions with n = 200.
Function ABCPSOWOAOBCWOASWWOA
best4.69 × 10 + 02 5.31 × 10 + 00 1.62 × 10 183 0.00 × 10 + 00 0.00 × 10 + 00
f 1 avg1.41 × 10 + 04 1.08 × 10 + 01 1.81 × 10 160 0.00 × 10 + 00 0.00 × 10 + 00
std1.06 × 10 + 05 1.68 × 10 + 01 3.45 × 10 159 0.00 × 10 + 00 0.00 × 10 + 00
best2.02 × 10 + 02 1.36 × 10 + 02 5.00 × 10 179 0.00 × 10 + 00 0.00 × 10 + 00
f 2 avg3.82 × 10 + 02 1.47 × 10 + 02 1.83 × 10 164 0.00 × 10 + 00 0.00 × 10 + 00
std7.26 × 10 + 02 5.92 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best9.80 × 10 + 01 8.66 × 10 + 00 8.53 × 10 31 1.09 × 10 193 0.00 × 10 + 00
f 3 avg9.83 × 10 + 01 9.61 × 10 + 00 1.52 × 10 18 6.38 × 10 179 0.00 × 10 + 00
std1.22 × 10 + 00 4.03 × 10 + 00 2.93 × 10 17 0.00 × 10 + 00 0.00 × 10 + 00
best2.52 × 10 01 4.44 × 10 16 8.13 × 10 271 0.00 × 10 + 00 0.00 × 10 + 00
f 4 avg1.11 × 10 + 00 6.11 × 10 14 2.03 × 10 227 0.00 × 10 + 00 0.00 × 10 + 00
std1.43 × 10 + 00 4.36 × 10 13 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best3.45 × 10 02 2.44 × 10 + 00 2.17 × 10 303 0.00 × 10 + 00 0.00 × 10 + 00
f 5 avg3.35 × 10 01 7.11 × 10 + 00 1.58 × 10 270 0.00 × 10 + 00 0.00 × 10 + 00
std1.21 × 10 + 00 2.71 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.76 × 10 + 04 1.38 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 6 avg2.16 × 10 + 04 3.21 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.98 × 10 + 04 6.05 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.00 × 10 + 11 6.13 × 10 + 03 2.23 × 10 + 03 8.80 × 10 + 02 4.50 × 10 08
f 7 avg1.04 × 10 + 13 6.92 × 10 + 03 3.37 × 10 + 03 4.27 × 10 + 03 1.46 × 10 + 00
std3.24 × 10 + 13 3.15 × 10 + 03 2.63 × 10 + 03 6.43 × 10 + 03 2.82 × 10 + 01
best2.17 × 10 + 05 1.20 × 10 + 03 8.76 × 10 02 2.52 × 10 + 00 1.96 × 10 + 02
f 8 avg2.19 × 10 + 05 1.34 × 10 + 03 1.23 × 10 + 01 1.30 × 10 + 02 1.98 × 10 + 02
std2.38 × 10 + 04 6.42 × 10 + 02 1.90 × 10 + 02 3.72 × 10 + 02 2.71 × 10 + 00
best1.52 × 10 + 06 1.43 × 10 + 04 1.55 × 10 + 04 0.00 × 10 + 00 0.00 × 10 + 00
f 9 avg1.52 × 10 + 06 2.08 × 10 + 04 1.59 × 10 + 06 0.00 × 10 + 00 0.00 × 10 + 00
std2.08 × 10 09 3.22 × 10 + 04 6.54 × 10 + 06 0.00 × 10 + 00 0.00 × 10 + 00
best9.62 × 10 + 00 1.49 × 10 + 01 8.18 × 10 118 5.75 × 10 223 0.00 × 10 + 00
f 10 avg1.38 × 10 + 02 1.68 × 10 + 01 3.49 × 10 103 1.28 × 10 209 0.00 × 10 + 00
std5.11 × 10 + 02 8.13 × 10 + 00 6.80 × 10 102 0.00 × 10 + 00 0.00 × 10 + 00
best9.11 × 10 + 00 3.01 × 10 + 00 4.75 × 10 302 0.00 × 10 + 00 0.00 × 10 + 00
f 11 avg1.04 × 10 + 01 4.46 × 10 + 00 3.93 × 10 265 0.00 × 10 + 00 0.00 × 10 + 00
std6.48 × 10 + 00 1.14 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.00 × 10 + 11 2.42 × 10 + 05 2.19 × 10 311 0.00 × 10 + 00 0.00 × 10 + 00
f 12 avg1.64 × 10 + 17 8.05 × 10 + 05 3.09 × 10 243 0.00 × 10 + 00 0.00 × 10 + 00
std1.27 × 10 + 18 2.76 × 10 + 06 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best5.59 × 10 + 01 2.33 × 10 + 00 1.24 × 10 117 6.32 × 10 225 0.00 × 10 + 00
f 13 avg9.31 × 10 + 01 3.09 × 10 + 00 5.09 × 10 108 6.65 × 10 210 0.00 × 10 + 00
std9.07 × 10 + 01 2.96 × 10 + 00 9.16 × 10 107 0.00 × 10 + 00 0.00 × 10 + 00
best7.21 × 10 + 02 2.43 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 14 avg1.06 × 10 + 03 2.75 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std6.07 × 10 + 02 1.24 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.45 × 10 + 02 1.01 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 15 avg3.57 × 10 + 04 1.21 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.34 × 10 + 05 4.76 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.00 × 10 01 3.33 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 16 avg2.08 × 10 + 00 4.37 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.55 × 10 + 01 3.78 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best8.81 × 10 + 01 2.41 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 17 avg1.46 × 10 + 02 2.56 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.38 × 10 + 02 4.11 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.23 × 10 + 01 5.35 × 10 + 00 4.44 × 10 16 4.44 × 10 16 4.44 × 10 16
f 18 avg1.45 × 10 + 01 6.17 × 10 + 00 1.51 × 10 15 4.44 × 10 16 4.44 × 10 16
std7.17 × 10 + 00 3.27 × 10 + 00 7.28 × 10 15 0.00 × 10 + 00 0.00 × 10 + 00
best5.00 × 10 01 4.99 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 19 avg5.00 × 10 01 4.99 × 10 01 7.77 × 10 03 6.32 × 10 03 0.00 × 10 + 00
std2.27 × 10 07 4.18 × 10 04 1.74 × 10 02 2.07 × 10 02 0.00 × 10 + 00
best7.58 × 10 + 01 1.29 × 10 + 01 1.41 × 10 90 0.00 × 10 + 00 0.00 × 10 + 00
f 20 avg7.68 × 10 + 01 1.32 × 10 + 01 6.49 × 10 02 4.99 × 10 03 0.00 × 10 + 00
std4.44 × 10 + 00 1.34 × 10 + 00 2.13 × 10 01 9.73 × 10 02 0.00 × 10 + 00
Table 6. Comparison results for 20 test functions with n = 500.
Table 6. Comparison results for 20 test functions with n = 500.
Function ABCPSOWOAOBCWOASWWOA
best1.33 × 10 + 05 1.56 × 10 + 02 9.48 × 10 181 0.00 × 10 + 00 0.00 × 10 + 00
f 1 avg9.21 × 10 + 05 2.00 × 10 + 02 2.22 × 10 160 0.00 × 10 + 00 0.00 × 10 + 00
std2.59 × 10 + 06 1.69 × 10 + 02 4.31 × 10 159 0.00 × 10 + 00 0.00 × 10 + 00
best7.75 × 10 + 04 7.47 × 10 + 03 1.19 × 10 182 0.00 × 10 + 00 0.00 × 10 + 00
f 2 avg7.63 × 10 + 05 9.79 × 10 + 03 1.11 × 10 158 0.00 × 10 + 00 0.00 × 10 + 00
std2.36 × 10 + 06 4.56 × 10 + 03 2.06 × 10 157 0.00 × 10 + 00 0.00 × 10 + 00
best9.90 × 10 + 01 1.13 × 10 + 01 1.06 × 10 29 9.88 × 10 190 0.00 × 10 + 00
f 3 avg9.92 × 10 + 01 1.28 × 10 + 01 1.92 × 10 20 7.55 × 10 177 0.00 × 10 + 00
std7.70 × 10 01 3.99 × 10 + 00 2.91 × 10 19 0.00 × 10 + 00 0.00 × 10 + 00
best1.99 × 10 + 00 6.44 × 10 12 6.58 × 10 292 0.00 × 10 + 00 0.00 × 10 + 00
f 4 avg2.72 × 10 + 00 1.38 × 10 05 2.00 × 10 217 0.00 × 10 + 00 0.00 × 10 + 00
std1.68 × 10 + 00 2.43 × 10 04 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best4.63 × 10 + 02 4.76 × 10 + 02 2.83 × 10 297 0.00 × 10 + 00 0.00 × 10 + 00
f 5 avg1.71 × 10 + 03 1.01 × 10 + 03 6.00 × 10 258 0.00 × 10 + 00 0.00 × 10 + 00
std7.30 × 10 + 03 1.92 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.22 × 10 + 05 1.06 × 10 + 04 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 6 avg1.34 × 10 + 05 1.43 × 10 + 04 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std5.16 × 10 + 04 1.14 × 10 + 04 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.00 × 10 + 11 1.98 × 10 + 04 7.15 × 10 + 03 2.55 × 10 + 03 4.90 × 10 10
f 7 avg5.55 × 10 + 17 9.49 × 10 + 09 8.17 × 10 + 03 9.20 × 10 + 03 1.07 × 10 + 03
std1.29 × 10 + 18 1.27 × 10 + 11 4.39 × 10 + 03 1.57 × 10 + 04 1.64 × 10 + 04
best1.13 × 10 + 08 1.81 × 10 + 04 1.33 × 10 01 3.36 × 10 + 00 4.98 × 10 + 02
f 8 avg1.64 × 10 + 08 2.43 × 10 + 04 5.96 × 10 + 00 2.79 × 10 + 02 4.98 × 10 + 02
std3.09 × 10 + 08 2.06 × 10 + 04 2.59 × 10 + 01 9.76 × 10 + 02 4.21 × 10 01
best6.72 × 10 + 06 1.25 × 10 + 05 1.07 × 10 + 06 0.00 × 10 + 00 0.00 × 10 + 00
f 9 avg7.41 × 10 + 06 2.03 × 10 + 05 1.12 × 10 + 07 0.00 × 10 + 00 0.00 × 10 + 00
std2.54 × 10 + 06 1.71 × 10 + 05 3.24 × 10 + 07 0.00 × 10 + 00 0.00 × 10 + 00
best3.55 × 10 + 02 1.18 × 10 + 02 3.17 × 10 117 6.23 × 10 220 0.00 × 10 + 00
f 10 avg7.67 × 10 + 21 1.31 × 10 + 02 2.87 × 10 105 4.90 × 10 209 0.00 × 10 + 00
std6.86 × 10 + 22 4.09 × 10 + 01 5.57 × 10 104 0.00 × 10 + 00 0.00 × 10 + 00
best3.98 × 10 + 01 6.66 × 10 + 00 2.48 × 10 299 0.00 × 10 + 00 0.00 × 10 + 00
f 11 avg6.83 × 10 + 01 1.39 × 10 + 01 4.53 × 10 246 0.00 × 10 + 00 0.00 × 10 + 00
std1.28 × 10 + 02 2.42 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.00 × 10 + 11 4.45 × 10 + 09 1.78 × 10 307 0.00 × 10 + 00 0.00 × 10 + 00
f 12 avg3.03 × 10 + 18 8.45 × 10 + 09 9.28 × 10 262 0.00 × 10 + 00 0.00 × 10 + 00
std1.71 × 10 + 19 1.11 × 10 + 10 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.35 × 10 + 02 4.33 × 10 + 01 5.30 × 10 118 1.88 × 10 219 0.00 × 10 + 00
f 13 avg4.07 × 10 + 02 5.50 × 10 + 01 4.57 × 10 103 4.43 × 10 208 0.00 × 10 + 00
std4.95 × 10 + 02 2.99 × 10 + 01 8.91 × 10 102 0.00 × 10 + 00 0.00 × 10 + 00
best2.51 × 10 + 03 1.34 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 14 avg3.60 × 10 + 03 1.56 × 10 + 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std2.83 × 10 + 03 4.47 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.19 × 10 + 04 7.42 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 15 avg1.47 × 10 + 05 8.53 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std8.77 × 10 + 05 3.16 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best4.25 × 10 + 00 3.27 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 16 avg3.12 × 10 + 01 3.87 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.56 × 10 + 02 1.77 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.55 × 10 + 02 7.62 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 17 avg4.68 × 10 + 02 7.95 × 10 + 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std9.75 × 10 + 02 6.32 × 10 + 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.62 × 10 + 01 6.82 × 10 + 00 4.44 × 10 16 4.44 × 10 16 4.44 × 10 16
f 18 avg1.70 × 10 + 01 7.55 × 10 + 00 1.69 × 10 15 4.44 × 10 16 4.44 × 10 16
std2.73 × 10 + 00 2.23 × 10 + 00 7.58 × 10 15 0.00 × 10 + 00 0.00 × 10 + 00
best5.00 × 10 01 5.00 × 10 01 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 19 avg5.00 × 10 01 5.00 × 10 01 6.80 × 10 03 4.37 × 10 03 0.00 × 10 + 00
std5.05 × 10 09 8.62 × 10 05 1.99 × 10 02 2.16 × 10 02 0.00 × 10 + 00
best1.23 × 10 + 02 1.77 × 10 + 01 3.22 × 10 92 0.00 × 10 + 00 0.00 × 10 + 00
f 20 avg1.24 × 10 + 02 1.93 × 10 + 01 6.49 × 10 02 3.50 × 10 02 0.00 × 10 + 00
std2.01 × 10 + 00 3.14 × 10 + 00 2.13 × 10 01 2.13 × 10 01 0.00 × 10 + 00
Table 7. Comparison results for 20 test functions with n = 1000.
Table 7. Comparison results for 20 test functions with n = 1000.
Function WOAOBCWOASWWOA
best1.40 × 10 181 0.00 × 10 + 00 0.00 × 10 + 00
f 1 avg1.07 × 10 157 0.00 × 10 + 00 0.00 × 10 + 00
std2.09 × 10 156 0.00 × 10 + 00 0.00 × 10 + 00
best1.18 × 10 179 0.00 × 10 + 00 0.00 × 10 + 00
f 2 avg3.41 × 10 165 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.41 × 10 27 7.14 × 10 189 0.00 × 10 + 00
f 3 avg9.83 × 10 21 4.49 × 10 171 0.00 × 10 + 00
std1.50 × 10 19 0.00 × 10 + 00 0.00 × 10 + 00
best6.62 × 10 291 0.00 × 10 + 00 0.00 × 10 + 00
f 4 avg2.29 × 10 224 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best5.30 × 10 297 0.00 × 10 + 00 0.00 × 10 + 00
f 5 avg2.08 × 10 269 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 6 avg0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.11 × 10 + 04 4.84 × 10 + 03 9.94 × 10 04
f 7 avg1.59 × 10 + 04 1.84 × 10 + 04 1.31 × 10 + 04
std8.17 × 10 + 03 2.66 × 10 + 04 6.18 × 10 + 04
best3.18 × 10 02 1.50 × 10 + 00 9.98 × 10 + 02
f 8 avg8.18 × 10 + 01 6.60 × 10 + 02 9.98 × 10 + 02
std9.74 × 10 + 02 1.87 × 10 + 03 4.41 × 10 01
best4.92 × 10 + 06 0.00 × 10 + 00 0.00 × 10 + 00
f 9 avg7.18 × 10 + 07 0.00 × 10 + 00 0.00 × 10 + 00
std2.33 × 10 + 08 0.00 × 10 + 00 0.00 × 10 + 00
best7.51 × 10 117 3.76 × 10 220 0.00 × 10 + 00
f 10 avg3.61 × 10 108 5.96 × 10 207 0.00 × 10 + 00
std5.34 × 10 107 0.00 × 10 + 00 0.00 × 10 + 00
best1.41 × 10 304 0.00 × 10 + 00 0.00 × 10 + 00
f 11 avg1.41 × 10 242 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.43 × 10 322 0.00 × 10 + 00 0.00 × 10 + 00
f 12 avg2.95 × 10 262 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.10 × 10 116 1.19 × 10 220 0.00 × 10 + 00
f 13 avg4.13 × 10 109 1.54 × 10 203 0.00 × 10 + 00
std5.08 × 10 108 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 14 avg0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 15 avg0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 16 avg0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 17 avg0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best4.44 × 10 16 4.44 × 10 16 4.44 × 10 16
f 18 avg2.04 × 10 15 4.44 × 10 16 4.44 × 10 16
std7.90 × 10 15 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 19 avg7.29 × 10 03 6.80 × 10 03 0.00 × 10 + 00
std1.88 × 10 02 1.99 × 10 02 0.00 × 10 + 00
best3.03 × 10 88 0.00 × 10 + 00 0.00 × 10 + 00
f 20 avg6.99 × 10 02 3.00 × 10 02 0.00 × 10 + 00
std2.05 × 10 01 2.05 × 10 01 0.00 × 10 + 00
Table 8. Comparison results with different dimensions.
Table 8. Comparison results with different dimensions.
n = 20n = 100n = 200n = 500n = 1000
WOA1.58 × 10 21 9.67 × 10 19 1.52 × 10 18 1.92 × 10 20 9.83 × 10 21
f 3 OBCWOA2.87 × 10 200 4.43 × 10 185 6.38 × 10 179 7.55 × 10 177 4.49 × 10 171
SWWOA0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
WOA6.87 × 10 + 01 1.56 × 10 + 03 3.37 × 10 + 03 8.17 × 10 + 03 1.59 × 10 + 04
f 7 OBCWOA3.15 × 10 + 01 2.07 × 10 + 03 4.27 × 10 + 03 9.20 × 10 + 03 1.84 × 10 + 04
SWWOA2.48 × 10 15 3.08 × 10 05 1.46 × 10 + 00 1.07 × 10 + 03 1.31 × 10 + 04
WOA3.80 × 10 105 2.25 × 10 110 5.09 × 10 108 4.57 × 10 103 4.13 × 10 109
f 13 OBCWOA1.47 × 10 231 3.78 × 10 214 6.65 × 10 210 4.43 × 10 208 1.54 × 10 203
SWWOA0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
WOA8.26 × 10 03 9.15 × 10 03 7.77 × 10 03 6.80 × 10 03 7.29 × 10 03
f 19 OBCWOA4.86 × 10 04 4.37 × 10 03 6.32 × 10 03 4.37 × 10 03 6.80 × 10 03
SWWOA0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
Table 9. Comparison of wins under different function types and dimensions.
Table 9. Comparison of wins under different function types and dimensions.
nABCPSOWOAOBCWOASWWOA
2000156
10000156
f 1 f 6 (US)20000156
50000156
1000156
2000135
10000135
f 7 f 12 (UN)20000135
50000135
1000135
2000223
10000223
f 13 f 15 (MS)20000223
50000223
1000223
2000245
10000235
f 16 f 20 (MN)20000235
50000235
1000235
Table 10. Comparison results for 8 shifted rotated functions with n = 20.
Table 10. Comparison results for 8 shifted rotated functions with n = 20.
Function ABCPSOWOAOBCWOASWWOA
best6.27 × 10 + 01 1.04 × 10 + 00 5.99 × 10 + 02 3.64 × 10 + 02 1.01 × 10 03
Shifted Rotated f 8 avg1.49 × 10 + 02 2.16 × 10 + 01 2.21 × 10 + 03 1.52 × 10 + 03 5.20 × 10 + 00
std1.92 × 10 + 02 1.25 × 10 + 02 5.54 × 10 + 03 3.19 × 10 + 03 2.91 × 10 + 01
best8.93 × 10 + 03 1.99 × 10 + 03 3.34 × 10 170 0.00 × 10 + 00 0.00 × 10 + 00
Rotated f 11 avg1.28 × 10 + 04 1.99 × 10 + 03 1.51 × 10 + 04 0.00 × 10 + 00 0.00 × 10 + 00
std3.98 × 10 + 03 2.03 × 10 12 1.11 × 10 + 05 0.00 × 10 + 00 0.00 × 10 + 00
best3.79 × 10 + 02 6.92 × 10 33 7.21 × 10 183 0.00 × 10 + 00 0.00 × 10 + 00
Rotated f 12 avg3.81 × 10 + 03 6.92 × 10 33 3.46 × 10 145 0.00 × 10 + 00 0.00 × 10 + 00
std1.54 × 10 + 04 1.22 × 10 47 6.74 × 10 144 0.00 × 10 + 00 0.00 × 10 + 00
best2.89 × 10 + 01 6.87 × 10 + 01 1.18 × 10 + 02 1.12 × 10 + 02 2.59 × 10 + 01
Shifted f 14 avg7.86 × 10 + 01 7.46 × 10 + 01 1.79 × 10 + 02 1.55 × 10 + 02 2.79 × 10 + 01
std1.06 × 10 + 02 1.54 × 10 + 01 1.22 × 10 + 02 7.29 × 10 + 01 9.75 × 10 + 00
best2.99 × 10 + 02 2.99 × 10 + 02 2.99 × 10 + 02 2.99 × 10 + 02 2.99 × 10 + 02
Shifted Rotated f 14 avg2.99 × 10 + 02 2.99 × 10 + 02 2.99 × 10 + 02 2.99 × 10 + 02 2.99 × 10 + 02
std2.54 × 10 13 2.54 × 10 13 2.54 × 10 13 2.54 × 10 13 2.54 × 10 13
best4.45 × 10 + 02 4.45 × 10 + 02 4.45 × 10 + 02 4.45 × 10 + 02 4.45 × 10 + 02
Shifted Rotated f 16 avg4.45 × 10 + 02 4.45 × 10 + 02 4.45 × 10 + 02 4.45 × 10 + 02 4.45 × 10 + 02
std7.63 × 10 13 7.63 × 10 13 7.63 × 10 13 7.63 × 10 13 7.63 × 10 13
best3.34 × 10 + 01 3.34 × 10 + 01 3.34 × 10 + 01 3.34 × 10 + 01 3.34 × 10 + 01
Shifted Rotated f 17 avg3.34 × 10 + 01 3.34 × 10 + 01 3.34 × 10 + 01 3.34 × 10 + 01 3.34 × 10 + 01
std6.36 × 10 14 6.36 × 10 14 6.36 × 10 14 6.36 × 10 14 6.36 × 10 14
best2.16 × 10 + 01 2.16 × 10 + 01 2.16 × 10 + 01 2.16 × 10 + 01 2.16 × 10 + 01
Shifted Rotated f 18 avg2.16 × 10 + 01 2.16 × 10 + 01 2.16 × 10 + 01 2.16 × 10 + 01 2.16 × 10 + 01
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
Table 11. Comparison results for eight shifted rotated functions with n = 100.
Table 11. Comparison results for eight shifted rotated functions with n = 100.
Function ABCPSOWOAOBCWOASWWOA
best6.30 × 10 + 02 1.87 × 10 + 03 5.47 × 10 + 04 3.84 × 10 + 04 3.16 × 10 + 02
Shifted Rotated f 8 avg1.20 × 10 + 03 4.97 × 10 + 03 7.36 × 10 + 04 5.29 × 10 + 04 7.48 × 10 + 02
std1.63 × 10 + 03 9.05 × 10 + 03 4.96 × 10 + 04 3.48 × 10 + 04 1.45 × 10 + 03
best2.18 × 10 + 04 1.03 × 10 + 04 5.87 × 10 07 0.00 × 10 + 00 0.00 × 10 + 00
Rotated f 11 avg3.49 × 10 + 04 1.56 × 10 + 04 1.58 × 10 + 05 0.00 × 10 + 00 0.00 × 10 + 00
std3.63 × 10 + 04 3.31 × 10 + 04 7.89 × 10 + 05 0.00 × 10 + 00 0.00 × 10 + 00
best9.01 × 10 + 08 6.83 × 10 + 04 3.19 × 10 178 0.00 × 10 + 00 0.00 × 10 + 00
Rotated f 12 avg3.94 × 10 + 09 4.42 × 10 + 05 1.67 × 10 152 0.00 × 10 + 00 0.00 × 10 + 00
std1.64 × 10 + 10 2.16 × 10 + 06 3.26 × 10 151 0.00 × 10 + 00 0.00 × 10 + 00
best3.35 × 10 + 02 5.42 × 10 + 02 1.24 × 10 + 03 1.03 × 10 + 03 3.20 × 10 + 02
Shifted f 14 avg4.74 × 10 + 02 6.17 × 10 + 02 1.32 × 10 + 03 1.20 × 10 + 03 4.08 × 10 + 02
std3.38 × 10 + 02 2.34 × 10 + 02 1.60 × 10 + 02 3.41 × 10 + 02 2.09 × 10 + 02
best1.70 × 10 + 03 1.70 × 10 + 03 1.70 × 10 + 03 1.70 × 10 + 03 1.70 × 10 + 03
Shifted Rotated f 14 avg1.70 × 10 + 03 1.70 × 10 + 03 1.70 × 10 + 03 1.70 × 10 + 03 1.70 × 10 + 03
std4.07 × 10 12 4.07 × 10 12 4.07 × 10 12 4.07 × 10 12 4.07 × 10 12
best3.47 × 10 + 03 3.47 × 10 + 03 3.47 × 10 + 03 3.47 × 10 + 03 3.47 × 10 + 03
Shifted Rotated f 16 avg3.47 × 10 + 03 3.47 × 10 + 03 3.47 × 10 + 03 3.47 × 10 + 03 3.47 × 10 + 03
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.81 × 10 + 02 1.81 × 10 + 02 1.81 × 10 + 02 1.81 × 10 + 02 1.81 × 10 + 02
Shifted Rotated f 17 avg1.81 × 10 + 02 1.81 × 10 + 02 1.81 × 10 + 02 1.81 × 10 + 02 1.81 × 10 + 02
std0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best2.17 × 10 + 01 2.17 × 10 + 01 2.17 × 10 + 01 2.17 × 10 + 01 2.17 × 10 + 01
Shifted Rotated f 18 avg2.17 × 10 + 01 2.17 × 10 + 01 2.17 × 10 + 01 2.17 × 10 + 01 2.17 × 10 + 01
std3.18 × 10 14 3.18 × 10 14 3.18 × 10 14 3.18 × 10 14 3.18 × 10 14
Table 12. p-values obtained from Wilcoxon’s rank sum test.
Table 12. p-values obtained from Wilcoxon’s rank sum test.
nFuncsABCPSOWOAOBCWOAFuncsABCPSOWOAOBCWOA
20 f 1 <0.001<0.001<0.0011 f 11 <0.001<0.001<0.0011
100 <0.001<0.001<0.0011 <0.001<0.001<0.0011
200 <0.001<0.001<0.0011 <0.001<0.001<0.0011
500 <0.001<0.001<0.0011 <0.001<0.001<0.0011
20 f 2 <0.001<0.001<0.0011 f 12 <0.001<0.001<0.0011
100 <0.001<0.001<0.0011 <0.001<0.001<0.0011
200 <0.001<0.001<0.0011 <0.001<0.001<0.0011
500 <0.001<0.001<0.0011 <0.001<0.001<0.0011
20 f 3 <0.001<0.001<0.001<0.001 f 13 <0.001<0.001<0.001<0.001
100 <0.001<0.001<0.001<0.001 <0.001<0.001<0.001<0.001
200 <0.001<0.001<0.001<0.001 <0.001<0.001<0.001<0.001
500 <0.001<0.001<0.001<0.001 <0.001<0.001<0.001<0.001
20 f 4 <0.001<0.001<0.0011 f 14 <0.001<0.00111
100 <0.001<0.001<0.0011 <0.001<0.00111
200 <0.001<0.001<0.0011 <0.001<0.00111
500 <0.001<0.001<0.0011 <0.001<0.00111
20 f 5 <0.001<0.001<0.0011 f 15 <0.001<0.00111
100 <0.001<0.001<0.0011 <0.001<0.00111
200 <0.001<0.001<0.0011 <0.001<0.00111
500 <0.001<0.001<0.0011 <0.001<0.00111
20 f 6 <0.001<0.00111 f 16 <0.001<0.00111
100 <0.001<0.00111 <0.001<0.00111
200 <0.001<0.00111 <0.001<0.00111
500 <0.001<0.00111 <0.001<0.00111
20 f 7 <0.001<0.001<0.001<0.001 f 17 <0.001<0.00111
100 <0.001<0.001<0.001<0.001 <0.001<0.00111
200 <0.001<0.001<0.001<0.001 <0.001<0.00111
500 <0.001<0.001<0.001<0.001 <0.001<0.00111
20 f 8 <0.001<0.001<0.0010.589 f 18 <0.001<0.0010.0031
100 <0.001<0.001<0.0010.194 <0.001<0.0010.0071
200 <0.001<0.001<0.001<0.001 <0.001<0.0010.0301
500 <0.001<0.001<0.001<0.001 <0.001<0.0010.1761
20 f 9 <0.001<0.001<0.0011 f 19 <0.001<0.001<0.001<0.001
100 <0.001<0.001<0.0011 <0.001<0.001<0.001<0.001
200 <0.001<0.001<0.0011 <0.001<0.001<0.001<0.001
500 <0.001<0.001<0.0011 <0.001<0.001<0.001<0.001
20 f 10 <0.001<0.001<0.001<0.001 f 20 <0.001<0.001<0.0011
100 <0.001<0.001<0.001<0.001 <0.001<0.001<0.001<0.001
200 <0.001<0.001<0.001<0.001 <0.001<0.001<0.0010.015
500 <0.001<0.001<0.001<0.001 <0.001<0.001<0.0010.003
Table 13. Ablation experiment with n = 20.
Table 13. Ablation experiment with n = 20.
Function WOAA1A2A3SWWOA
best1.89 × 10 27 3.53 × 10 30 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 3 avg4.73 × 10 19 4.00 × 10 21 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std9.17 × 10 18 7.15 × 10 20 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best1.73 × 10 07 1.05 × 10 + 01 3.51 × 10 + 02 1.15 × 10 + 03 4.71 × 10 21
f 7 avg8.87 × 10 + 01 1.39 × 10 + 02 3.30 × 10 + 03 3.12 × 10 + 03 8.54 × 10 16
std4.55 × 10 + 02 4.72 × 10 + 02 4.50 × 10 + 03 4.30 × 10 + 03 7.74 × 10 15
best7.37 × 10 121 1.22 × 10 121 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 13 avg1.43 × 10 112 2.23 × 10 111 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.28 × 10 111 4.10 × 10 110 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
best0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
f 19 avg7.29 × 10 03 6.32 × 10 03 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
std1.88 × 10 02 2.07 × 10 02 0.00 × 10 + 00 0.00 × 10 + 00 0.00 × 10 + 00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, P.; Cheng, W.; Liu, N.; Zhang, H.; Lu, J. A Modified Whale Optimization Algorithm with Single-Dimensional Swimming for Global Optimization Problems. Symmetry 2020, 12, 1892. https://doi.org/10.3390/sym12111892

AMA Style

Du P, Cheng W, Liu N, Zhang H, Lu J. A Modified Whale Optimization Algorithm with Single-Dimensional Swimming for Global Optimization Problems. Symmetry. 2020; 12(11):1892. https://doi.org/10.3390/sym12111892

Chicago/Turabian Style

Du, Pengzhen, Weiming Cheng, Ning Liu, Haofeng Zhang, and Jianfeng Lu. 2020. "A Modified Whale Optimization Algorithm with Single-Dimensional Swimming for Global Optimization Problems" Symmetry 12, no. 11: 1892. https://doi.org/10.3390/sym12111892

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop