Next Article in Journal
Fiber Bragg Grating-Based Optical Signal Processing: Review and Survey
Previous Article in Journal
Acoustic, Myoelectric, and Aerodynamic Parameters of Euphonic and Dysphonic Voices: A Systematic Review of Clinical Studies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Studying the Impact of Initialization for Population-Based Algorithms with Low-Discrepancy Sequences

1
IT Support Center, GC Women University Sialkot, Punjab 51310, Pakistan
2
Department of Computer Science, Abasyn University, Islamabad 45710, Pakistan
3
Faculty of Computing and Informatics, Universiti Malaysia Sabah, Jalan UMS, Kota Kinabalu 88400, Malaysia
4
Department of Computer Science, University of Gujrat, Punjab 50700, Pakistan
5
Centro de Tecnologia, Campus Petrônio Portela, Federal University of Piauí (UFPI), Teresina 64049-550, Brazil
6
Instituto de Telecomunicações, 6201-001 Covilhã, Portugal
7
Data Science and Cybersecurity Center, Department of Electrical Engineering and Computer Science, Howard University, Washington, DC 20059, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(17), 8190; https://doi.org/10.3390/app11178190
Submission received: 7 April 2021 / Revised: 29 April 2021 / Accepted: 3 May 2021 / Published: 3 September 2021
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
To solve different kinds of optimization challenges, meta-heuristic algorithms have been extensively used. Population initialization plays a prominent role in meta-heuristic algorithms for the problem of optimization. These algorithms can affect convergence to identify a robust optimum solution. To investigate the effectiveness of diversity, many scholars have a focus on the reliability and quality of meta-heuristic algorithms for enhancement. To initialize the population in the search space, this dissertation proposes three new low discrepancy sequences for population initialization instead of uniform distribution called the WELL sequence, Knuth sequence, and Torus sequence. This paper also introduces a detailed survey of the different initialization methods of PSO and DE based on quasi-random sequence families such as the Sobol sequence, Halton sequence, and uniform random distribution. For well-known benchmark test problems and learning of artificial neural network, the proposed methods for PSO (TO-PSO, KN-PSO, and WE-PSO), BA (BA-TO, BA-WE, and BA-KN), and DE (DE-TO, DE-WE, and DE-KN) have been evaluated. The synthesis of our strategies demonstrates promising success over uniform random numbers using low discrepancy sequences. The experimental findings indicate that the initialization based on low discrepancy sequences is exceptionally stronger than the uniform random number. Furthermore, our work outlines the profound effects on convergence and heterogeneity of the proposed methodology. It is expected that a comparative simulation survey of the low discrepancy sequence would be beneficial for the investigator to analyze the meta-heuristic algorithms in detail.

1. Introduction

The term ‘optimization’ refers to the best solution for a problem with minimum cost in aspect of memory, time, and resources. Sometimes processing time is fast but it may be using a lot of memory while sometimes the processing speed and memory both work fine but the accuracy may get affected. Optimization targets the best solution of any problem [1]. The solution is considered to be the best solution if it is satisfactory in terms of processing speed, resource utilization, and accuracy of the result [2]. Optimization algorithms are utilized to determine the problems of local and global search. A typical target behind the utilization of these optimization algorithms is to discover the optima for contribution as indicated by known inputs model that describes the problem which is to be solved [3]. Optimization algorithms have turned out to be the most generally adopted algorithms that are operational in all application areas, like, enterprises, sports, medical, agriculture, and finance [4].
Evolutionary algorithms (EAs) have been introduced and strongly employed in the different field of science and engineering tracks [5]. EAs have been broadly utilized to determine optimization problems of maximization and minimization to find best optimal value. Rather than ordinary strategies dependent on mathematical programming or formal rationales, EAs are observed to be all the more dominant and adaptable [6]. Despite this, in solving the complex optimization problems, EAs faces the problem of local optima in which the computation to be caught in nearby local optima and a resist the convergence speed, for example, complex nonlinear problems [7]. To enhance the performance of EAs and to avoid premature convergence, there is a need to develop new variants of evolutionary algorithms. Furthermore, dependent on genetic evolution procedures, several researchers have been given the task to improve existing EAs or developing new EAs. The most generally used EAs algorithms involve the genetic algorithm (GA) [8] and differential evolution [9]. DE is recognized as a simplistic yet strong evolutionary algorithm that has been utilized to tackle different hard optimization problems in various science and engineering disciplines [10]. As another component of an EA, the DE algorithm yields a comparative structure [11] by EA, which incorporates three essential basic genetic operators, i.e., mutation, crossover, and selection. These genetic operators contribute major roles to the performance of the DE [12].
The intelligent attitude of non-intelligent species like ants (going for searching food) or birds (during flying in flocks) or fish in school is termed as swarm intelligence (SI). SI inspired by the experience of ants, bees, birds, and fishes to fulfill their goals as a swarm [13]. If every member of a swarm in SI works individually without social interaction, it will become complicated to achieve their goals due to their individuality which results in lack of intelligence. However, when they cooperate with each other, their social interaction is improved, and they also interact with the environment which makes it easier for them to accomplish difficult tasks [14]. SI-based algorithms are ant colony optimization (ACO), bat algorithm (BA) [15,16,17,18], and particle swarm optimization (PSO) [19]. PSO [20] has pulled in much consideration because of its simplicity of execution and strong search abilities. It is impelled by the social foraging fashions of fish and birds that seek for food in the form of groups.
The major issue with these meta-heuristic algorithms while applying those complex numerical optimization problems is premature convergence [21]. Regardless of the nature of the non-linear problem, this issue is confronted while running a heuristic algorithm, for example, meta-heuristic algorithms like PSO [22] and DE get stuck in the local optima after little number of epochs. The population convergence fails to produce a new population of the swarm due to inappropriate animalization strategies to explore the whole search space [23]. In the field of evolutionary computing, the performance of meta-heuristics algorithms is affected by the generation of random numbers while initializing the population into the multidimensional search space [24]. The meta-heuristics algorithm tends to reach the optimum value while solving the problems in low dimensional search space. However, the performance is supposed to be insignificant when the dimensionality of the problem is high, and this causes the particles to stick in the local optima [25]. Meta-heuristics population-based algorithm initialization can be performed by using chaotic initialization [26,27,28], opposition-based initialization, and quasi-random sequences. This paper presents the impact of quasi-random sequences for the initialization of the population of the meta-heuristics algorithm.
In accordance with the optimization problem, population initialization plays a significant role in meta-heuristic algorithms. These algorithms can influence diversity, convergence, and also help to find an efficiently optimal solution. Particularly, recognizing the importance of diversity, several researchers have worked on performance for the improvement of meta-heuristic algorithms. In order to improve the convergence, rather applying the random distribution for initialization, quasi-random sequences are more useful to initialize the population [29].
Quasi-random sequences suffer from many issues while solving the problems of different dimensionality in real world [30]. Some of the sequences of quasi-random sequences give better results on large dimensions and vice versa [31,32]. Our objective is to find the most suitable quasi-random sequence for meta-heuristic algorithms, which gives superior results without considering the dimensionality problem.
Considering this fact, we have proposed three novel pseudo-random initialization strategies called WELL sequence, Knuth sequence [33], and Torus sequence to initialize the population in the search space. We initialized PSO, BA, and DE algorithm with these proposed pseudo-random strategies (WELL sequence, Knuth sequence, and Torus sequence). In our first contribution, we have compared the novel PSO technique with the simple random distribution [34] and family of low discrepancy sequences [35] on several unimodal and multi modals complex benchmark functions and training of the artificial neural network [36]. The experimental results have shown that PSO with Knuth-based initialization (KN-PSO) outperforms the other traditional PSO, PSO with Sobol-based initialization (SO-PSO), PSO with Halton-based initialization (H-PSO), PSO with Torus-based initialization (TO-PSO), and PSO with WELL-based initialization (WE-PSO) [37]. Similarly, in the second contribution, DE is initialized by these proposed pseudo-random strategies (WELL sequence, Knuth sequence, and Torus sequence) for function optimization and training of the neural network. The simulation results depict that DE with Halton-based initialization (DE-H) is superior to the standard DE, DE with Sobol-based initialization (DE-SO), DE with Knuth-based initialization (DE-KN), DE with Torus-based initialization (DE-TO), and DE with WELL-based initialization (DE-WE). BA is initialized by these proposed pseudo-random strategies (WELL sequence, Knuth sequence, and Torus sequence) for function optimization and training of the neural network. The simulation results depict that BA with WELL-based initialization (BA-WE) is superior to the standard BA, BA with Sobol-based initialization (BA-SO), BA with Halton-based initialization (BA-HA), BA with Torus-based initialization (BA-TO), and BA with Knuth-based initialization (KN-BA) [38,39,40].
The rest of the paper is organized as: Section 2 overviews the previous work. In Section 3, the different algorithm methodology is represented with six initialization strategies. Section 4 contains the experimental setup. In Section 5, the results and discussion about the implementation and comparison of algorithms using initialization techniques on sixteen benchmark tests functions are presented. Section 6 presents the comparison of PSO, BA, and DE regarding data classification. Lastly, Section 7 concludes the paper.

2. Previous Work

Many research studies proposed different variants based on initialization techniques and we have discussed some of them in detail in this chapter. Initialization of the swarm in a good way helps the PSO to search more efficiently [41]. In this work, the initialization of swarm with nonlinear simplex method (NSM) has been done. NSM requires only function evaluations without any derivatives for computation. NSM starts with initial simplex and produces sequence of steps moving the highest function value vertex in opposite direction of the lowest one. They initialized the particle with the initial simplex in the D dimensional search areas, where D + 1 vertices of the simplex are D + 1 particle of the swarm and the MSN method is applied for N-D + 1 steps for N size swarm. In this way, each particle in the swarm has the information of the region. In the last, they compared their results with simple PSO and found significant improvement. This variant was introduced by Mark Richards and Dan Ventura in 2004.
In their work [42], they proposed to use centroidal Voronoi tessellations for initializing the swarm. Voronoi tessellations is a technique of partitioning any region into compartments, each partition contains group of generators. Each partition is associated with one generator and it consists of all the particles closer to that generator. In the same way, the generators are selected for the initial position of the particle. In this way, they initialized the particle swarm optimization algorithm. They compared it with basic SPO on many benchmark functions and found improved performance in high-dimensional spaces.
Halton sampling was introduced by Nguyen Xuan Hoai, Nguyen Quang Uy, and R.I. McKay in 2007 [43]. Halton sequence is a low discrepancy deterministic sequence used to generate point in space. Halton sequence is not a fully random. To randomize, X.Wang and F.J. Hickernell proposed a new function called randomize Haltom sequence by using von Neuman–Kakutani transformation. They used this sequence to initialize the global best of the PSO. They performed a test on various benchmark functions and compared the result with the PSO and initialized with uniform random numbers. They found better performance especially for complex and smaller populations.
VC−PSO was introduced by Millie Pant et al. in 2008. They used Vandor Corput sequence for initializing the swarm for large dimensions’ search [44] (Pant, Thangaraj, Grosan, and Abraham, 2008). The Vandor Corput and Sobol sequence generates semi random number which is more suitable for computational purposes. They tested the new variant with different benchmark functions and compared the result with BPSO and SO-PSO and found significant improvement especially for large search space dimension. The main purpose of this variant is to see the performance of large search space problems. That is why they used search space with different dimensions ranging [−5.12, 5.12] to [−1000, 1000]. The performance is showing prominent when the dimension increases to [−100, 100].
SMPO was introduced by Millie Pant et al. in 2008. They used quasi-random Sobol sequence to initialize the particles instead of normal random numbers [45]. They used a new operator called systematic mutation operator which is used to improve the performance of the PSO. Instead of using the normal random number, the new operator uses the quasi-random Sobol sequence to initialize the swarm as the QRS is less random as compared to pseudorandom sequences which is helpful for computational methods. They proposed two variants, SM-PSO1 and SM-PSO2. The main difference between the two versions is that in MSPSO1, the best particle is mutated while in MS-PSO2, the worst particle is mutated. They found better results comparing with BPSO and other variants.
This work is done by Jiyong et al. 2011 [46]. In this paper, researchers proposed a new method of initialization. In their work, they added the functionality to detect automatically when the particle is prematurely converged and initializes the swarm. They also added functionality to redesign the inertia weight to balance the searching ability globally and locally. They named it IAWPSO.
This variant was proposed by P. Murugan in 2012 and applied on the transmission expansion problem to decide installation of new circuits in an increasing usage of electricity and found this variant fruitful [47]. In this work, he used the new initialization technique called population monitored for complementary magnitudes initialization. In initialization, he used decision variables. All particles are initialized with and integer within the limit of the upper and lower values of the decision variable in such a way that each particle should be unique. The initial population is created in a way that each particle can have the ability of the possible solution and they are unique. Almost 50% of the particles are opposite to another 50% considering the lower and upper limit of decision variable. The important thing in this initialization is to maintain uniqueness and diversity among the particles of the swarm generated initially.
SISP SO was introduce by Liang Yin, Xiao−Min Hu, and Jun Zhang in 2013 [48]. In this paper, the authors introduced a new initialization technique named space-based initialization strategy. In this work, they broke down each dimension of the search area into two segments, S1i and S2i. The borders of the areas are [li,(li + ui)/2] and [(li + ui)/2, ui], with each segment linked with a probability and initialized with 0.5. They applied SIS-PSO on thirteen functions and compared results with GPSO and CLPSO and found significant improvement.
This variant was introduced by Moaath Shatnawi, Mohammad Faidzul Nasrudin, and Shahnorbanun Sahran in 2017 [49]. In this work, they introduced a new variant of PSO called polar PSO. They explained that most of the distortion was occurring due to polar particles. Hence, they introduced a new method for reinitialization of the polar particles by redefining the distance based on the dimensionality of the point. By using this method, it removed the distortion occurring during the computation. He compared the results with BPSO and found some improvement.
This variant was proposed by Laxmi et al. in 2017 [50]. In this work, they used the Nawaz–Enscore–Ham heuristic technique to initialize the swarm. This variant is named PHPSO. The sequence generated by NEH jobs is placed in ascending order of the sums of their total flow time. To construct a job sequence, it depends on its initial order. The minimum TFT sequence is the current sequence for the upcoming iteration among all the sequences. The resulting population generated by the NEH method is used to initialize the population of PSO. They applied this algorithm for the no-wait flow shop scheduling problem. They compared the result with DPSO and HPSO and found the comparatively better result.
A new variant of PSO combining with stochastic gradient decent was proposed by Hayder M. Albeahdili, Tony Han, and Naz E. Islam in 2015 and named it the PSO–SGD algorithm for training the convolution neural network [51]. The proposed technique was divided into two phases. PSO was used to train and initialize the CNN parameters in the first phase. When it showed slow progress of the PSO for few iterations, SGD was used in the second phase. Additionally, they used PSO combined with the genetic algorithm (GA) which helped the particle for simulation and overcame the slowness of SGD. They applied the new algorithm on different benchmark datasets and performed well for three different datasets. The proposed technique avoided the occurrence of local optimum and premature saturation as it was in the known problem by using any single algorithm.
The authors in [52] examined the impact of initiating the initial population by excluding traditional techniques like random numbers or quasi-random numbers. The authors applied the non-linear simplex method for generating the initial population of DE, where the proposed algorithm was termed NSD. The working of the proposed algorithm is measured with twenty benchmark functions and compared with the standard DE and opposition-based DE (ODE) algorithm. Numerical results illustrate that the proposed technique enhances the convergence rate.
To tackle the thresholding problem of the image, an enhanced variant of the standard DE algorithm with a local search (termed as LED) and low discrepancy sequences is introduced [53]. Experimental results conclude that the performance of the introduced algorithm is superior for finding the optimum threshold.
For the steelmaking continuous (SCC) problem, in [54], the authors presented a novel enhanced technique of DE based on the two-step procedure for producing an initial population, as well as, a novel mutation approach. Furthermore, an incremental methodology for generating the initial population was also incorporated in DE to handle dynamic events. Computational experiments conducted with the presented approach show the effectiveness of the presented approach than others. Additionally, as per concern, in the application area, the authors utilized BA for the antenna optimization problem in [55], moreover, pan evaporation was estimated by using BA [56]. Beside this, in [57], the authors applied a new variant of DE for path-planning of mobile robots.
According to the above-mentioned studies, we conclude that the efficiency of meta-heuristic algorithms is affected by using the random number for the initialization of the population. Due to this reason, various articles used the quasi-random number sequences for the population’s initialization in meta-heuristic algorithms. However, the majority of the researchers used limited quasi-random sequences for initializing the population and did not perform any comparative analysis for their effect on the initialization of population algorithms. Similarly, Knuth, Well, and Torus sequences from quasi-random sequences are still not proposed in DE and BA for the initialization of the population. After analyzing all the literature, we found the above-mentioned gaps and try to fill it.

3. Methodology

The most important step in any meta-heuristic algorithm is to initialize its population properly. If the initialization is not proper, then it may go to search in unnecessary areas and may fail to search the optimum solution. Proper initialization is very important for any algorithm for its performance. The objective of this paper is to figure out the purity of quasi-random sequences. PSO is random in nature, so it does not have a specific pattern to ensure the global optimum point. Therefore, by taking the advantage of this randomness and considering this fact, we proposed three novel quasi-random initialization strategies called WELL sequence, Knuth sequence, and Torus sequence to initialize the population in the search space. We initialized the PSO, DE, and BA algorithm with these proposed pseudo-random strategies (WELL sequence, Knuth sequence, and Torus sequence). We have compared the novel techniques with the simple random distribution and family of low discrepancy sequences on several unimodal and multi modals complex benchmark functions and training of the artificial neural network. A brief description of quasi sequences approaches and proposed algorithms using WELL sequence, Knuth sequence, and Torus sequence for PSO, DE, and BA are discussed in below.
It has been stated above; the goal of this study is to analyze the purity of low discrepancy sequences. Therefore, we compare the proposed algorithm based on WELL, Torus, and Knuth distribution with the simple PSO, BA, and DE based on pseudo-random uniform distribution and other low discrepancy distributions based on the Sobol sequence and Halton sequence.

3.1. Low Discrepancy Sequences

Discrepancy is the measure of how uniform the numbers are distributed. Consider the set of points P = (x1, x2, ……, xn) be set of n points in s dimensions in [0, 1)s. For a vector y = (y1, y2, …, ys) [0, 1)s, let J be:
j = i = 1 s [ 0 , y i ) s
Although there exit other measures of discrepancy, the star discrepancy is commonly used. A low value of discrepancy means more uniform distribution in space.

3.1.1. Uniform Random Numbers

Random numbers are generated through a pseudo random sequence by pursuing uniform-distribution [44], which can be typified using the probability-density function of constant uniform-distribution. Given below is the probability-density function in (2) as:
f ( t ) = { 1 p q   f o r   p < t < q 0   f o r   t < p   o r   t > q
where u and v describe the features that fit the maximum likelihood. At the edge of u and v, the cost of f(w) is unproductive because of 0 impacts on the integrals of f(w)dw at any range. The likelihood function of assessment helps to simulate the assessment of features of maximum likelihood, likelihood function of assessment is given below in (3) as:
l ( p , q | t ) = n log ( q p )

3.1.2. Sobol

The Sobol sequence is firstly introduced by a Russian mathematician, Sobol [45]. Then, reconstruct the coordinates. Coordinates have liner recurrence relation for each dimension. Let the non-negative instance s containing a binary expression in (4) where s:
a = a 1 2 0 + a 2 2 1 + a 3 2 2 + + a z 2 z 1
Then, the ith instance of the D dimension can be generated using the (5):
x i D = i 1 v 1 D + i 2 v 2 D + + i z v z D
where   v 1 D represents the binary function which is followed by the D dimension and ith direction instance and these direction instances can be generated using the (6):
V k D = c 1 v k 1 D + c 2 v k 2 D + + c z v z 1 D + ( v i z D 2 z )  
where cq is a polynomial coefficient where i > q.

3.1.3. Halton

Halton sequences was carried out by J. Halton [43] and can be considered as the enhanced version of Van Dar Corput (Gentle, 2006). Halton sequence constructs random points pattern by using the base as coprime. Halton sequences: The pseudo code to generate Haltom sequences is as follow:
Halton Sequences:
  • //input: Initial index = s and base = coprime
  • //output: instances = h
  • Set the interval over
  • min = 0
  • max = 1
For each iteration k1, k2, k3…kn:do
  • For each particle p1, p2, p3, …, pn
max = max/coprime
min = min + max ∗ smodb
s = s/b
Return h

3.1.4. WELL

WELL equi-distributed long-period linear (WELL) sequence was proposed in [58]. Initially, it was carried out as updated version of the Mersenne twister algorithm. The algorithm for generating the WELL distribution is given as:
WELL Sequences
  • WELL ():
  • t 0 = ( m x & v k , r 1 ) + ( m x & v k , r 2 )
  • t 1 = ( A 0 v k , 0 ) + ( A 1 v k , m 1 )
  • t 2 = ( A 2 v k , m 2 ) + ( A 3 v k , m 3 )
  • t 3 = t 2 + t 1
  • t 4 = t 0 A 4 + t 1 A 5 + t 2 A 6 + t 3 A 7
  • v k + 1 , r 1 = v k , r 2   &   m x
  • f o r   i r 2 ..2   d o   v k + 1 , i = v k , i 1
  • v k + 1 , 1 = t 3
  • v k + 1 , 0 = t 4
  • R e t u r n   y k = v k , 0
The algorithm stated above describes the general recurrence for the WELL distribution. The description for the algorithm is as: x and r two integers with the interval of r > 0 and 0 < x < k and k = rwx, where w is the weight factor of distribution. A0 to A7 represent the binary matrix of size rw having r bit block. mx describes the bit mask that holds the first wx bits. t0 to t7 are temporary vector variables.

3.1.5. Knuth

As discussed above, inbuilt library function is used, Knuth(x(min,)xmax) to generate Knuth sequences random points. Following is the pseudo code to generate Knuth sequences. Knuth sequence is designed and was proposed by the authors in [33]. Following is the pseudo code to generate Knuth sequences.
  • To shuffle an array a of n elements (indices 0…n − 1):
  • F orifrom0ton − 2do
  • jrandomintegersuchthati ≤ j < n
  • exchangea[i]anda[j]

3.1.6. Torus

Torus is a geometric term and was firstly used by the authors in [59] to generate a Torus mesh for the geometric coordinate system. In game development, Torus mesh is commonly used and can be generated using the left hand coordinate system or right hand coordinate system. The shape for the Torus at 1D, 2D, and 3D are circle, donut, and 2D rectangle, respectively. The Torus in 3D can be represented by the following (7)–(9):
a ( θ , δ ) = ( D + r cos θ ) cos δ ,
b ( θ , δ ) = ( D + r cos θ ) sin δ ,
c ( θ , δ ) = r sin δ ,
where the angles of circles are θ, δ and D is the distance from tube center to Torus center, r denotes to the radius of 6circle. Inspired by this mesh having Torus, low discrepancy sequences have been generated that were initialized with the prime series as Torus effect. In (10), the mathematical notation for Torus series is shown:
α k = ( f ( k s 1 ) , , f ( k s d ) ) , .  
where s1 denotes the series of ith prime number and f is a fraction which can be calculated by f = a − floor(a). Due to the prime constraints, the dimension for the Torus is limited to the 100,000 only if we use parameter prime in Torus function. For more than 100,000 dimensions, the number must be provided through manual way.
In Figure 1, uniform random distribution; Figure 2, Sobol distribution; Figure 3, Halton distribution; Figure 4, WELL distribution; Figure 5, Knuth distribution; Figure 6, Torus distribution are presented by bubble plot, in which y-axis is representing the random values and the x-axis is showing the relevant index of the concerned point in the table. We made our first major contribution to this study first by introducing three novel methods of initialization of population: WE-PSO, KN-PSO, and TO-PSO. The algorithm shows the flow chart of the proposed distribution-based PSO initialization.
The Algorithm 1 shows the flow chart of the proposed distribution-based PSO initialization.
Algorithm 1 Proposed Pseudo Code of PSO Using Novel Method of Initialization
  • Initialize the swarm
  • Set epoch count I = 0 , population size N z , Dimension of the problem D z , w m a x and w m i n
  • For each particle  P z .
  • Initialize x z , as x z =   WELL ,   Knuth ,   Torus   ( x m i n , x m a x )
  • Initialize the Particle velocity as, v z = R a n d ( x m i n , x m a x . )
  • Compute the fitness score f z
  • Set global best position g z b e s t as m a x ( f 1 , f 2 , f 3 .. f z ) where f z g l o b a l l y   o p t i m a l   f i t n e s s
  • Set local best position p z b e s t as m a x ( f 1 , f 2 , f 3 .. f z ) where f z l o c a l l y   o p t i m a l   f i t n e s s
  • Compare the current particle’s fitness score   x z in the swarm and its old local best location p z b e s t If the current fitness score   x z is greater than p z b e s t , then substitute p z b e s t , with   x z ; else retain the   x z unchanged
  • Compare the current particle’s fitness score   x z in the swarm and its old global best location g z b e s t If the current fitness score   x z . is greater than g z b e s t , then substitute g z b e s t , with   x z ; else retain the   x z unchanged
  • Compute   v z + 1 → updated velocity vector
  • Compute   x z + 1 → updated position vector
  • Go to step 2; If the stopping criteria does not met; else terminate
In our other contribution in the paper, we introduced three novel methods of initialization of population: DE-WE, DE-KN, and DE-TO. The Algorithm 2 shows the flow chart of the proposed distribution-based DE initialization.
Algorithm 2 Proposed Pseudo Code of DE Using Novel Method of Initialization
Input:xi = (xi,1, xi,2, xi,3, ..., xi,D), Population size ‘N-P’, Problem Size ‘D’, Mutation Rate ‘F’, Crossover Rate ‘C-R’; Stopping Criteria {Number of Generation, Target}, Upper Bound ‘U’, Lower Bound ‘L’,
Output: xi = Global fitness vector with minimal fitness value
1.
Pop = Initialize of Paraments (N-P, D, U, L);
a.
Generate initial population Using WELL,Knuth,Torus
2.
While (Stopping Criteria ≠ True) do
3.
Best Vector = Evaluate Pop (Pop);
4.
vx = Select Rand Vector (Pop);
5.
I = Find Index Vector (vx);
6.
Select Rand Vector (Pop,v1,v2,v3) where v1 ≠ v2 ≠ v3 ≠ vx
7.
vy = v1, + F(v2−v3)
8.
For (i = 0; i++; i < D−1)
9.
If (randj [0, 1) < C-R) Then
10.
U[i] = vx [i]
11.
else
12.
U[i] = vy [i]
13.
End For loop
14.
If (Cost Fun Vector(U) ≤ Cost Fun Vector (vx)) Then
15.
Update Pop (U, I, Pop);
16.
End IF
17.
End While
18.
Retune Best Vector;
In our last contribution in the paper, we introduced three novel methods of initialization of population: BA-WE, BA-KN, and BA-TO. The Algorithm 3 shows the flow chart of the proposed distribution-based BA initialization.
Algorithm 3 Proposed Pseudo Code of BA Using Novel Method of Initialization
(1)
Bat-Initialization();Using WELL,Knuth,Torus
(2)
E = newly_evluated_population;
(3)
f m i n = current_solution(best);
(4)
While termination_condition_not_meet do
(5)
fori = 1 to population do
(6)
x t = compute_solution(best);
(7)
if rand(0,1) > r i then
(8)
x t = update_the_current-solution(best)
(9)
end if {searching locally}
(10)
if f n e w = compute_new_solution( x t ) ;
(11)
E = E + 1; Addition in evaluation
(12)
if f n e w < f i     and N(0,1) < A i then
(13)
  x i = x t ;
(14)
f i = f n e w ;
(15)
end if {Simulation annealing}
(16)
  f m i n = explore_for_best_solution(best);
(17)
end for
(18)
end while

4. Experimental Setup

To achieve the effective working of the algorithms, it is compulsory to adjust the parameters coupled with all approaches to their most suitable value. Largely, these parameters are observed before the implementation of the algorithm and maintain uniformity throughout the execution. In various studies, it is suggested that the most appropriate methodology for selecting the parameters of any algorithm is predicted through exhaustive experiments for obtaining the optimal parameters. In this study, the experimental setting of the parameters is employed in Table 1 and the parameters setting of the algorithms in Table 2, respectively, which is on the basis of the literature stated in this section. Along with this, objective functions and their details are in Table 3. The search space boundary is [−100, 100] population size kept 50 with 10 number of runs. Further, 10, 20, and 30 dimensions are used for 1000, 2000, and 3000 iterations, respectively.

5. Results and Discussion

This section briefly describes the simulation results of the proposed approaches and their graphical representation. Primarily, this section is divided into three sub-sections, where each sub-section is specifically dedicated to EAs simulation results such as PSO, DE, and BA, respectively. In addition to this, each EA is also examined through the statistical tests which are also stated in sub-sections.

5.1. Discussion on PSO Results

The simulation was simulated in C + + and applied on a computer using the C + + language on the computer having the Windows 10 operating system with the specifications of 8 Gigabyte ram and 2.3 GHz Core (M) 2 Duo CPU processor. In order to measure the execution of the proposed approaches, WELL-based PSO (WE-PSO), Torus-based PSO (TO-PSO), and Knuth-based PSO (KN-PSO), a group of fifteen non-linear benchmark test functions were utilized to do the comparison of WE-PSO, TOPSO, and KN-PSO with standard PSO, SO-PSO, and HPSO. These functions are normally applied to investigate the performance of any technique. Hence, in our study, we used them to examine the optimization outcomes of the quasi-random-based approach of WE-PSO, TO-PSO, KN-PSO, SO-PSO, and H-PSO. The list of those functions is available in Table 3. In Table 1, D(Dimensions) shows the dimensionality of the problem, S (Search Space) represents the interval of the variables, I(Iterations), Pop(Population size) and in Table 2, fmin denotes the common global optimum minimum value. The parameters for the simulation use c1 = c2 = 1.45, inertia weight w is used in the interval [0.9, 0.4], and swarm size is 50. For simulation, the function dimensions are D = 10, 20, and 30 and a maximum number of epochs is 3000. All techniques were applied to similar parameters for comparatively effective results. In order to check the performance of each technique, all algorithms were tested for 30 runs.
The purpose of this study continues to observe whereby the unique characteristics of experimental results rely on dimensions of these standard benchmark functions.
The objective of this study is to find the most suitable initializing approach for the PSO and during the first experiment, the proposed WE-PSO, TO-PSO, and KN-PSO with other approaches SO-PSO, H-PSO, and standard PSO was investigated. The objective of the second simulation is to find the nature of the dimension regarding standard function optimization. Lastly, the simulation results of WE-PSO, TO-PSO, and KN-PSO were compared with standard PSO, SO-PSO, and H-PSO. In the rest of the paper, simulation results are discussed in detail.
Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15, Figure 16, Figure 17, Figure 18, Figure 19, Figure 20, Figure 21 and Figure 22 contain the graphical representation of the comparisons of the proposed WE-PSO, TO-PSO, and KN-PSO with standard PSO, H-PSO, and SO-PSO. In the x-axis, dimensions of the problem 10, 20, and 30 are presented while the y-axis represents the mean best against each dimension of the problem.
We can see that the majority of the figures contain a better convergence curve for KN-PSO on functions F1, F2, F3, F4, F5, F6, F17, F8, F9, F10, F11, F12, F13, F14, and F15 over WE-PSO, TO-PSO, H-PSO, SO-PSO, and standard PSO on all dimensions comprehensively. The other proposed approach TO-PSO provides better results over WE-PSO on functions F1, F2, F4, F6, F17, F8, F9, F10, F14, F15, and F16 and beats on all functions for PSO, SO-PSO, and TO-PSO.
In this simulation, PSO is initialized with the WELL sequence (WE-PSO), Torus sequence (TO-PSO), and Knuth sequence (KN-PSO) instead of uniform distribution. The variant proposed WE-PSO, TO-PSO, and KN-PSO is compared with other initialized approaches, Sobol sequence (SO-PSO), Halton sequence (H-PSO), and standard PSO. The experimental results give superior results in higher dimensions for KN-PSO on other SO-PSO, H-PSO, PSO, and proposed approach TO-PSO and WE-PSO.
The core objective of this simulation setup is to find the superiority of results depending upon the dimension of the functions that are to be optimized. In experiments, three dimensions for benchmark functions D = 10, D = 20, and D = 30 were used. Simulation results are presented in Table 4. From these simulation results, it was found that functions having larger dimensions were tougher to optimize and it can be seen from the Table 4 when dimension size id is D = 20 and D = 30 and our proposed approach KN-PSO shows belter result on higher dimensions on other approaches WE-PSO, TO-PSO, standard PSO, H-PSO, and SO-PSO.
KN-PSO is compared with the other approaches like WE-PSO, TO-PSO, SO-PSO, H-PSO, and standard PSO where each technique true value is presented for comparison with other techniques for the same nature of problems. Standard benchmark functions are presented in the Table 3 and their parameter settings are also shown in the Table 1. Table 4 shows that with dimension D-30, KN-PSO is more superior and outperforms in convergence than the WE-PSO, TO-PSO, standard PSO, SO-PSO, and H-PSO. The comparative analysis can be seen from Table 4 that with smaller dimension size, standard PSO performs well (D = 10); while the size of the dimension increases, KN-PSO outperforms in convergence significantly. Hence, KN-PSO is best for higher dimensions. The experimental results from Table 4 show that KN-PSO outclassed the results of WE-PSO, TO-PSO, SO-PSO, H-PSO, and traditional PSO for all functions. It can be seen that the TO-PSO outperformed the results of the other techniques in all functions on SO-PSO, H-PSO, standard PSO; while in the other approaches, H-PSO performs better on functions f4, f1, f2 for 20D but H-PSO gives overall poor result on higher dimensions and SO-PSO gives slightly better results on functions f8, f9, f15 on 10-D but worst results on larger dimensions. Standard PSO did not provide better results. Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14 and Figure 15 depict that WE-PSO outperforms in simulation results than other approaches for solving the standard benchmark tests functions for dim size D = 10, D = 20, D = 30.
To validate the numerical results, mean ranks obtained by Kruskal–Wallis and Friedman tests for KN-PSO, WE-PSO, TO-PSO, SO-PSO, HA-PSO, and standard PSO are given in Table 5.

5.2. Discussion on DE Results

Population initialization is a vital factor in the evolutionary computing-based algorithm, which considerably influences the diversity and convergence. In order to improve the diversity and convergence, rather applying the random distribution for initialization, quasi-random sequences are more useful to initialize the population. In this paper, the capability of DE was extended to make it suitable for the optimization problem by introducing new initialization techniques: Knuth sequence-based (DE-KN), the Torus-based sequence-based (DE-TO), and the WELL sequence-based (DE-WE) by using low discrepancies sequence, Torus to solve the optimization problems in large dimension search spaces.
For global optimization, the most considerable variety of benchmark problems can be used. All benchmark problems have their own individual abilities and the variety of detailed characteristics of such functions explains the level of complexity for benchmark problems. For the efficiency analysis of the above-mentioned optimization algorithms, Table 3 displays the benchmark problems that are utilized. Table 3 explains the following contents of benchmark problems: name, range, domain, and formulas. In this study, those benchmark problems are incorporated, which have been extensively utilized in the literature for conveying a deep knowledge of the performance related to the above-mentioned optimization algorithms.
To measure the effectiveness and robustness of optimization algorithms, benchmark functions are applied. In this study, fifteen computationally-expensive black box functions are applied with their various abilities and traits. The purpose to utilize these benchmark functions is to examine the effectiveness of the above-mentioned proposed approaches.
In this section, a comparison among the low discrepancies sequence, methods, is performed with each other with reference to capabilities and efficiency with the help of highdimensional fifteen benchmark functions. Nevertheless, the whole performance of optimization algorithms varies on the basis of setting parameters, and also with other testing criteria. Benchmark problems may be embeded to demonstrate the performance of the low discrepancies sequence approaches, at various complex levels. Table 6 contains the experimental simulation results on benchmark functions. The exhaustive statistical results are explained in Table 7. From Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38, the experimental results of constrained benchmark test functions are only exhibited by having the surface with D = 10, 20, and 30. The experimental results of this work may not contemplate the entire competency of the new proposed low discrepancies sequence in accordance with all the possible conditions.
The core objective of this section is to review the consequences of tested optimization approaches in high dimensionality, regarding the accuracy and reliability of achieved solutions at the time of solving complex and computationally-expensive optimization problems. From Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38, the performance of following methods: DE-KN, DE-WE, DE-TO, DE-S, DE-H, and traditional DE is compared for sixteen benchmark functions. In the graphs, the horizontal axis displays the total number of iterations, while on the other hand, the vertical axis displays the mean value of objective functions at the fixed number of function computations. Correspondingly, the value achieved in each iteration operates as a performance measure. As a result, the exploitation ability of traditional DE is moderately low, particularly for high-dimensional problems. The results are also disclosed that traditional DE, DE-S, DE-H are only effective in performance, while they tackle with expensive design problems having low dimensionality.
Beside this, DE-TO has excellent control on the high dimensionality problems than other methods in spite of complexity and the superficial topology of the examined problems. Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37 and Figure 38 show the achievements of traditional DE, DE-S, DE-H, DE-KN, and DE-WE algorithms with to regard to their efficiency and capability. The results demonstrate that DE-TO outperforms in higher dimensionality problems. By summarizing it, the dimensionality strongly influences the working of most algorithms, however, it is observed that DE-TO is more consistent during the increment of dimensions of the problem. Due to this consistency of DE-TO, it is proven that the DE-TO algorithm has greater capability of exploration.
For statistical comparison, a widely known mean ranks obtained by Kruskal–Wallis and Friedman tests is implemented to compare the implications between the DE-TO algorithm and other algorithms in DE-KN, DE-WE, DE-S, DE-H and standard DE are given in the Table 7.

5.3. Discussion on BA Results

The initialization technique plays a vital role in evolutionary and swarm-based stochastic algorithms. As, the traditional BA is not good in the process of global search. Therefore, the performance of BA can be increased by assigning the robust initial fitness to the particles. This may cause the enhancement in the diversity of swarm. For improving the performance of BA in terms of minimizing the global solution, we proposed three novel techniques of population initialization: the Knuth sequence-based BA (BA-KN), the Torus sequence-based-BA (BA-TO), and the WELL sequence-based BA (BA-WE).
For evolutionary comparison, the proposed techniques: Knuth sequence-based BA(BA-KN), the Torus sequence-based BA (BA-TO), and the WELL sequence-based BA (BA-WE) are compared with BA with Halton distribution (BA-HA), BA with Sobol distribution (BA-SO), are tested on sixteen standard benchmark functions and compared with the standard BA. The experimental results showed that the proposed techniques, BA based on Halton distribution (BA-HA), BA with Sobol distribution (BA-SO), and BA based on Torus sequence (BA-TO), perform better as compared to the standard BA, where among these three proposed techniques, (BA-TO) outperforms the others such as (BA-HA) and the (BA-SO). The convergence curves for all the techniques are presented in Figure 39, Figure 40, Figure 41, Figure 42, Figure 43, Figure 44, Figure 45, Figure 46, Figure 47, Figure 48, Figure 49, Figure 50, Figure 51, Figure 52, Figure 53 and Figure 54. The results depict that proposed techniques can help BA to avoid from premature convergence and to find the optimum solution quickly in the search space. The performance of the proposed techniques (BA-KN), (BA-WE), and (BA-TO) are compared with the standard BA and (BA-SO), (BA-HA), and (BA-TO) on standard benchmark functions with different dimensional size. It can be concluded from results that quasi-random sequences are best to create the random number sequences for BA and also for other population-based algorithms.
In this work, the primary concern is to reach the optimal solution, which is 0 in the ideal case. We investigated the different distribution approaches such as Knuth, WELL, Torus, Sobol, Halton, and random to initialize the BA for ensuring the swarm diversity in the very initial stage of the process. It is observed from Table 8 that Knuth distribution-based BA initialization gives better results as compared to the other quasi-random sequences. To validate the numerical results, mean ranks obtained by Kruskal–Wallis and Friedman tests for BA-KN, BA-WE, BATO, BA-SO, BA-HA, and standard BA are given in the Table 9.

6. Comparison of PSO, BA, and DE Regarding Data Classification

6.1. NN Classifications with PSO-Based Initialization Approaches

For further verification of performance of proposed algorithms TO-PSO, WE-PSO, and KN-PSO, a comparative study for real world benchmark datasets problem is tested for training of the neural network. We performed experiments using seven benchmark datasets (Diabetes, Heart, Wine, Seed, Vertebral, Blood Tissue, and Mammography) exerted from the worldwide famous machine-learning repository of UCI. Training weights are initialized within interval [−50, 50]. Accuracy of the feed forward neural network is tested in the form of root mean squared error (RMSE). Table 10 shows the characteristics of the datasets used.

Discussion

The multi-layer feed forward neural network is trained with the back propagation algorithm, standard PSO, SO-PSO, H-PSO, and proposed TO-PSO, KN-PSO, and WE-PSO. Comparison of these training approaches tested on real classification problem datasets are taken from the UCI repository. The cross validation method is used to compare the performances of different classification techniques. In the paper, the k-fold cross validation method used for the comparison of classification performances for the training of the neural network with back propagation, standard PSO, SO-PSO, H-PSO, and proposed TO-PSO, KN-PSO, and WE-PSO is used. The k-fold cross validation method was proposed and used in the experimental with value k = 10. The dataset divided into 10 chunks where each chunk of data contains the same proportion of each class of dataset. One chunk is used as testing while nine chunks are used as the training phase. The experimental results of algorithms such as with back propagation, standard PSO, SOPSO, H-PSO, and proposed TO-PSO, KN-PSO, and WE-PSO are compared with each other on seven well-known real datasets taken from UCI and their performances are evaluated. In Table 11, the simulation results show that the training of the neural network with the KN-PSO algorithm outperforms in accuracy and is capable to provide good classification accuracy than the other traditional approaches. The KN-PSO algorithm may be used effectively for data classification and statistical problems in the future as well. Figure 55 represents the accuracy graph for seven datasets.
The classification testing accuracy were imported from Microsoft Excel Spreadsheet to the software RStudio version 1.2.5001 to get assurance of the winner approach statistically among all the other approaches. The testing accuracy of all seven variants of PSONN were analyzed by the one-way ANOVA test and post-hoc Tukey’s multi-comparison test [60] having a 0.05 significance level. Table 12 depicts the results of one-way ANOVA of the testing accuracy of classification data. The significance value in Table 11 is 0.04639 which is less than 0.05, giving evidence that there is a significant difference among all variants of PSONN with a 95% confidence level. According to this, the variants of PSONN are significantly distinct from each other. Figure 56 represents the graph of one-way ANOVA results, which conclude that KN-PSONN significantly outperforms than all other variants of PSONN. Figure 57 represents the results of multi-comparisons of PSONN variants through the post-hoc Tukey’s test. The resultant graph depicts that the KN-PSONN variant is significantly different from all other variants. According to the results in Figure 55, KN-PSONN is proved statistically different from all other approaches of PSONN with a 95% confidence level.

6.2. NN Classifications with DE-Based Initialization Approaches

The proposed approaches, DE-KN, DE-TO, and DE-WE and family of low discrepancy sequences, are extremely suitable for tackling global optimization problems. A comparative study for real-world benchmark datasets problems is tested for the training of the neural network. We performed experiments using seven benchmark datasets (Diabetes, Heart, Wine, Seed, Vertebral, Blood Tissue, and Mammography) exerted from the worldwide famous machine-learning repository of UCI. Training weights are initialized within the interval [−50, 50]. Accuracy of the feed-forward neural network is tested in the form of root mean squared error (RMSE).

Discussion

The multi-layer feed-forward neural network is trained with the back propagation algorithm, standard DE, DE-S, DE-H, and proposed DE-TO, DE-KN, and DE-WE. For this goal, we prepared the multi-layer feed-forward neural network utilizing the process of weight optimization. The performance of the DE, DE-S, DE-H, DE-TO, DE-KN, and DE-WE and state of the art NN algorithms are examined on 10 well-known datasets which have been taken directly from the worldwide UCI repository of machine learning. The features of those informational indexes are given in Table 10. These features include the total units participated against each dataset, the number of total input instances, the dataset nature, and the number of classes against each dataset i.e., binary class problem or multi-class problem. The impact of increasing the number of target classes is independent as the proposed strategy is purely concerned with weight optimization rather feature selection or reducing high dimensionality. The 10-fold cross validation method has been carried out for the training and testing process. The experimental results of algorithms such as with back propagation, standard DE, DE-S, DE-H, DE-WE, DE-TO, and DE-KN are compared with each other on seven well-known real datasets taken from UCI and their performances are evaluated. In Table 13, the simulation results show that training of neural networks with the DE-H algorithm outperforms in accuracy and is capable of providing the good classification accuracy than the other traditional approaches. The DE-H algorithm may be used effectively for data classification and statistical problems in the future as well. Figure 58 represents the accuracy graph for seven datasets.
To prove the experimental results statistically, the testing accuracy of classification datasets were loaded to the software RStudio (1.2.5001 version). The classification results of seven approaches of DE were tested along with the one-way ANOVA statistical test and post-hoc Tukey’s pair-wise comparison statistical test [60] using significance level 0.05. The findings of the classification dataset with one-way ANOVA are illustrated in Table 14, where the significance = 0.02043 is less than the above-mentioned threshold of significance level. The findings in Table 14 prove that there are significant dissimilarities in all variants of DE with a 95% confidence level. Figure 59 demonstrates the graph of one-way ANOVA which gives the evidence that H-DE is significantly better than other approaches of DE. Figure 60 models the findings of pairwise comparisons of DE approaches with the post-hoc Tukey’s statistical test. The simulated graph describes that the H-DE approach is statistically significant dissimilar as compared to other approaches of DE having a 95% confidence level.

6.3. NN Classifications with BA-Based Initialization Approaches

The multi-layer feed forward neural network is trained with the back propagation algorithm, standard BA, BA-SO, BA-H, and proposed BA-TO, BA-KN, and BA-WE. The comparison of these training approaches is tested on real classification problem datasets taken from the UCI repository. The cross validation method used to compare the performances of different classification techniques. In the paper, the k-fold cross validation method used for comparison of classification performances for the training of the neural network with back propagation, standard BA, BA-SO, BA-H, and proposed BA-TO, BAKN, and BAWE is used. The k-fold cross validation method was proposed and used in the experimental with value k = 10. The dataset is divided into 10 chunks where each chunk of data contains the same proportion of each class of dataset. One chunk is used as testing while nine chunks are used as training phase. The experimental results of algorithms such as with back propagation, standard BA, BA-SO, BA-H, and proposed BA-TO, BA-KN, and BA-WE are compared with each other on seven well-known real datasets taken from UCI and their performances are evaluated. In Table 15, the simulation results show that training of the neural network with the BA-KN algorithm outperforms in accuracy and is capable of providing good classification accuracy than the other traditional approaches. The BA-WE algorithm may be used effectively for data classification and statistical problem in the future as well. Figure 61 represents the accuracy graph for seven datasets.
For giving the evidence of simulation results, the results of seven variants of BA initialization obtained from classification were examined by the statistical tests such as one-way ANOVA and post-hoc Tukey tests [60] for pair-wise likeliness (pair-wise comparisons) under the condition of 0.05 significance level. The outcomes of the one-way ANOVA test are presented in Table 16, which show the significance level less than 0.05 is 0.03623. The outcomes of Table 16 revealed that there is significant divergence in all initialization variants of BA with a 95% confidence level. Figure 62 displays the graph of one-way ANOVA which gives proof that KN-BANN is significantly superior to the other initialization variants of BANN. Figure 63 displays the outcomes of pair-wise likeliness comparisons (pair-wise comparisons) of initialization variants of BANN by using post-hoc Tukey’s statistical test. The plotted graph shows that the KN-BANN initialization variant is statistically divergent than other initialization variants of BANN having a 95% confidence level.

7. Conclusions

This paper introduces the new WELL sequence, Knuth sequence, and Torus sequence pseudorandom initialization strategies that are used to initialize the population in PSO, BA, and DE algorithms. Using the low discrepancy sequence family, the theoretical validation of the suggested methods is assessed on a robust suite of benchmark test functions and artificial neural network learning. The results of the simulation show that the use of the low discrepancy sequence family preserves the swarm’s diversity, increases the pace of convergence, and identifies a better swarm area. The suggested low discrepancy sequence families contain wider diversity and improved local searchability. The experimental findings indicate that KN-PSO, BA-KN, and DE-H have excellent convergence precision and improved avoidance of local optima. The proposed methods are contrasted with both a random distribution family of low discrepancy sequence approaches and traditional algorithms for PSO, BA, and DE, producing better performance. According to our analysis, an inference can be drawn that the quasi-random sequence for all population-based algorithms is substantially stronger and more feasible. Our goal is to work on higher-dimensional problems and constrained optimization problems for future perspectives. Moreover, we have not improved other additional algorithm operators such as mutations in this study. However, the results of such operators on low discrepancy sequences would be fascinating to examine. The main goal of this research is to extend to other stochastic meta-heuristic algorithms that establish the future directions of our work.

Author Contributions

Formal analysis, K.N. and D.B.R.; Investigation, W.H.B.; Methodology, S.P.; Project administration, W.H.B.; Resources, J.j.P.C.R.; Software, W.H.B.; Validation, A.A.A.I.; Writing—original draft, A.A.; Writing—review & editing, K.N. All authors have read and agreed to the published version of the manuscript.

Funding

The manuscript APC is supported by Universiti Malaysia Sabah, Jalan UMS, 88400, KK, Malaysia. Furthermore, this work is partially funded by FCT/MCTES through national funds and when applicable co-funded EU funds under the Project UIDB/50008/2020; and by Brazilian National Council for Scientific and Technological Development—CNPq, via Grant No. 313036/2020-9.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krishna, G.J.; Ravi, V. Mining top high utility association rules using binary differential evolution. Eng. Appl. Artif. Intell. 2020, 96, 103935. [Google Scholar] [CrossRef]
  2. Baró, G.B.; Martínez-Trinidad, J.F.; Rosas, R.M.V.; Ochoa, J.A.C.; González, A.Y.R.; Cortés, M.S.L. A pso-based algorithm for mining association rules using a guided exploration strategy. Pattern Recognit. Lett. 2020, 138, 8–15. [Google Scholar] [CrossRef]
  3. Fister, I.; Fong, S.; Brest, J. A novel hybrid self-adaptive bat algorithm. Sci. World J. 2014, 2014, 1–12. [Google Scholar] [CrossRef] [Green Version]
  4. Mandal, J.K.; Dutta, P.; Mukhopadhyay, S. Advances in Intelligent Computing; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  5. Drugan, M.M. Reinforcement learning versus evolutionary computation: A survey on hybrid algorithms. Swarm Evol. Comput. 2019, 44, 228–246. [Google Scholar] [CrossRef]
  6. Liu, J.; Abbass, H.A.; Tan, K.C. Evolutionary Computation; Springer: Berlin/Heidelberg, Germany, 2019; pp. 3–22. [Google Scholar]
  7. Zou, F.; Wang, L.; Hei, X.; Chen, D.; Yang, D. Teaching–learning-based optimization with dynamic group strategy for global optimization. Inf. Sci. 2014, 273, 112–131. [Google Scholar] [CrossRef]
  8. Davis, L. Handbook of Genetic Algorithms; Van Nostrand Reinhold: New York, NY, USA, 1991. [Google Scholar]
  9. Storn, R.; Price, K. Differential evolution–A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  10. Li, T.; Dong, H.; Sun, J. Binary differential evolution based on individual entropy for feature subset optimization. IEEE Access 2019, 7, 24109–24121. [Google Scholar] [CrossRef]
  11. Lei, Y.-X.; Gou, J.; Wang, C.; Luo, W.; Cai, Y.-Q. Improved differential evolution with a modified orthogonal learning strategy. IEEE Access 2017, 5, 9699–9716. [Google Scholar] [CrossRef]
  12. Meng, Z.; Yang, C.; Li, X.; Chen, Y. Di-de: Depth information-based differential evolution with adaptive parameter control for numerical optimization. IEEE Access 2020, 8, 40809–40827. [Google Scholar] [CrossRef]
  13. Tawhid, M.A.; Ali, A.F. Multi-directional bat algorithm for solving unconstrained optimization problems. Opsearch 2017, 54, 684–705. [Google Scholar] [CrossRef]
  14. Kolias, C.; Kambourakis, G.; Maragoudakis, M. Swarm intelligence in intrusion detection: A survey. Comput. Secur. 2011, 30, 625–642. [Google Scholar] [CrossRef]
  15. Wang, Y.; Wang, P.; Zhang, J.; Cui, Z.; Cai, X.; Zhang, W.; Chen, J. A novel bat algorithm with multiple strategies coupling for numerical optimization. Mathematics 2019, 7, 135. [Google Scholar] [CrossRef] [Green Version]
  16. Xue, F.; Cai, Y.; Cao, Y.; Cui, Z.; Li, F. Optimal parameter settings for bat algorithm. Int. J. Bio Inspired Comput. 2015, 7, 125–128. [Google Scholar] [CrossRef]
  17. Cui, Z.; Li, F.; Zhang, W. Bat algorithm with principal component analysis. Int. J. Mach. Learn. Cybern. 2019, 10, 603–622. [Google Scholar] [CrossRef]
  18. Chen, G.; Qian, J.; Zhang, Z.; Sun, Z. Applications of novel hybrid bat algorithm with constrained pareto fuzzy dominant rule on multi-objective optimal power flow problems. IEEE Access 2019, 7, 52060–52084. [Google Scholar] [CrossRef]
  19. Eberhart, R.; Kennedy, J. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; Citeseer: Princeton, NJ, USA, 2002; Volume 4, pp. 1942–1948. [Google Scholar]
  20. Sakri, S.B.; Rashid, N.B.A.; Zain, Z.M. Particle swarm optimization feature selection for breast cancer recurrence prediction. IEEE Access 2018, 6, 29637–29647. [Google Scholar] [CrossRef]
  21. Arora, S.; Singh, S. Butterfly optimization algorithm: A novel approach for global optimization. Soft Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  22. Abbas, G.; Gu, J.; Farooq, U.; Asad, M.U.; El-Hawary, M. Solution of an economic dispatch problem through particle swarm optimization: A detailed survey-part i. IEEE Access 2017, 5, 15105–15141. [Google Scholar] [CrossRef]
  23. Laskar, N.M.; Guha, K.; Chatterjee, I.; Chanda, S.; Baishnab, K.L.; Paul, P.K. Hwpso: A new hybrid whale-particle swarm optimization algorithm and its application in electronic design optimization problems. Appl. Intell. 2019, 49, 265–291. [Google Scholar] [CrossRef]
  24. Al-Betar, M.A.; Awadallah, M.A. Island bat algorithm for optimization. Expert Syst. Appl. 2018, 107, 126–145. [Google Scholar] [CrossRef]
  25. Cervantes, A.; Galván, I.M.; Isasi, P. Ampso: A new particle swarm method for nearest neighborhood classification. IEEE Trans. Syst. Man Cybern. Part B 2009, 39, 1082–1091. [Google Scholar] [CrossRef] [Green Version]
  26. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. Novel chaotic grouping particle swarm optimization with a dynamic regrouping strategy for solving numerical optimization tasks. Knowl. Based Syst. 2020, 194, 105568. [Google Scholar] [CrossRef]
  27. Yüzgeç, U.; Eser, M. Chaotic based differential evolution algorithm for optimization of baker’s yeast drying process. Egypt. Inform. J. 2018, 19, 151–163. [Google Scholar] [CrossRef]
  28. Jordehi, A.R. Chaotic bat swarm optimisation (CBSO). Appl. Soft Comput. 2015, 26, 523–530. [Google Scholar] [CrossRef]
  29. Grosan, C.; Abraham, A.; Nicoara, M. Search optimization using hybrid particle sub-swarms and evolutionary algorithms. Int. J. Simul. Syst. Sci. 2005, 6, 60–79. [Google Scholar]
  30. Bhat, C.R. Simulation estimation of mixed discrete choice models using randomized and scrambled halton sequences. Transp. Res. Part B Methodol. 2003, 37, 837–855. [Google Scholar] [CrossRef] [Green Version]
  31. Sobol’, I.M. On the distribution of points in a cube and the approximate evaluation of integrals. Zhurnal Vychislitel’noi Matematiki i Matematicheskoi Fiziki 1967, 7, 784–802. [Google Scholar] [CrossRef]
  32. Lazzús, J.A.; Vega-Jorquera, P.; López-Caraballo, C.H.; Palma-Chilla, L.; Salfate, I. Parameter estimation of a generalized lotka–volterra system using a modified pso algorithm. Appl. Soft Comput. 2020, 96, 106606. [Google Scholar] [CrossRef]
  33. Knuth, D.E. Fundamental Algorithms; Addison-Wesley: Boston, MA, USA, 1973. [Google Scholar]
  34. Gentle, J.E. Random Number Generation and Monte Carlo Methods; Springer Science & Business Media: Berlin, Germany, 2006. [Google Scholar]
  35. Wang, X.; Sloan, I.H. Low discrepancy sequences in high dimensions: How well are their projections distributed? J. Comput. Appl. Math. 2008, 213, 366–386. [Google Scholar] [CrossRef] [Green Version]
  36. Ali, M.H.; al Mohammed, B.A.D.; Ismail, A.; Zolkipli, M.F. A new intrusion detection system based on fast learning network and particle swarm optimization. IEEE Access 2018, 6, 20255–20261. [Google Scholar] [CrossRef]
  37. Shi, Y.; Eberhart, R. A modified particle swarm optimizer. In Proceedings of the 1998 IEEE International Conference on Evolutionary Computation, IEEE World Congress on Computational Intelligence (Cat. No. 98TH8360), Anchorage, AK, USA, 4–9 May 1998; pp. 69–73. [Google Scholar]
  38. Bangyal, W.H.; Ahmad, J.; Rauf, H.T. Optimization of neural network using improved bat algorithm for data classification. J. Med. Imaging Health Inform. 2019, 9, 670–681. [Google Scholar] [CrossRef]
  39. Sacco, W.F.; Rios-Coelho, A.C. On Initial Populations of Differential Evolution for Practical Optimization Problems. In Computational Intelligence, Optimization and Inverse Problems with Applications in Engineering; Springer: Berlin/Heidelberg, Germany, 2019; pp. 53–62. [Google Scholar]
  40. Devika, K.; Jeyakumar, G. Solving multi-objective optimization problems using differential evolution algorithm with different population initialization techniques. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–5. [Google Scholar]
  41. Parsopoulos, K.E.; Vrahatis, M.N. Initializing the particle swarm optimizer using the nonlinear simplex method. Adv. Intell. Syst. Fuzzy Syst. Evol. Comput. 2002, 216, 1–6. [Google Scholar]
  42. Richards, M.; Ventura, D. Choosing a starting configuration for particle swarm optimization. In Proceedings of the International Joint Conference on Neural, Budapest, Hungary, 25–29 July 2004; IEEE: Piscataway, NJ, USA, 2005; Volume 3, pp. 2309–2312. [Google Scholar]
  43. Uy, N.Q.; Hoai, N.X.; McKay, R.I.; Tuan, P.M. Initialising pso with randomised low-discrepancy sequences: The comparative results. In Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, 25–28 September 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 1985–1992. [Google Scholar]
  44. Pant, M.; Thangaraj, R.; Grosan, C.; Abraham, A. Improved Particle Swarm Optimization with Low-Discrepancy Sequences. In Proceedings of the 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence), Piscataway, NJ, USA, 1–6 June 2008; pp. 3011–3018. [Google Scholar]
  45. Pant, M.; Thangaraj, R.; Singh, V.P.; Abraham, A. Particle Swarm Optimization Using Sobol Mutation. In Proceedings of the 2008 First International Conference on Emerging Trends in Engineering and Technology, Nagpur, India, 16–18 July 2008; pp. 367–372. [Google Scholar]
  46. Du, J.; Zhang, F.; Huang, G.; Yang, J. A new initializing mechanism in particle swarm optimization. In Proceedings of the 2011 IEEE International Conference on Computer Science and Automation Engineering, Shanghai, China, 10–12 June 2011; IEEE: Piscataway, NJ, USA, 2011; Volume 4, pp. 325–329. [Google Scholar]
  47. Murugan, P. Modified particle swarm optimisation with a novel initialisation for finding optimal solution to the transmission expansion planning problem. IET Gener. Transm. 2012, 6, 1132–1142. [Google Scholar] [CrossRef]
  48. Yin, L.; Hu, X.-M.; Zhang, J. Space-based initialization strategy for particle swarm optimization. In Proceedings of the 15th Annual Conference Companion on Genetic and Evolutionary Computation, Amsterdam, The Netherlands, 6–10 July 2013; ACM: New York, NY, USA, 2010; pp. 19–20. [Google Scholar]
  49. Shatnawi, M.; Nasrudin, M.F.; Sahran, S. A new initialization technique in polar coordinates for particle swarm optimization and polar pso. Int. J. Adv. Sci. Eng. Inf. Technol. 2017, 7, 242–249. [Google Scholar] [CrossRef]
  50. Bewoor, L.; Prakash, V.C.; Sapkal, S. Evolutionary hybrid particle swarm optimization algorithm for solving np-hard no-wait flow shop scheduling problems. Algorithms 2017, 10, 121. [Google Scholar] [CrossRef] [Green Version]
  51. Albeahdili, H.M.; Han, T.; Islam, N.E. Hybrid algorithm for the optimization of training convolutional neural network. Int. J. Adv. Comput. Sci. Appl. 2015, 1, 79–85. [Google Scholar]
  52. Ali, M.; Pant, M.; Abraham, A. Simplex differential evolution. Acta Polytech. Hung. 2009, 6, 95–115. [Google Scholar]
  53. Nakib, A.; Daachi, B.; Siarry, P. Hybrid differential evolution using low-discrepancy sequences for image segmentation. In Proceedings of the 2012 IEEE 26th International Parallel and Distributed Processing Symposium Workshops & PhD Forum, Shanghai, China, 21–25 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 634–640. [Google Scholar]
  54. Tang, L.; Zhao, Y.; Liu, J. An improved differential evolution algorithm for practical dynamic scheduling in steelmaking-continuous casting production. IEEE Trans. Evol. Comput. 2013, 18, 209–225. [Google Scholar] [CrossRef]
  55. Dong, J.; Wang, Z.; Mo, J. A phase angle-modulated bat algorithm with application to antenna topology optimization. Appl. Sci. 2021, 11, 2243. [Google Scholar] [CrossRef]
  56. Dong, L.; Zeng, W.; Wu, L.; Lei, G.; Chen, H.; Srivastava, A.K.; Gaiser, T. Estimating the pan evaporation in northwest china by coupling catboost with bat algorithm. Water 2021, 13, 256. [Google Scholar] [CrossRef]
  57. Rodriguez-Molina, A.; Solis-Romero, J.; Villarreal-Cervantes, M.G.; Serrano-Perez, O.; Flores-Caballero, G. Path-planning for mobile robots using a novel variable-length differential evolution variant. Mathematics 2021, 9, 357. [Google Scholar] [CrossRef]
  58. Panneton, F.; L’ecuyer, P.; Matsumoto, M. Improved long-period generators based on linear recurrences modulo 2. ACM Trans. Math. Softw. 2006, 32, 1–16. [Google Scholar] [CrossRef]
  59. Nikulin, V.V.; Shafarevich, I.R. Geometries and Groups; Springer Science & Business Media: Berlin, Germany, 2012. [Google Scholar]
  60. Ulusoy, U. Application of anova to image analysis results of talc particles produced by different milling. Powder Technol. 2008, 188, 133–138. [Google Scholar] [CrossRef]
Figure 1. Population initialization using uniform distribution.
Figure 1. Population initialization using uniform distribution.
Applsci 11 08190 g001
Figure 2. Population initialization using Sobol distribution.
Figure 2. Population initialization using Sobol distribution.
Applsci 11 08190 g002
Figure 3. Population initialization using Halton distribution.
Figure 3. Population initialization using Halton distribution.
Applsci 11 08190 g003
Figure 4. Population initialization using WELL distribution.
Figure 4. Population initialization using WELL distribution.
Applsci 11 08190 g004
Figure 5. Population initialization using Knuth distribution.
Figure 5. Population initialization using Knuth distribution.
Applsci 11 08190 g005
Figure 6. Sample data generated using Torus distribution.
Figure 6. Sample data generated using Torus distribution.
Applsci 11 08190 g006
Figure 7. Convergence curve on F1.
Figure 7. Convergence curve on F1.
Applsci 11 08190 g007
Figure 8. Convergence curve on F2.
Figure 8. Convergence curve on F2.
Applsci 11 08190 g008
Figure 9. Convergence curve on F3.
Figure 9. Convergence curve on F3.
Applsci 11 08190 g009
Figure 10. Convergence curve on F4.
Figure 10. Convergence curve on F4.
Applsci 11 08190 g010
Figure 11. Convergence curve on F5.
Figure 11. Convergence curve on F5.
Applsci 11 08190 g011
Figure 12. Convergence curve on F6.
Figure 12. Convergence curve on F6.
Applsci 11 08190 g012
Figure 13. Convergence curve on F7.
Figure 13. Convergence curve on F7.
Applsci 11 08190 g013
Figure 14. Convergence curve on F8.
Figure 14. Convergence curve on F8.
Applsci 11 08190 g014
Figure 15. Convergence curve on F9.
Figure 15. Convergence curve on F9.
Applsci 11 08190 g015
Figure 16. Convergence curve on F10.
Figure 16. Convergence curve on F10.
Applsci 11 08190 g016
Figure 17. Convergence curve on F11.
Figure 17. Convergence curve on F11.
Applsci 11 08190 g017
Figure 18. Convergence curve on F12.
Figure 18. Convergence curve on F12.
Applsci 11 08190 g018
Figure 19. Convergence curve on F13.
Figure 19. Convergence curve on F13.
Applsci 11 08190 g019
Figure 20. Convergence curve on F14.
Figure 20. Convergence curve on F14.
Applsci 11 08190 g020
Figure 21. Convergence curve on F15.
Figure 21. Convergence curve on F15.
Applsci 11 08190 g021
Figure 22. Convergence curve on F16.
Figure 22. Convergence curve on F16.
Applsci 11 08190 g022
Figure 23. Convergence curve on F1.
Figure 23. Convergence curve on F1.
Applsci 11 08190 g023
Figure 24. Convergence curve on F2.
Figure 24. Convergence curve on F2.
Applsci 11 08190 g024
Figure 25. Convergence curve on F3.
Figure 25. Convergence curve on F3.
Applsci 11 08190 g025
Figure 26. Convergence curve on F4.
Figure 26. Convergence curve on F4.
Applsci 11 08190 g026
Figure 27. Convergence curve on F5.
Figure 27. Convergence curve on F5.
Applsci 11 08190 g027
Figure 28. Convergence curve on F6.
Figure 28. Convergence curve on F6.
Applsci 11 08190 g028
Figure 29. Convergence curve on F7.
Figure 29. Convergence curve on F7.
Applsci 11 08190 g029
Figure 30. Convergence curve on F8.
Figure 30. Convergence curve on F8.
Applsci 11 08190 g030
Figure 31. Convergence curve on F9.
Figure 31. Convergence curve on F9.
Applsci 11 08190 g031
Figure 32. Convergence curve on F10.
Figure 32. Convergence curve on F10.
Applsci 11 08190 g032
Figure 33. Convergence curve on F11.
Figure 33. Convergence curve on F11.
Applsci 11 08190 g033
Figure 34. Convergence curve on F12.
Figure 34. Convergence curve on F12.
Applsci 11 08190 g034
Figure 35. Convergence curve on F13.
Figure 35. Convergence curve on F13.
Applsci 11 08190 g035
Figure 36. Convergence curve on F14.
Figure 36. Convergence curve on F14.
Applsci 11 08190 g036
Figure 37. Convergence curve on F15.
Figure 37. Convergence curve on F15.
Applsci 11 08190 g037
Figure 38. Convergence curve on F16.
Figure 38. Convergence curve on F16.
Applsci 11 08190 g038
Figure 39. Convergence curve on F1.
Figure 39. Convergence curve on F1.
Applsci 11 08190 g039
Figure 40. Convergence curve on F2.
Figure 40. Convergence curve on F2.
Applsci 11 08190 g040
Figure 41. Convergence curve on F3.
Figure 41. Convergence curve on F3.
Applsci 11 08190 g041
Figure 42. Convergence curve on F4.
Figure 42. Convergence curve on F4.
Applsci 11 08190 g042
Figure 43. Convergence curve on F5.
Figure 43. Convergence curve on F5.
Applsci 11 08190 g043
Figure 44. Convergence curve on F6.
Figure 44. Convergence curve on F6.
Applsci 11 08190 g044
Figure 45. Convergence curve on F7.
Figure 45. Convergence curve on F7.
Applsci 11 08190 g045
Figure 46. Convergence curve on F8.
Figure 46. Convergence curve on F8.
Applsci 11 08190 g046
Figure 47. Convergence curve on F9.
Figure 47. Convergence curve on F9.
Applsci 11 08190 g047
Figure 48. Convergence curve on F10.
Figure 48. Convergence curve on F10.
Applsci 11 08190 g048
Figure 49. Convergence curve on F11.
Figure 49. Convergence curve on F11.
Applsci 11 08190 g049
Figure 50. Convergence curve on F12.
Figure 50. Convergence curve on F12.
Applsci 11 08190 g050
Figure 51. Convergence curve on F13.
Figure 51. Convergence curve on F13.
Applsci 11 08190 g051
Figure 52. Convergence curve on F14.
Figure 52. Convergence curve on F14.
Applsci 11 08190 g052
Figure 53. Convergence curve on F15.
Figure 53. Convergence curve on F15.
Applsci 11 08190 g053
Figure 54. Convergence curve on F16.
Figure 54. Convergence curve on F16.
Applsci 11 08190 g054
Figure 55. Classification Testing Accuracy Results.
Figure 55. Classification Testing Accuracy Results.
Applsci 11 08190 g055
Figure 56. Box plot visualization of the results achieved by the training of FFNN for all PSO-based initialization approaches and BPA for given datasets of the classification problem.
Figure 56. Box plot visualization of the results achieved by the training of FFNN for all PSO-based initialization approaches and BPA for given datasets of the classification problem.
Applsci 11 08190 g056
Figure 57. Multi-comparison post-hoc Tukey test graph of all PSO-based.
Figure 57. Multi-comparison post-hoc Tukey test graph of all PSO-based.
Applsci 11 08190 g057
Figure 58. Classification testing accuracy results.
Figure 58. Classification testing accuracy results.
Applsci 11 08190 g058
Figure 59. Box plot visualization of the results achieved by the training of FFNN for all DE-based initialization approaches and BPA for given datasets of classification problem.
Figure 59. Box plot visualization of the results achieved by the training of FFNN for all DE-based initialization approaches and BPA for given datasets of classification problem.
Applsci 11 08190 g059
Figure 60. Multi-comparison post-hoc Tukey test graph of all DE-based.
Figure 60. Multi-comparison post-hoc Tukey test graph of all DE-based.
Applsci 11 08190 g060
Figure 61. Classification testing accuracy results.
Figure 61. Classification testing accuracy results.
Applsci 11 08190 g061
Figure 62. Box plot visualization of the results achieved by the training of FFNN for all BA-based initialization approaches and BPA for given datasets of the classification problem.
Figure 62. Box plot visualization of the results achieved by the training of FFNN for all BA-based initialization approaches and BPA for given datasets of the classification problem.
Applsci 11 08190 g062
Figure 63. Multi-comparison post-hoc Tukey test graph of all BA-based.
Figure 63. Multi-comparison post-hoc Tukey test graph of all BA-based.
Applsci 11 08190 g063
Table 1. Experimental setting of parameters.
Table 1. Experimental setting of parameters.
ParameterValue
Search Space[100, −100]
Dimensions102030
Iterations100020003000
Population size50
Number of PSO Runs10
Table 2. Parameters setting of parameters.
Table 2. Parameters setting of parameters.
AlgorithmParameters
PSOc1=c2=1.49, w = linearly decreasing
BA N p =40, r i j t [ 0 , 1 ] , A i j t [ 0 , 2 ]
DEF ∈ [0.4, 1], CR ∈ 0.6
Table 3. Function table with characteristics.
Table 3. Function table with characteristics.
Sr.#Function NameObjective FunctionSearch SpaceOptimal Value
01Sphere M i n f ( x ) = i = 1 D x i 2 5.12 x i 5.12 0
02Rastrigin M i n f ( x ) = 10 D + i = 1 D [ x i 2 10 cos ( 2 π x ) ] i 5.12 x i 5.12 0
03Axis parallel hyper-ellipsoid M i n f ( x ) = i = 1 D ( i . x i 2 ) 5.12 x i 5.12 0
04Rotated hyper ellipsoid M i n f ( x ) = i = 1 D j = 1 i ( x j 2 ) 65.536 x i 65.536 0
05Moved Axis M i n f ( x ) = i = 1 D 5 i . x i 2 5.12 x i 5.12 0
06Sum of different power M i n f ( x ) = i = 1 D | x i | ( i + 1 ) 1 x i 1 0
07ChungReynolds M i n f ( x ) = ( i = 1 D x i 2 ) 2 100 x i 100 0
08Csendes M i n f ( x ) = i = 1 D x i 6 ( 2 + sin 1 x i ) 1 x i 1 0
09Schaffer M i n f ( x ) = 0.5 + sin 2 ( x 1 2 + x 2 2 ) 2 0.5 1 + 0.001 ( x 1 2 + x 2 2 ) 2 100 x i 100 0
10Schumer_Steiglitz M i n f ( x ) = i = 1 D x i 4 100     x i     100 0
11Schwefel M i n f ( x ) = ( i = 1 D x i 2 ) α 100 x i 100 0
12Schwefel1.2 M i n f ( x ) = i = 1 D ( j = 1 i x j ) 2 100 x i 100 0
13Schwefel 2.21 M i n f ( x ) = max 1 i D | x i | 100 x i 100 0
14Schwefel 2.22 M i n f ( x ) = i = 1 D | x i | + i = 1 D | x i | 100 x i 100 0
15Schwefel 2.23 M i n f ( x ) = i = 1 D x i 10 10 x i 10 0
16Zakharov M i n f ( x ) = i = 1 D x i 2 + ( 1 2 i = 1 n i x i ) 2 + ( 1 2 i = 1 n i x i ) 4 5 x i 10 0
Table 4. Comparative results for all PSO-based approaches on 16 standard benchmark functions.
Table 4. Comparative results for all PSO-based approaches on 16 standard benchmark functions.
FunctionsDIM × ItrPSOSO−PSOH−PSOTO−PSOWE−PSOKN−PSO
MeanMeanMeanMeanMeanMean
F110 × 10002.33 × 10−742.74 × 10−763.10 × 10−775.57 × 10−785.91 × 10−780.0000 × 10+00
20 × 20001.02 × 10−848.20 × 10−881.76 × 10−901.30 × 10−904.95 × 10−903.14001 × 10−217
30 × 30001.77 × 10−267.67 × 10−204.13 × 10−321.25 × 10−511.30 × 10−428.91595 × 10−88
F210 × 10004.97 × 10−014.97 × 10−017.96 × 10−013.98 × 10−012.98 × 10−01−8602.02
20 × 20008.17 × 10+006.47 × 10+003.58 × 10+002.89 × 10+003.11 × 10+00−31,433.3
30 × 30001.01 × 10+019.86 × 10+009.45 × 10+008.16 × 10+007.76 × 10+00−60,711.8
F310 × 10008.70 × 10−801.79 × 10−794.87 × 10−793.91 × 10−824.40 × 10−810.0000 × 10+00
20 × 20002.621447.864322.621447.07 × 10−901.78 × 10−894.78718 × 10−237
30 × 30002.62 × 10+011.57 × 10+011.05 × 10+017.70 × 10−353.87 × 10−571.57084 × 10−97
F410 × 10004.46 × 10−1473.86 × 10−1479.78 × 10−1457.29 × 10−1481.24 × 10−1500.0000 × 10+00
20 × 20003.14 × 10−1559.27 × 10−1542.75 × 10−1595.14 × 10−1584.96 × 10−1590.0000 × 10+00
30 × 30001.82 × 10−1332.36 × 10−1358.53 × 10−1303.13 × 10−1382.54 × 10−1361.6439 × 10−228
F510 × 10004.35 × 10−798.95 × 10−792.43 × 10−782.04 × 10−802.20 × 10−800.0000 × 10+00
20 × 20001.31 × 10+013.93 × 10+011.31 × 10+013.54 × 10−893.12 × 10−892.39359 × 10−236
30 × 30001.31 × 10+027.86 × 10+015.24 × 10+013.85 × 10−341.94 × 10−562.9093 × 10−87
F610 × 10001.70 × 10−614.45 × 10−647.29 × 10−662.46 × 10−664.62 × 10−663.04226 × 10−318
20 × 20003.25 × 10−1124.39 × 10−1125.01 × 10−1092.56 × 10−1154.45 × 10−1138.59557 × 10−277
30 × 30007.21 × 10−1354.10 × 10−1241.51 × 10−1346.22×10−1376.96 × 10−1352.33033 × 10−223
F710 × 10002.96 × 10−1572.39 × 10−1571.28 × 10−1574.89 × 10−1592.47 × 10−1630.0000 × 10+00
20 × 20008.79 × 10−1771.77 × 10−1843.49 × 10−1833.09 × 10−1873.41 × 10−1860.0000 × 10+00
30 × 30001.23 × 10−821.25 × 10−1165.99 × 10−1305.01 × 10−1354.60 × 10−1348.03288 × 10−175
F810 × 10004.39 × 10−2001.98 × 10−1944.51 × 10−1971.26 × 10−2028.99 × 10−2014.9228 × 10−67
20 × 20001.57 × 10−201.04 × 10−931.10 × 10−1482.84 × 10−1574.09 × 10−1514.5887 × 10−16
30 × 30001.89 × 10−094.54 × 10−101.14 × 10−081.40 × 10−101.34 × 10−092.2334 × 10−08
F910 × 10005.49 × 10−011.30 × 10−012.02 × 10−011.26 × 10−011.42 × 10−010.824968
20 × 20002.05 × 10+007.83 × 10−016.83 × 10−015.84 × 10−014.32 × 10−014.56265
30 × 30001.12 × 10+009.99 × 10−019.56 × 10−019.06 × 10−019.12 × 10−017.25675
F1010 × 10002.23 × 10−1382.23 × 10−1384.35 × 10−1371.02 × 10−1401.10 × 10−1390.0000 × 10+00
20 × 20003.79 × 10−1487.87 × 10−1494.19 × 10−1473.78 × 10−1518.73 × 10−1530.0000 × 10+00
30 × 30004.43 × 10−1267.52 × 10−1331.57 × 10−1282.03 × 10−1341.38 × 10−1332.26229 × 10−221
F1110 × 10003.75 × 10−1871.57 × 10−1922.15 × 10−1915.57 × 10−1988.99 × 10−1980.0000 × 10+00
20 × 20005.29 × 10−1932.53 × 10−1958.45 × 10−1958.45 × 10−1959.83 × 10−1970.0000 × 10+00
30 × 30004.82 × 10−1548.84 × 10−1595.49 × 10−1682.04 × 10−1705.75 × 10−1739.00586 × 10−278
F1210 × 10001.13 × 10−011.67 × 10−022.28 × 10−024.78 × 10−032.89 × 10−032.739 × 10−12
20 × 20001.39 × 10+015.03 × 10+002.95 × 10+001.28 × 10+001.67 × 10+007.819 × 10+00
30 × 30007.45 × 10+001.22 × 10+018.74 × 10+002.94 × 10+004.94 × 10+002.239 × 10+01
F1310 × 10008.04 × 10−268.01 × 10−273.59 × 10−271.24 × 10−271.41 × 10−270.0000 × 10+00
20 × 20001.42 × 10−082.64 × 10−113.29 × 10−102.99 × 10−102.14 × 10−120.0000 × 10+00
30 × 30006.20 × 10−031.41 × 10−039.36 × 10−031.12 × 10−031.41 × 10−030.0000 × 10+00
F1410 × 10003.62 × 10−383.62 × 10−385.92 × 10−366.92 × 10−391.95 × 10−387.78286 × 10−197
20 × 20006.27 × 10−101.38 × 10−097.91 × 10−132.49 × 10−121.17 × 10−136.6163 × 10−12
30 × 30002.56 × 10−064.80 × 10+011.34 × 10−065.40 × 10−114.88 × 10−099.3032 × 10−06
F1510 × 10001.10 × 10−2943.19 × 10−3012.78 × 10−3071.94 × 10−3073.21 × 10−3086.26612 × 10−138
20 × 20006.16 × 10−2715.09 × 10−2763.74 × 10−2701.60 × 10−2764.85 × 10−2681.29033 × 10−25
30 × 30003.08 × 10−2071.04 × 10−2008.12 × 10−2092.34 × 10−2153.06 × 10−2122.27 × 10−06
F1610 × 10005.48353858.5299 × 10 3.3074 × 10−161.2248038.3354 × 10−072.26476 × 10−27
20 × 200083.4671.63440.1803749.168415.13227.17014 × 10−72
30 × 3000265.90708282.186445.0408133.967967.03015.45179 × 10−251
Table 5. Mean ranks obtained by Kruskal–Wallis and Friedman tests for all.
Table 5. Mean ranks obtained by Kruskal–Wallis and Friedman tests for all.
ApproachesFriedman Valuep-ValueKruskal–Wallisp-Value
PSO39.090.00139.330.001
SO-PSO37.470.00138.390.001
H-PSO38.500.00138.910.001
TO-PSO41.790.00042.670.000
WE-PSO41.880.00042.500.000
KN-PSO18.240.00123.310.002
Table 6. Comparative results for all DE-based approaches on 16 standard benchmark functions.
Table 6. Comparative results for all DE-based approaches on 16 standard benchmark functions.
FunctionsDIM × IterDEDE−HDE−SDE−TODE−WEDE−KN
F110 × 10001.1464 × 10−442.1338 × 10−445.8561 × 10−447.4117 × 10−457.4827 × 10−395.7658 × 10−39
20 × 20003.3550 × 10−467.2338 × 10−461.3545 × 10−451.2426 × 10−459.6318 × 10−457.1501 × 10−45
30 × 30008.8946 × 10−471.2273 × 10−459.4228 × 10−461.6213 × 10−466.2007 × 10−465.7425 × 10−46
F210 × 10000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
20 × 20000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
30 × 30001.8392 × 10+011.1846 × 10+011.8871 × 10+013.7132 × 10−015.0821 × 10+006.6313 × 10+00
F310 × 10005.00325 × 10−441.5019 × 10−389.3956 × 10−444.7807 × 10−441.6251 × 10−381.3411 × 10−38
20 × 20002.56987 × 10−454.1485 × 10−441.5339 × 10−443.0262 × 10−459.5984 × 10−441.3606 × 10−43
30 × 30001.01692 × 10−452.7349 × 10−454.0581 × 10−454.5726 ×10−454.5686 × 10−455.4659 × 10−45
F410 × 10005.81825 × 10−423.0950 × 10−362.2300 × 10−411.6903 × 10−411.1331 × 10−363.8869 × 10−36
20 × 20002.70747 × 10−431.0658 × 10−411.6730 × 10−421.3490 × 10−421.3094 × 10−416.0053 × 10−42
30 × 30002.99887 × 10−431.4032 × 10−424.4442 × 10−425.9186 × 10−434.6922 × 10−431.4829 × 10−42
F510 × 10001.65318 × 10−434.7939 × 10−387.0329 × 10−434.8106 × 10−434.3219 × 10−383.5770 × 10−38
20 × 20001.39082 × 10−443.6325 × 10−434.2191 × 10−442.7448 × 10−445.8557 × 10−431.4008 × 10−43
30 × 30006.07162 × 10−451.7557 × 10−441.6295 × 10−442.0582 × 10−448.6773 × 10−454.2285 × 10−44
F610 × 10007.8201 × 10−963.8819 × 10−969.7956 × 10−962.3292 × 10−958.4774 × 10−942.8037 × 10−95
20 × 20001.6847 × 10−1258.6880 × 10−1245.9005 × 10−1228.7800 × 10−1233.7438 × 10−1241.3947 × 10−124
30 × 30002.4533 × 10−1401.5487 × 10−1395.7211 × 10−1384.4492 × 10−1376.5749 × 10−1403.4442 × 10−137
F710 × 10008.0217 × 10−757.3243 × 10−675.7807 × 10−661.0243 × 10−731.9035 × 10−671.4359 × 10−65
20 × 20004.0682 × 10−711.5037 × 10−701.5747 × 10−691.0623 × 10−705.5546 × 10−702.3507 × 10−70
30 × 30008.5895 × 10−686.6009 × 10−683.3919 × 10−672.6036 × 10−671.1587 × 10−672.1901 × 10−67
F810 × 10007.0221 × 10−1203.4271 × 10−1082.7718 × 10−1086.3092 × 10−1183.9423 × 10−1069.9394 × 10−108
20 × 20005.2096 × 10−1087.7158 × 10−891.4732 × 10−1068.8720 × 10−1073.4490 × 10−1072.2539 × 10−106
30 × 30001.2538 × 10−981.8071 × 10−981.1085 × 10−957.2462 × 10−982.5375 × 10−995.8040 × 10−98
F910 × 10000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
20 × 20000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
30 × 30000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+000.0000 × 10+00
F1010 × 10001.3459 × 10−752.6493 × 10−662.6884 × 10−663.6168 × 10−673.8397 × 10−671.8408 × 10−66
20 × 20003.0478 × 10−711.6106 × 10−695.5253 × 10−692.7746 × 10−705.3662 × 10−705.0931 × 10−70
30 × 30008.2514 × 10−681.0937 × 10−664.1120 × 10−671.3055 × 10−673.5397 × 10−686.073 × 10−67
F1110 × 10002.3417 × 10−421.2483 × 10−411.3726 × 10−416.3337 × 10−424.8161 × 10−427.34640 × 10−42
20 × 20008.4769 × 10−443.5140 × 10−433.3777 × 10−432.4721 × 10−431.9553 × 10−433.6961 × 10−43
30 × 30003.6888 × 10−446.9938 × 10−442.5123 × 10−431.4710 × 10−434.0019 × 10−443.9503 × 10−43
F1210 × 10002.3304 × 10+004.4354 × 10+003.4520 × 10+005.1229 × 10+003.8782 × 10+002.7840 × 10+00
20 × 20003.1768 × 10+043.9596 × 10+043.8814 × 10+042.9488 × 10+044.1181 × 10+044.0914 × 10+04
30 × 30001.1760 × 10+061.0300 × 10+061.3402 × 10+061.2008 × 10+061.0916 × 10+061.0160 × 10+06
F1310 × 10001.3940 × 10−651.3756 × 10−643.1956 × 10−669.3609 × 10−645.4864 × 10−639.2695 × 10−63
20 × 20002.0163 × 10−1118.5333 × 10−1108.5260 × 10−1113.9836 × 10−1095.0102 × 10−1154.4624 × 10−110
30 × 30001.4146 × 10−1564.3434 × 10−1564.4702 × 10−1544.3862 × 10−1511.0781 × 10−1531.0142 × 10−149
F1410 × 10009.1259 × 10−242.1900 × 10−232.5559 × 10−232.9039 × 10−231.9174 × 10−233.3427 × 10−23
20 × 20002.6867 × 10−253.8631 × 10−251.5177 × 10−245.5714 × 10−254.5049 × 10−255.6503 × 10−25
30 × 30005.9241 × 10−268.6401 × 10−268.4348 × 10−261.4630 × 10−259.7932 × 10−261.4921 × 10−25
F1510 × 10001.0493 × 10−1854.0276 × 10−1815.0331 × 10−1823.1770 × 10−1831.1698 × 10−1802.6563 × 10−182
20 × 20002.9407 × 10−1599.9152 × 10−1592.1401 × 10−1589.0345 × 10−1563.8871 × 10−1588.0144 × 10−160
30 × 30004.6769 × 10−1381.0737 × 10−1377.0544 × 10−1388.0376 × 10−1384.9091 × 10−1391.1054 × 10−137
F1610 × 10001.8635 × 10−041.8109 × 10−024.9798 × 10−025.8605 × 10−041.4858 × 10−023.7220 × 10−02
20 × 20001.1032 × 10+001.6605 × 10+001.7157 × 10+001.4875 × 10+001.5697 × 10+001.2008 × 10+00
30 × 30002.8283 × 10+012.2049 × 10+012.9388 × 10+012.8205 × 10+012.5794 × 10+012.9526 × 10+01
Table 7. Mean ranks obtained by Kruskal–Wallis and Friedman tests for all.
Table 7. Mean ranks obtained by Kruskal–Wallis and Friedman tests for all.
ApproachesFriedman Valuep-ValueKruskal–Wallisp-Value
DE63.740.00065.110.000
DE-H59.310.00060.410.000
DE-S64.010.00065.050.000
DE-TO63.760.00065.350.000
DE-WE63.350.00063.930.000
DE-KN63.330.00064.060.000
Table 8. Comparative results for all BA-based approaches on 16 standard benchmark functions.
Table 8. Comparative results for all BA-based approaches on 16 standard benchmark functions.
BABA−SOBA−HABA−TOBA−WEBA−KN
F#DIM × IterMeanMeanMeanMeanMeanMean
F110 × 10001.59 × 10−071.03 × 10−071.32 × 10−078.95 × 10−080.632020.88186
20 × 20001.02 × 10−848.20 × 10−881.76 × 10−901.30 × 10−904.95 × 10−903.14001 × 10−217
30 × 30001.77 × 10−267.67 × 10−204.13 × 10−321.25 × 10−511.30 × 10−428.91595 × 10−88
F210 × 10004.13 × 10+013.17 × 10+011.82 × 10+013.55 × 10+0137.988340.8852
20 × 20001.06 × 10+025.78 × 10+011.15 × 10+029.47 × 10+01140.2023147.8938
30 × 30002.05 × 10+021.46 × 10+021.83 × 10+021.69 × 10+02271.307275.8626
F310 × 10005.93 × 10−074.70 × 10−073.99 × 10−074.0 × 10−072.91255.2009
20 × 20001.57 × 10−061.05 × 10−061.29 × 10−061.05 × 10−0649.301172.4834
30 × 30003.53 × 10−063.48 × 10−063.27 × 10−062.20 × 10−06197.4826257.9855
F410 × 10002.19 × 10+051.11 × 10+051.66 × 10−142.07 × 10+029.26813.5548
20 × 20002.56 × 10+072.42 × 10+072.31 × 10+071.50 × 10+07160.0394255.3367
30 × 30001.43 × 10+081.38 × 10+083.30 × 10+081.34 × 10+08656.5592946.3934
F510 × 10002.81 × 10−062.67 × 10−061.69 × 10−062.47 × 10−0619.765122.0461
20 × 20007.43 × 10−065.77 × 10−066.25 × 10−065.33 × 10−06250.8679293.7174
30 × 30001.59 × 10−051.43 × 10−051.50 × 10−059.59 × 10−061029.05951277.0077
F610 × 10006.19 × 10−044.54 × 10−043.89 × 10−043.39 × 10−0410.117310.1222
20 × 20007.96 × 10−046.32 × 10−047.77 × 10−045.78 × 10−0420.211920.1467
30 × 30001.05 × 10−031.01 × 10−031.01 × 10−036.65 × 10−0430.362330.2845
F710 × 100015.896618.85118.254415.346516.942918.4835
20 × 2000839.1846686.8456762.1919690.0657876.2518496.7506
30 × 30004892.68644877.70724476.01524482.30355361.48083860.4327
F810 × 10000.0234550.015570.0181520.0167350.0242640.020123
20 × 20000.452220.441010.394780.397190.394450.24522
30 × 30001.72661.09691.35121.37171.43750.99905
F910 × 10004.13943.80033.86874.07393.40243.9516
20 × 20008.5878.80198.56868.5858.83198.5709
30 × 300013.087813.50213.418813.229113.283513.4514
F1010 × 10006.66 × 10−153.31 × 10−151.90 × 10−152.43 × 10−151.33572.0298
20 × 20003.65 × 10−152.03 × 10−151.55 × 10−151.12 × 10−1524.941553.16
30 × 30001.71 × 10−151.06 × 10−151.06 × 10−159.24 × 10−16115.9262318.7949
F1110 × 1000218.1498113.8805105.502758.898769.565652.7079
20 × 200031293.809617609.149320760.649817638.313926832.698513990.4835
30 × 3000651165.7416338621.8857323118.6791268752.5102432441.3838213235.6129
F1210 × 10002.96 × 10+032.21 × 10+032.27 × 10+031.49 × 10+03253.722272.7033
20 × 20002.21 × 10+041.28 × 10+041.49 × 10+045.93 × 10+0311265.96169512.9456
30 × 30002.65 × 10+057.06 × 10+041.65 × 10+057.19 × 10+0471723.277667828.8796
F1310 × 10001.44531.28271.27661.2981.38841.3271
20 × 20002.73032.84042.7462.79732.96882.6632
30 × 30003.99753.99934.25884.05094.26463.9435
F1410 × 10003.83 × 10+067.39 × 10+062.33 × 10+052.69 × 10+044.71 × 10+084.23 × 10+08
20 × 20007.28 × 10+191.55 × 10+171.30 × 10+201.69 × 10+187.78 × 10+192.94 × 10+19
30 × 30005.15 × 10+333.30 × 10+318.00 × 10+321.21 × 10+301.99 × 10+335.50 × 10+33
F1510 × 10002.40 × 10−361.11 × 10−372.08 × 10−372.19 × 10−375.27 × 10−023.76 × 10−02
20 × 20001.50 × 10−374.91 × 10−385.70 × 10−384.72 × 10−391.83 × 10+021.30 × 10+02
30 × 30001.39 × 10−372.20 × 10−387.43 × 10−381.15 × 10−381.08 × 10+042.04 × 10+04
F1610 × 10004.31973.27673.05132.88334.09233.0495
20 × 200021.988124.509322.6082.883323.91722.102
30 × 300089.005380.300474.439166.781267.706163.0376
Table 9. Mean ranks obtained by Kruskal–Wallis and Friedman tests for all.
Table 9. Mean ranks obtained by Kruskal–Wallis and Friedman tests for all.
ApproachesFriedman Valuep-ValueKruskal–Wallisp-Value
BA44.880.00046.150.000
BA-SO44.820.00046.000.000
BA-HA40.290.00040.900.000
BA-TO44.710.00045.160.000
BA-WE40.120.00132.670.005
BA-KN39.530.00032.320.006
Table 10. Characteristics of UCI benchmarks DataSets.
Table 10. Characteristics of UCI benchmarks DataSets.
S. NoData SetContinuousNatureNo. of InputsNo. of Classes
1Diabetes8Real82
2Heart13Real132
3Wine13Real133
4Seed7Real73
5Vertebral6Real62
6Blood Tissue5Real52
7Memo Graphy6Real62
Table 11. Results of 10-fold classification rates of ANN-training methods in 7 datasets for accuracy.
Table 11. Results of 10-fold classification rates of ANN-training methods in 7 datasets for accuracy.
S. NoData SetsTypeBPAPSONNSO-PSONNH-PSONNTO-PSONNWE-PSONNKN-PSONN
Ts. AccTs. AccTs. AccTs. AccTs. AccTs. AccTs. Acc
1Diabetes2-Class65.3%69.1%69.1%71.6%73.3%74.1%78.5%
2Heart2-Class68.3%72.5%67.5%72.5%77.577.5%79%
3Wine3-Class62.17%61.11%66.66%67.44%69.44%69.6%72%
4Seed3-Class70.56%77.77%84.44%77.77%88.88%91.11%93%
5Vertebral2-Class84.95%92.85%92.85%92.85%94.64%94.64%96%
6Blood Tissue2-Class73.47%78.6%78.66%70%82.66%84%87%
7Memo Graphy2-Class71.26%76.66%63%85%88.88%96.66%98%
Table 12. One-way ANOVA results of PSO variants.
Table 12. One-way ANOVA results of PSO variants.
ParameterRelationSum of SquaresdfMean SquareFSignificance
Testing AccuracyAmong groups1318.26219.6972.36760.04639
Table 13. Results of 10-fold classification rates of ANN-training methods in 7 datasets for accuracy.
Table 13. Results of 10-fold classification rates of ANN-training methods in 7 datasets for accuracy.
S. NoData SetsTypeBPADENNSO-DENNH-DENNTO-DENNKN-DENNWE-DENN
Ts. AccTs. AccTs. AccTs. AccTs. AccTs. AccTs. Acc
2Diabetes2-Class65.3%66.1%68.16%69.6%71.30%67.17%75.50%
3Heart2-Class68.3%70.5%72.5%71.5%74.50%72.56%76.34%
4Wine3-Class62.17%64.7%65.19%66.20%66.59%68.25%70.51%
5Seed3-Class70.56%75.16%75.29%75.77%82.13%86.76%91.54%
6Vertebral2-Class84.95%87.13%89.26%91.15%93.64%90.17%96.25%
7Blood Tissue2-Class73.47%76.23%74.16%72..21%84.76%81.34%86.45%
9Memo Graphy2-Class71.26%74.39%68.37%82.45%86.17%96.66%99.21%
Table 14. One-way ANOVA results of DE variants.
Table 14. One-way ANOVA results of DE variants.
ParameterRelationSum of SquaresdfMean SquareFSignificance
Testing AccuracyAmong groups11806196.6722.84530.02043
Table 15. Results of 10-fold classification rates of ANN-training methods in for 7 datasets for accuracy.
Table 15. Results of 10-fold classification rates of ANN-training methods in for 7 datasets for accuracy.
S. NoData SetsTypeBPABANNSO-BANNH-BANNTO-BANNKN-BANNWE-BANN
Ts. AccTs. AccTs. AccTs. AccTs. AccTs. AccTs. Acc
1Diabetes2-Class65.31%66.23%67.40%67.28%69.62%71.39%72.68%
2Heart2-Class68.34%69.39%69.11%69.65%70.12%73.19%72.47%
3Wine3-Class62.17%63.7%63.22%65.53%67.33%69.27%69.08%
4Seed3-Class70.56%72.29%72.97%74.41%77.76%84.53%81.54%
5Vertebral2-Class84.95%86.47%86.39%89.72%90.11%92.38%94.19%
6Blood Tissue2-Class73.47%75.28%75.23%72.21%84.76%81.34%83.19%
7Memo Graphy2-Class71.26%73.17%71.29%74.71%79.23%91.32%94.34%
Table 16. One-way ANOVA results of BA variants.
Table 16. One-way ANOVA results of BA variants.
ParameterRelationSum of SquaresdfMean SquareFSignificance
Testing AccuracyAmong groups845.86140.9672.51130.03623
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ashraf, A.; Pervaiz, S.; Haider Bangyal, W.; Nisar, K.; Ag. Ibrahim, A.A.; Rodrigues, J.j.P.C.; Rawat, D.B. Studying the Impact of Initialization for Population-Based Algorithms with Low-Discrepancy Sequences. Appl. Sci. 2021, 11, 8190. https://doi.org/10.3390/app11178190

AMA Style

Ashraf A, Pervaiz S, Haider Bangyal W, Nisar K, Ag. Ibrahim AA, Rodrigues JjPC, Rawat DB. Studying the Impact of Initialization for Population-Based Algorithms with Low-Discrepancy Sequences. Applied Sciences. 2021; 11(17):8190. https://doi.org/10.3390/app11178190

Chicago/Turabian Style

Ashraf, Adnan, Sobia Pervaiz, Waqas Haider Bangyal, Kashif Nisar, Ag. Asri Ag. Ibrahim, Joel j. P. C. Rodrigues, and Danda B. Rawat. 2021. "Studying the Impact of Initialization for Population-Based Algorithms with Low-Discrepancy Sequences" Applied Sciences 11, no. 17: 8190. https://doi.org/10.3390/app11178190

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop