Next Article in Journal
Dual-Polarized Transparent Antenna and Its Application for Capsule Endoscopy System
Next Article in Special Issue
CAS-UNet: A Retinal Segmentation Method Based on Attention
Previous Article in Journal
Intelligent Recognition of Seismic Station Environmental Interference Based on YOLOv5
Previous Article in Special Issue
A Shallow Pooled Weighted Feature Enhancement Network for Small-Sized Pine Wilt Diseased Tree Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust Adaptive Hierarchical Learning Crow Search Algorithm for Feature Selection

1
School of Computer Science and Engineering, Wuhan Institute of Technology, Wuhan 430073, China
2
Hubei Key Laboratory of Intelligent Robot, Wuhan Institute of Technology, Wuhan 430073, China
3
School of Computer Science, China University of Geosciences, Wuhan 100083, China
4
School of Artificial Intelligence, Shenzhen Polytechnic, Shenzhen 518055, China
5
Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(14), 3123; https://doi.org/10.3390/electronics12143123
Submission received: 7 June 2023 / Revised: 14 July 2023 / Accepted: 17 July 2023 / Published: 18 July 2023
(This article belongs to the Special Issue Application of Machine Learning in Graphics and Images)

Abstract

:
Feature selection is a multi-objective problem, which can eliminate irrelevant and redundant features and improve the accuracy of classification at the same time. Feature selection is a great challenge to balance the conflict between the two goals of selection accuracy and feature selection ratio. The evolutionary algorithm has been proved to be suitable for feature selection. Recently, a new meta-heuristic algorithm named the crow search algorithm has been applied to the problem of feature selection. This algorithm has the advantages of few parameters and achieved good results. However, due to the lack of diversity in late iterations, the algorithm falls into local optimal problems. To solve this problem, we propose the adaptive hierarchical learning crow search algorithm (AHL-CSA). Firstly, an adaptive hierarchical learning technique was used to adaptive divide the crow population into several layers, with each layer learning from the top layer particles and the topmost layer particles learning from each other. This strategy encourages more exploration by lower individuals and more exploitation by higher individuals, thus improving the diversity of the population. In addition, in order to make full use of the search information of each level in the population and reduce the impact of local optimization on the overall search performance of the algorithm, we introduce an information sharing mechanism to help adjust the search direction of the population and improve the convergence accuracy of the algorithm. Finally, different difference operators are used to update the positions of particles at different levels. The diversity of the population is further improved by using different difference operators. The performance of the method was tested on 18 standard UCI datasets and compared with eight other representative algorithms. The comparison of experimental results shows that the proposed algorithm is superior to other competitive algorithms. Furthermore, the Wilcoxon rank-sum test was used to verify the validity of the results.

1. Introduction

Feature selection is a technique used in machine learning and data analysis [1] to select the most relevant and important features from a dataset. It can improve the performance of a machine learning model by reducing dimensionality, removing irrelevant or redundant features, and focusing on the most informative ones. Feature selection methods comprise three categories: the filter method [2], embedded method [3] and wrapper method [4]. The filter method enjoys low computational cost and can avoid overfitting. However, it suffers from neglecting future dependencies, limited adaptability and insensitivity to classification boundaries. The embedded method automatically selects the optimal subset of features for a model, saving time and effort. By integrating feature selection into the model training process, it effectively reduces the risk of overfitting, but it increases the computational intensity during model training. The wrapper approach captures feature interactions and non-linear relationships, tailoring feature selection for specific machine learning algorithms. It is effective with high-dimensional data. Therefore, we adopt the wrapper approach in our proposed algorithm.
A major challenge in feature selection is balancing the trade-off between accuracy and rate. The results demonstrate that using a meta-heuristic algorithm effectively balances this conflict by leveraging meta-heuristic information to narrow down the search space. In recent years, a heuristic search algorithm known as the crow search algorithm, inspired by crows, has been introduced. The CSA is a population-based strategy inspired by the behavior of crows, which involves storing excess food in hidden places for later retrieval. The algorithm uses awareness probability to detect if the crow is being followed and flight length to update the crow’s position iteratively. Prior studies have demonstrated the effectiveness of the CSA algorithm in tackling complex optimization problems involving feature selection.
In group-based algorithms, achieving a balance between exploration and exploitation is a challenging task. In the early stage of algorithm iteration, global exploration becomes particularly crucial. In the late stage of algorithm iteration, the focus shifts to local development, where a refined search is needed to find the optimal solution.
Parameter setting in optimization algorithms is time-consuming and considered a disadvantage. Algorithms with fewer tuning parameters are easier to implement. CSA is a population-based optimization algorithm with adjustable parameters: flight length and awareness probability. Adjusting only these two parameters makes implementation easier. However, the simplistic setting of flight distance parameters limits the algorithm’s exploration and exploitation. The author sets the awareness probability to 0.1, indicating the consideration of local search ability but lack of optimal individual guidance, resulting in slightly insufficient performance. However, mere parameter changes are insufficient for the algorithm to achieve improved results. Therefore, the algorithm undergoes the following improvements:
  • To enhance population diversity, we employed the adaptive hierarchical strategy, utilizing the Hamming stratification method to determine the level number.
  • Multi-strategy variance is a reinforcement method based on differential evolutionary algorithms. It generates new individuals using randomly selected strategies and incorporates the best individuals into the subsequent population. This technique helps populations avoid local optima and discover superior alternatives.
  • The information-sharing mechanism selects parts of top individuals at each level and involves them in particle updates, accelerating convergence rate.

2. Related Works

Feature selection [5] is a challenging NP-hard combinatorial problem, which can be effectively addressed using meta-heuristic algorithms. These algorithms use defined rules and randomness to simulate natural phenomena, thus avoiding the gradient problem with the optimization process [6], which is different from traditional mathematical programming methods [7]. So, some evolutionary algorithms [8] are generally applied to feature selection.
For example, the binary particle swarm optimization algorithm (BPSO) [9] simulates the behavior of birds in a flock, seeking optimal solutions through cooperation and information sharing between individuals in the flock. Mafarja et al. proposed a dragonfly-based optimization algorithm that uses binary to solve the feature selection problem (BDA)  [10]. The genetic algorithm [11,12,13] (GA) is a computational model simulating the biological evolution process of Darwinian biology. It adopts selection, crossover, mutation and other operations to make the algorithm avoid local mimina and increase the diversity. The ant colony algorithm [14] (ACO) is a probabilistic algorithm used to find the optimal path, which is inspired by the behavior of ants looking for the path in the process of finding food. The search process adopts the distributed computing method, and multiple individuals carry out parallel calculations at the same time. The computational power and operation efficiency of the algorithm are greatly improved. The artificial bee colony Algorithm  [15,16] (ABC) is an optimization algorithm to simulate the behavior of bees. It is composed of a simple individual and represents extremely complex behavior. Its advantages are strong global optimization ability and fast convergence speed. The gray wolf-based optimization algorithm [17] (GWO) is inspired by the predatory behavior of the gray wolf population. It has the advantages of relatively high diversity, fast convergence and wide application. Okwu et al. proposed a locust optimization algorithm to solve some problems, which mainly uses the social forces among locusts to divide the space into attractive space, comfortable space and repellent space, and it changes the space scope according to the changes in the distance between different locusts to find the optimal location [18]. The whale optimization algorithm (WOA) [19] is a new group intelligent optimization algorithm simulating whale hunting behavior in nature, but the algorithm has the shortcomings of slow convergence, low accuracy, and easily falling into the local optimal. The crow search algorithm [20] (CSA) mimics the behavior of crows searching for food, but because of the greedy newer method, the algorithm is prone to falling into local optimization.
To address issues with local optimization and lack of diversity in meta-heuristic algorithms, researchers have proposed optimizing these algorithms by modifying their search mechanisms. Adaptive parameters such as non-dominant sorting, positive and corotational functions, transformation function optimization, multi-strategy mechanism, etc. can help achieve the purpose of algorithm optimization.
A particle swarm optimization algorithm with adaptive inertial weights is also used [21,22]. With the progress of iteration, by changing the weights in the particle swarm, the global search and local search capabilities of the algorithm can be better balanced. Mafarja and others have also optimized the dragonfly algorithm. Changing the fly length of the dragonfly so that the flight length varies with iteration is a strategy [23]. In genetic algorithms, non-dominated sorting [24] is used to increase the diversity of the population in the selection process and can well avoid local optimization problems. To solve the problem whereby the crow search algorithm easily fall into local optimal, a binary crow search algorithm based on time flight was proposed [25]. A strategy that can vary the flight distance of each crow as the iteration progresses can be used to optimize problems that fall into local optimality. Another is the chaos crow search algorithm [26,27]. After the use of chaos, instead of a random function, using a positive rotation function to determine the size of the data also plays an optimization role. De Souza et al. proposed a V-shaped transfer function to optimize the crow search algorithm. A good transfer function can better reflect the advantages of the algorithm [28]. Aiming at the problems of the slow convergence and easy convergence of the gray wolf optimization algorithm, a dynamic gray wolf optimizer is proposed. After completing its own search, the optimizer can directly compare with the first wolf without waiting for other wolves, thus improving the iterative convergence speed. Hichem et al. proposed a new way of thinking to solve the feature selection problem of the binary grasshopper optimization algorithm and also improve the performance of the algorithm [29,30]. The whale optimization algorithm has the advantages of simple operation, few adjustment parameters and strong ability to jump out of local optimization. However, some people have proposed optimization ideas for this algorithm, proposing the idea of non-dominant sorting and adaptive parameters  [31], which also applies to the whale optimization algorithm. Wang et al. proposed an adaptive hierarchical learning algorithm [32,33], which can also jump out of the local optimal. Xiaoyu et al. proposed a multi-strategy differential evolution algorithm [34,35,36], which also greatly improves the algorithm. These methods can improve the optimization performance of algorithms. The above are some optimization ideas for individual algorithms as well as adaptive weights [37]. However, millimeter-wave communication remains challenging. Jamil et al. analyzed 5G networks in terms of throughput and energy efficiency, and they proposed OAA [38] to solve the formulaic problem. Nyiam et al. proposed an extension of the multi-objective simplex algorithm (MSA) to generate a set of all non-dominated and non-redundant points  [39]. Wambua et al. proposed a bat algorithm for prioritizing test cases in regression testing [40].

3. Proposed Method

3.1. Crow Search Algorithm

The crow search algorithm is a meta-heuristic algorithm, known by its name, based on the behavior of the crow population. The crow search algorithm is a group-based technology. Its working principle is that crows store surplus food in a hidden place and take it out when they need it. Crows are very greedy. They will track each other and steal each other’s food. However, tracking actions is not an easy task. In fact, crows will stalk another crow to find a new source of food and steal it from there. On the other hand, if the crow realized that another crow was following it, it did not go to the spot where it had hidden its food but instead went to a random spot and tricked the following crow.
The implementation principles of the crow search algorithm follow:
  • Crows live in groups.
  • Crows will remember where they hide their food.
  • Crows will follow other crows to steal food.
  • Crows protect their own food from theft.
First, the population size N and dimension d should be initialized, and two parameters AP and FL should be defined. After calculating the position of each crow, initialize the position in the crow’s memory. In the Algorithm 1, N represents the size of the population, d represents the dimension of the population, and the position of the crow is represented by the vector X = ( x 1 , x 2 , x 3 , , x d ) .
Algorithm 1 Crow search algorithm.
Input: pop
Output: popmem
 1: Initialize the flock of N crows randomly in d dimensional search space;
 2: Define awareness probability AP;
 3: Evaluate the position of each crow;
 4: Initialize the memory of each crow;
 5: while  i t e r a t i o n < m a x _ i t e r a t i o n  do
 6:       for i = 1 to N do
 7:            Randomly select one crow j to follow i
 8:            if  r j A P  then
 9:                  x n e w = x c u r r e n t + r i × f l × ( m e m j , c u r r e n t x i , c u r r e n t )
10:            else
11:                  x n e w = any random position in search space
12:            end if
13:       end for
14:       Check the feasibility of new position;
15:       Evaluate the new position of the crow;
16:       Update the memory of the crow;
17: end while
To describe the specific steps of the crow search algorithm process more visually, we drew a flowchart of the crow search algorithm. The graphical representation of the algorithm makes it more intuitive and easier to understand. Figure 1 contains the diagram chart of the crow search algorithm:

3.2. Binary Crow Search Algorithm

The solution space of the simple crow search algorithm is continuous, but the feature selection in the input machine learning model requires a discrete feature subset. The crow search algorithm cannot directly solve the binary feature selection problem. To apply the crow search algorithm to the feature selection problem, we need to design a binary coding version. Since the feature selection solution will be distributed between binary numbers 0 and 1, there are two types of transform functions: S-type and V-type. Next, we will introduce these two conversion functions.
The formula of the one S-type transfer function is:
T ( x ) = 1 1 + e 2 x
The formula of the one V-type transfer function is:
T ( x ) = x 1 + x 2
There are two types of transfer functions with varying slopes, resulting in different exploration outcomes. A steep transfer function curve leads to poor exploration, while a less steep or flat curve leads to poor exploitation and a higher likelihood of falling into a local minimum [41]. Please refer to the diagram (Figure 2) below, illustrating the two different transfer functions.
This paper aims to apply the crow search algorithm for the feature selection of discrete classes. Therefore, selecting the appropriate binary transfer functions is crucial. The two formulas presented here are employed in the feature selection process using the binary crow search algorithm, and the results obtained with these two transfer functions are compared.

3.3. Adaptive Hierarchical Learning Crow Search Algorithm

The proposed method AHL-CSA is described here. This paper introduces the operation of the system from three aspects: adaptive hierarchical operation, information sharing mechanism and differential operator with multi-policy integration.
The following is a detailed introduction to each step of the algorithm, first to determine the dimension of the dataset. N is the number of initialized crows, and d is the data dimension. The fitness of each crow was then initialized. The initialization is complete. Enter the loop, then check the feasibility of the new location, evaluate the crow’s new location, and update the crow’s memory. Algorithm 2 is the pseudo-code for the algorithm.
The average Hamming value of the population is calculated according to Formulas (3) and (4), and then the number of layers is calculated according to the average Hamming value of the population, using Formulas (5) and (6). Equations (7)–(10) can be used for migration learning in the formula. Particles of different levels will be used according to different difference formulas.
Feature selection is a multi-objective problem. After the feature vectors are obtained by the evolutionary algorithm, these feature vectors are input into the machine learning model, and their performance is evaluated to obtain the classification accuracy and feature ratio. Feature selection is to select important features to improve the classification accuracy and select the least number of features. These two goals are in conflict with each other, so we use the following formula to obtain the fitness value:
F i t n e s s ( x ) = δ × γ R ( x ) + β × x n
In the formula, Fitness(x) is the fitness value of the x vector, γ R ( x ) is the classification error rate obtained from the feature vector x, x n is the feature selection ratio of x, and δ and β represents fixed values that satisfy δ + β = 1 , β = 0.01 .
Algorithm 2 Hierarchical learning crow search algorithm.
 1: Determine the number of feature in the dataset, call it d;
 2: Initialize the flock of N crows randomly in d-dimensional binary vectors;
 3: Evaluate the position of each crow;
 4: Initialize the memory of each crow;
 5: while  i t e r a t i o n < m a x _ i t e r a t i o n  do
 6:  Calculate the average Hamming value of the population;
 7:  Calculate the number of layering;
 8:  Sort the population according to fitness and find out the position of the best value and the position of the worst value;
 9:  Calculate the average position of the population at this time;
10:  Through the difference formula, the different offsets required for different particles are l e a r n i , c u r r e n t ;
11:  for i = 1 to N do
12:     Randomly select one crow j to follow i
13:     Define awareness probability AP
14:     if  r j A P then
15:       x n e w = x c u r r e n t + r i × f l × ( m e m j , c u r r e n t x i , c u r r e n t ) + l e a r n i , c u r r e n t
16:   else
17:       x n e w = any random position in search space
18:   end if
19: > end for
20:  Check the feasibility of the new position;
21:  Evaluate the new position of the crow;
22:  Update the memory of the crow;
23: end while

3.3.1. Adaptive Hierarchical Operation

The BCSA-TVFL proposed by Abhilasha et al. changes the flight length to an adaptive value, which means that the flight length (FL) will gradually decrease linearly with the iterations of the algorithm. Although the flight length will change with iterations and is consistent with the habitat of crows in nature, this simple optimization parameter is still flawed.
Therefore, we use a hierarchical strategy to optimize this problem. Since the population has different levels, the particles in the population can maintain a good diversity. High-order particles focus on the exploitation of the solution space, and low-order particles focus on the exploration of the solution space, which is balanced by the self-organization of the population. The hierarchy is implemented as follows:
  • Arrange population individuals in ascending order based on their fitness values after calculating the fitness for each particle;
  • The ranked population is divided into NL layers, with higher ranked particles assigned to higher levels and lower ranked particles assigned to lower levels. The number of particles in each stratum is the same, except when the number of population cannot be evenly divided by the number of strata. In this case, the last NP%NL particles are placed in the lowest stratum;
  • Each individual learns from the random individual in the higher level, and the highest level individual updates its position by learning from the random individual in the same level.
The number of individuals that it learns varies across different levels. Higher-level individuals learn fewer individuals compared to lower-level ones. This approach facilitates the exploration of low-level individuals and the exploitation of high-level individuals. Additionally, the performance of feature selection is influenced by the number of NL layers.
Yang et al. proposed a dynamic version of the method to determine the number of layers [42]. First, initialize an array with different numbers that represent the number of layers of the population. Then, at each iteration, select the number of strata for this iteration based on the probability of each number and record the performance using this number of strata at the end of this iteration. The more the fitness value of the current optimal solution in the array improves, the more likely it is to choose this integer in the next iteration. One major difficulty in determining the number of layers by this method would be the need to initially give a set of candidate integers. It is very time and effort consuming to give the set of layers, so an adaptive method is proposed to determine the number of layers, and the calculation formula for obtaining the number of layers (NL) is as follows:
  • First, use the following formula to calculate the Hamming value of pairwise in the population. Because the Hamming value can reflect the distance and similarity between individuals:
    H i , j = d = 1 D x i , d x j , d
    In the formula, H i , j is the Hamming distance between x i and x j , x i and x j are two different individuals in the population, and D is the dimension of the population.
  • Then, calculate the average Hamming value of the whole population:
    A v e t = m e a n i = 1 i = N 1 j = i + 1 N H i , j
    A v e t is used to find the average Hamming value of the whole population, and N is the population size.
  • In the initial iteration, we will divide the whole population into more levels so that the population can avoid falling into local optimum at an early stage. At the end of the iteration, fewer levels will be obtained so that the population can conduct a more refined search in a small space. Improving the learning probability of dominant individuals is conducive to a more intensive use of the search space by groups. For this purpose, the rate of decline of the average Hamming value in the population is used to determine NL. There is the following formula:
    r a t e = A v e t A v e 0 × N
    N L = 4 ,   r a t e < 4 f i x ( r a t e ) ,   4 r a t e N [ N / 2 ] ,   o t h e r
    In the formula, rate is the decline rate of the average Hamming value in the population, N is the size of the population, and the fix(rate) function is the maximum integer that returns, which is no greater than the rate.

3.3.2. Information Sharing Mechanism

In the process of searching, the search information of the population cannot be updated in time, which affects the convergence performance of the algorithm. Therefore, we propose an information sharing mechanism in which population particles cooperate to search under the guidance of their neighbors within a search cycle. Once a particle falls into the local optimal, the information sharing mechanism will adjust the search direction in time, avoid local mimina and search toward the population optimal direction. This paper divides the whole population into several levels. The information sharing mechanism uses the average location of the optimal solution from each level in the population. On the one hand, individuals in the population use optimal strategies to find better solutions. On the other hand, it also obtains information from other individuals, with evolutionary information at different levels. Information from other levels might push stagnant particles back to life.
Since the simple difference operator cannot solve the diversity problem of the population well, an information sharing formula is added to the difference operator formula to solve the diversity problem and local optimal problem of the simple difference operator by combining the difference operator and information sharing. The optimal average position is calculated as:
x ¯ = 1 p k = 1 p x l b e s t k
In the formula, p is the optimal number of selected individuals in each layer, x ¯ is the optimal average position, and x l b e s t k is the optimal particle in each layer.
The average location guidance mechanism of the information sharing mechanism is illustrated in Figure 3. The P point represents the guide point individuals selected in the population search space, while the other points (Xi) represent the individuals selected from the optimal part in each layer. These vectors provide the general direction for the next population update and enable escaping local optima.

3.3.3. Differential Operator with Multi-Strategy Integration

The difference algorithm [43] is a meta-heuristic random search algorithm based on group differences. The differential evolution [44] algorithm has the advantages of a simple principle, few controlled parameters and strong robustness. Different from the general evolutionary algorithm [45], the difference algorithm uses mutation operators and crossover operators to generate new particles [46]. The updating process of DE is realized by three steps: mutation, crossover and selection.
In the iteration of the algorithm, each individual particle will evolve to different degrees [47]. Due to the different characteristics of each particle, the learning strategies of each layer of particles are also different. According to the characteristics of different populations, different mutation strategies [48] and control parameters are used. At the same time, the mutation strategy [49,50] of individuals is determined by probability judgment for the common population to accelerate the convergence rate. Particles with higher fitness are more likely to approach the optimal location and exhibit superior accuracy in search ability. They play a pivotal role in guiding the population toward the optimal direction. Conversely, particles with lower fitness are more likely to deviate from the optimal position. However, they possess greater exploration capability, aiding in space expansion and avoiding local minima. So, there are three difference formulas to update each particle in the population:
  • The update operator performs a more refined search in a smaller area:
    L i , j = F × x g b e s t x i , j + F × x r 1 x r 2 + σ × x ¯ x i , j
  • In order to find the particles in the middle of the population carefully and extensively, we will give consideration to both:
    L i , j = F × x r 1 x i , j + F × x r 1 x r 2 + σ × x ¯ x i , j
  • It facilitates the population in avoiding local optima by enabling extensive searches for particles positioned in the middle of the population:
    L i , j = F × x g b e s t x w o r s t + F × x g b e s t x r + σ × x ¯ x i , j
In the above formula, F is a random number in a [−1,1] interval, and L i , j is the location offset l e a r n i , c u r r e n t mentioned above. x g b e s t is the global optimal value in the population, and x w o r s t is the global worst in the population. x i , j is the current particle, while x r 1 and x r 2 are two different random particles in the interval with better fitness than the current particle. σ is a constant value of 0.04, and x ¯ is the average value of the current population.

4. Experiments and Results

4.1. Experimental Design and Parameters

This experiment is conducted on a WIN10 system, with an NVIDIA GTX 1660 graphics card, Inter Core i5-11400 processor, 2.6 GHz main frequency and 16 GB running memory. The 2020b version of Matlab is used.
In order to verify the correctness of the idea of feature selection based on the hierarchical learning crow algorithm, the dataset information, comparison algorithm information and experimental parameter settings for the experiment are given. The experimental results of classification accuracy and dimension reduction of this algorithm and the comparison algorithm are given. The following datasets are used (Table 1):
To create a meta-heuristic algorithm, the parameters of a specific algorithm need to be adjusted. Different algorithms have different algorithm parameters. These parameters are obtained through a large number of experiments, and the setting of these parameters is a complex and time-consuming task. There are parameters to the following algorithms: binary particle swarm optimization (BPSO) (M. Mafarja, Jarrar, et al., 2018 [9]), genetic algorithm (GA) (Sikora & Piramuthu, 2007 [11,12,13]), binary dragonfly optimization algorithm (BDA) (M. M. Mafarja et al., 2017 [10]), binary crow search algorithm (BCSA1) (Laabadi, Naimi, Amri, & Achchab, 2020 [20]), V-shaped binary crow search algorithm (BCSA2) (De Souza et al., 2018 [26]), binary gray wolf optimization (BGWO) (Emary, Zawbaa, & Hassanien, 2016 [17]), binary grasshopper optimization algorithm (BGOA) (M. Mafarja et al., 2019 [29]), and binary crow search algorithm with time-varying flight length (BCSA-TVFL) (Abhilasha et al., 2020 [25]).
In these algorithms, certain parameters are derived from extensive experiments conducted by other researchers. These parameters remain constant throughout the iterations the algorithm. In BPSO, the authors use the inertia weight, particle renewal rate, global learning factor, individual learning factor and other parameters. The BCSA uses two parameters: flight length and perceived probability. The BDA uses separation weight, alignment weight, cohesion weight, food factor, enemy factor and other parameters. The GA uses crossover probability, mutation probability and other parameters. BCSA-TVFL has the same parameters as the BCSA, with an additional interval [1.5, 2.5]. A value is set for each parameter during initialization, which remains constant through all iterations. The values of these parameters are empirically determined, so setting these parameters is a complex and time-consuming task. Therefore, we can say that the fewer the number of parameters, the better. The following are the parameters used in the above article (Table 2).

4.2. Experimental Results and Analysis

Abhilasha et al. compared eight transfer functions in terms of average fitness [25], average accuracy and feature selection ratio. Through a lot of experiments, it can be concluded that the best transfer function is S1 type transfer function, which is Formula (1). Therefore, we use this transfer function in the following experiments.
To explore the effect of the proposed optimization strategy on the performance of the crow search algorithm, we compare the results of the AHL-CSA algorithm with other algorithms. The mentioned algorithm and other representative algorithms use 30 independent replicate experiments to guarantee the validity of the results. The dataset was divided into a 20% test set and an 80% training set. kNN classifiers are stable, simple, and have few parameters. To validate the classification of feature subsets, we used k = 5 as the nearest neighbors of the kNN classifier. The number of populations N in each experiment was set to 50. There were three evaluation criteria. The average fitness and average accuracy, standard deviation, feature selection rate, and maximum accuracy for 30 independent runs were obtained under the same experimental conditions. In the table, the best results are shown in bold, and the number of optimal results obtained for each algorithm is counted.

4.2.1. Results Comparison

The AHL-CSA proposed in this paper is compared with other algorithms. This is a comparison of three numbers. Table 3 records accuracy, Table 4 records fitness value, and Table 5 records feature ratio. These three numbers are the average of 30 separate operations. At the same time, the deviation value obtained by each algorithm is recorded to reflect the stability of each algorithm. After recording the data obtained in each experiment, the superiority of the algorithm is proved from three aspects: the number of times that the algorithm ranks first, the number of times that the algorithm ranks first with other algorithms, and the number of times that the algorithm ranks last. Finally, each algorithm has a maximum precision to reflect the superiority of the algorithm.
The transfer function adopts the above Equation (1), as used by Abhilasha et al. According to the average accuracy measurement, the transfer function S1 ranks first, with an average fitness and standard deviation of 30 independent runs (Table 6). The transfer function S1 has the best result 70% of the time [25]. The results of Mafarja et al. show the importance of the selection of transfer function in the algorithm, because selecting the appropriate transfer function can significantly improve the performance of the algorithm. Furthermore, it is clear that the behavior of adjusting the transfer function through the optimization process has a big impact on improving the performance of the algorithm [23]. To better evaluate the proposed method, the experimental results of a comparison between AHL-CSA and other algorithms are calculated on the same computer. The standard deviation of each datum effectively reflects the quality of the data processed by the algorithm. The higher the standard deviation, the more unstable the algorithm processing results, and the less convincing the results. The lower the standard deviation, the better the stability of the algorithm for data processing, and the more convincing the data results.
The faster the convergence curve reaches stability, the better the data processing effect of the algorithm and vice versa. The convergence curve shows the exploration and development ability of the algorithm, and whether the two methods can be compromised in the process of evolution is also an evaluation criterion of the algorithm.
To ensure that the experimental comparison is fair and equitable, the ending criteria should be consistent with the original text. In each algorithm, the population size is 50, the number of iterations is 100, and the results of each algorithm are generated from 30 independent runs.
The average accuracy of Table 3 shows that the proposed algorithm achieves very good results. The proposed algorithm achieves the best results on nine datasets, and there are five datasets where it achieves the same results as the other algorithms. The standard deviation of the proposed algorithm for each dataset is almost all 0, which indicates that the proposed algorithm has good stability. In Table 4, for the mean accuracy, the proposed algorithm outperforms other representative algorithms on 10 datasets and achieves the same results as other algorithms on one dataset. In Table 5, for the feature selection rate, the proposed algorithm achieves the best results on five datasets and achieves the same results as the compared algorithms on one dataset.
The adaptive hierarchical strategy, differential operator with multi-strategy integration and information sharing mechanism proposed in this paper can effectively improve the performance of the algorithm. Because the proposed three innovations can effectively improve the diversity of the population, each particle has the potential to achieve the optimal.

4.2.2. Results of Convergence Comparison

The following Figure 4 is the accuracy convergence diagram of each algorithm. From the following 18 datasets, we can see that the AHL-CSA algorithm converges faster than other algorithms, and the final accuracy value is also better. In each algorithm, the population size is 50 and the number of iterations is 100 so as to run the convergence diagram of the accuracy algorithm.
As can be seen from the convergence rate table above, the algorithm can converge quickly in most datasets and achieve good results in algorithm accuracy. This also shows that the algorithm has a strong ability to avoid the local mimina problem.

4.2.3. Wilcoxon Rank-Sum Test

The rank-sum test (Table 7) is used to infer whether there is a difference between two population distribution locations from two independent measurements or samples of rank data. This is a non-parametric paired two-sample t test that can be used to compare two independent sample groups. It is used when the data are not normally distributed and still do not meet the normality requirement after some numerical conversion attempts. It compares the non-normal sample values of two independent sample groups.
To test the validity of the results, a Wilcoxon rank-sum test was performed on the data [51]. The p-values were tested at the significance level of 0.05 by the accuracy of 30 independent runs for each dataset. p-values greater than 0.05 are indicated in bold. From this table, we can see that the data are statistically valid.

4.2.4. Results Comparison with Well-Known Filter Methods in Literature

Filter methods have gained popularity in feature selection as they require low computational cost and show high processing speed. These methods reduce the dimension of feature space by evaluating and ranking features, thereby finding the optimal predictive subset. The following table compares the representative filter methods and our proposed algorithm in terms of the maximum accuracy.
Filter methods often ignore the interrelationships between features. They typically select features by measuring a single feature, such as variance, correlation coefficient, Chi-square test, and information gain. These indicators solely consider the predictive power of each feature itself, however, neglecting the relationship between features. There may be complex interactions between features that can have an important impact on the accuracy of classification or regression models.
Another type of approach is to use a wrapper or embedded method. These methods directly consider feature selection as part of model training, take into account the interrelationships between features, and optimize through techniques such as cross-validation. However, these methods are more computationally intensive and time consuming than filter methods. Therefore, in practical applications, we need to weigh the relationship between computational cost and accuracy and choose the appropriate method.
Table 8 shows that the proposed algorithm takes the lead compared with the filter method. The hierarchical strategy makes the number of candidate samples of individuals differ at every level, thus improving the population diversity. The introduction of a difference strategy makes the particles with poor fitness mutate and cross, increasing the diversity of particles, boosting the possibility of obtaining solutions in space, and thus obtaining stronger global search ability. Furthermore, the combination of a hierarchical strategy and differential strategy takes our algorithm to a better level, achieving promising results.

5. Conclusions and Future Work

In this paper, a meta-heuristic search strategy is used in feature selection. Problems such as local optimality and population diversity cannot be guaranteed and may arise when using this method. Based on a meta-heuristic crow search algorithm, we propose an adaptive hierarchical learning crow search algorithm (AHL-CSA). To improve the population diversity, we adopted an adaptive stratification strategy. This strategy divides the population into several levels according to fitness level. In addition, we also use multi-strategy differential and information sharing strategies to further avoid the local optimal problem. Multi-strategy difference is an improved method based on the differential evolution algorithm, which uses multiple randomly selected strategy vectors to generate new individuals and selects the best individuals from them to join the next generation population. This method avoids the current local optimum and enables the discovery of a superior alternative. To verify the validity and accuracy of these methods, 18 UCI standard datasets were selected for fitness, classification accuracy and feature selection ratio tests. The comparison table and graph clearly demonstrate the algorithm’s strong convergence speed and effectiveness. The proposed method can effectively balance exploration and exploitation in the process of evolution. The proposed algorithm outperforms other representative algorithms in terms of accuracy, selection ratio, and fitness value. Finally, we have uploaded the algorithm code to GitHub: https://github.com/eagdog/Algorithm-code.git (accessed on 21 June 2023).
Although the proposed algorithm utilizes multiple strategies to calculate various levels, it incurs a slight time disadvantage. Furthermore, the algorithm is limited to text datasets on UCI and cannot process other data types. In the future work, we will try this algorithm for feature selection in high-dimensional data so that this algorithm can better play to its strengths. We will also focus on applying the crow search algorithm to a wider range of applications. The algorithm has applications not only in feature selection but also in remote sensing image classification, medical image classification, and various fused object detection methods. Our future research direction aims to fulfill the diverse needs of different application areas.

Author Contributions

Conceptualization, Z.Y.; Methodology, Z.Y.; Software, Z.Y.; Validation, Z.Y.; Investigation, Z.Y.; Writing—original draft, Z.Y. and B.G.; Writing—review and editing, Y.C., Y.W., X.Y. and X.L.; Supervision, Y.C. and X.L.; Project administration, Y.C. and X.L.; Funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by The National Key Research and Development Program of China under Grant 2020YFB1313900, Innovation Fund of Hubei Key Laboratory of Intelligent Robotics under Grant HBIRL202107, Wuhan Institute of Technology Scientific Research Fund under Grant 21QD53, Science and Technology Research Project of Hubei Provincial Education Department under Grant Q20221501, Shenzhen Science and Technology Program under Grant JCYJ20200109115201707 and JCYJ20220818101408019, Stable Supporting Program for Universities of Shenzhen under Grant 20220812102547001, National Natural Science Foundation of China under Grant 62102268, Shenzhen Polytechnic Research Fund under Grant 6022312044K, Shenzhen Polytechnic Research Fund under Grant 6023310030K.

Data Availability Statement

The dataset used in the experiments is publicly available from UCI https://archive.ics.uci.edu/ml/datasets.php (accessed on 17 April 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CSAcrow search algorithm
AHLCSAadaptive hierarchical learning crow search algorithm
PSOparticle swarm optimization algorithm
GAgenetic algorithm
BDAdragonfly optimization algorithm
GOAgrasshopper optimization algorithm
GWOgray wolf optimization algorithm
CSATVFLcrow search algorithm with time-varying flight length

References

  1. Zhang, L.; Li, N.; Li, Z. An Overview on Supervised Semi-structured Data Classification. In Proceedings of the 2021 IEEE 8th International Conference on 526 Data Science and Advanced Analytics (DSAA), Porto, Portugal, 6–9 October 2021; pp. 1–10. [Google Scholar]
  2. Nie, P.Y.; Ma, C.F. A trust region filter method for general non-linear programming. Appl. Math. Comput. 2006, 172, 1000–1017. [Google Scholar] [CrossRef]
  3. Kim, P.; Kim, J.; Jung, W.K.; Bu, S. An error embedded method based on generalized Chebyshev polynomials. J. Comput. Phys. 2016, 306, 55–72. [Google Scholar] [CrossRef]
  4. Maldonado, S.; Weber, R. A wrapper method for feature selection using Support Vector Machines. Inf. Sci. 2009, 179, 2208–2217. [Google Scholar] [CrossRef]
  5. Nahid, A.A.; Sikder, N.; Abid, M.H.; Toma, R.N.; Talin, I.A.; Ali, L.E. Home Occupancy Classification Using Machine Learning Techniques along with Feature Selection. Int. J. Eng. Manuf. 2022, 12, 38. [Google Scholar] [CrossRef]
  6. Oh, M.; Yeak, S. A Hybrid Multiscale Finite Cloud Method and Finite Volume Method in Solving High Gradient Problem. Int. J. Comput. Methods 2022, 19, 271–279. [Google Scholar] [CrossRef]
  7. Li, Y.; Han, T.; Han, B.; Zhao, H.; Wei, Z. Whale Optimization Algorithm with Chaos Strategy and Weight Factor. J. Phys. Conf. Ser. 2019, 1213, 032004. [Google Scholar] [CrossRef] [Green Version]
  8. Ming, F.; Gong, W.; Wang, L. A Two-Stage Evolutionary Algorithm wth Balanced Convergence and Diversity for Many-Objective Optimization. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 6222–6234. [Google Scholar] [CrossRef]
  9. Tawhid, M.A.; Dsouza, K.B. Hybrid Binary Bat Enhanced Particle Swarm Optimization Algorithm for solving feature selection problems. Appl. Comput. Inform. 2018, 16, 117–136. [Google Scholar] [CrossRef]
  10. Hammouri, A.I.; Mafarja, M.; Al-Betar, M.A.; Awadallah, M.A.; Abu-Doush, I. An improved Dragonfly Algorithm for feature selection. Knowl. Based Syst. 2020, 203, 106131. [Google Scholar]
  11. Kabir, M.M.; Shahjahan, M.; Murase, K. A new local search based hybrid genetic algorithm for feature selection. Neurocomputing 2011, 74, 2914–2928. [Google Scholar] [CrossRef]
  12. Weidong, L.I.; Suhayb, M.K.; Thangavelu, L.; Marhoon, H.A.; Pustokhina, I.; Alqsair, U.F.; El-Shafay, A.S.; Alashwal, M. Implementation of AdaBoost and genetic algorithm machine learning models in prediction of adsorption capacity of nanocomposite materials. J. Mol. Liq. 2022, 350, 118527. [Google Scholar]
  13. Hiremath, G.S.; Kumar, N.; CL, S.N. A Study on Ancient Temple Structural Elements Recognition Using Genetic Algorithm. Int. J. Eng. Manuf. 2023, 3, 34–47. [Google Scholar] [CrossRef]
  14. Shunmugapriya, P.; Kanmani, S. A hybrid algorithm using ant and bee colony optimization for feature selection and classification (AC-ABC Hybrid). Swarm Evol. Comput. 2017, 36, 27–36. [Google Scholar] [CrossRef]
  15. Karaboga, D.; Basturk, B. On the performance of artificial bee colony (ABC) algorithm. Appl. Soft Comput. 2008, 8, 687–697. [Google Scholar] [CrossRef]
  16. Zhang, M.; Palade, V.; Wang, Y.; Ji, Z. Attention-based word embeddings using Artificial Bee Colony algorithm for aspect-level sentiment classification. Inf. Sci. 2021, 545, 713–738. [Google Scholar] [CrossRef]
  17. Al-Tashi, Q.; Abdulkadir, S.J.; Rais, H.M.; Mirjalili, S.; Alhussian, H. Binary Optimization Using Hybrid Grey Wolf Optimization for Feature Selection. IEEE Access 2019, 7, 39496–39508. [Google Scholar] [CrossRef]
  18. Mafarja, M.; Aljarah, I.; Faris, H.; Hammouri, A.I.; Ala’M, A.Z.; Mirjalili, S. Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst. Appl. 2019, 117, 267–286. [Google Scholar] [CrossRef]
  19. Mafarja, M.M.; Mirjalili, S. Whale Optimization Approaches for Wrapper Feature Selection. Appl. Soft Comput. 2017, 62, 441–453. [Google Scholar] [CrossRef]
  20. Oliva, D.; Hinojosa, S.; Cuevas, E.; Pajares, G.; Avalos, O.; Galvez, J. Cross entropy based thresholding for magnetic resonance brain images using Crow Search Algorithm. Expert Syst. Appl. 2017, 79, 164–180. [Google Scholar] [CrossRef]
  21. Hu, J. A Particle Swarm Optimization Algorithm with Distributed Adaptively Weighted Delays. Adv. Appl. Math. 2021, 10, 753–762. [Google Scholar] [CrossRef]
  22. Mafarja, M.; Jarrar, R.; Ahmad, S.; Abusnaina, A.A. Feature Selection Using Binary Particle Swarm Optimization with Time Varying Inertia Weight Strategies. In Proceedings of the International Conference on Future Networks and Distributed Systems, Amman, Jordan, 26–27 June 2018; pp. 1–9. [Google Scholar]
  23. Mafarja, M.; Aljarah, I.; Heidari, A.A.; Faris, H.; Fournier-Viger, P.; Li, X.; Mirjalili, S. Binary Dragonfly Optimization for Feature Selection using Time-Varying Transfer functions. Knowl. Based Syst. 2018, 161, 185–204. [Google Scholar] [CrossRef]
  24. Bs, A.; Fg, B. A new approach to generate pattern-efficient sets of non-dominated vectors for multi-objective optimization—ScienceDirect. Inf. Sci. 2020, 530, 22–42. [Google Scholar]
  25. Chaudhuri, A.; Sahu, T.P. Feature selection using Binary Crow Search Algorithm with time varying flight length. Expert Syst. Appl. 2020, 168, 114288. [Google Scholar] [CrossRef]
  26. Rizk-Allah, R.M.; Hassanien, A.E.; Bhattacharyya, S. Chaotic Crow Search Algorithm for Fractional Optimization Problems. Appl. Soft Comput. 2018, 71, 1161–1175. [Google Scholar] [CrossRef]
  27. Chung, S.W.; Freund, J.B. An optimization method for chaotic turbulent flow. J. Comput. Phys. 2022, 457, 111077. [Google Scholar] [CrossRef]
  28. Jafari-Asl, J.; Azizyan, G.; Monfared, S.; Rashki, M.; Andrade-Campos, A.G. An enhanced binary dragonfly algorithm based on a V-shaped transfer function for optimization of pump scheduling program in water supply systems (case study of Iran). Eng. Fail. Anal. 2021, 123, 105323. [Google Scholar] [CrossRef]
  29. Algamal, Z.Y.; Qasim, M.K.; Lee, M.H.; Ali, H. Improving grasshopper optimization algorithm for hyperparameters estimation and feature selection in support vector regression. Chemom. Intell. Lab. Syst. 2021, 208, 104196. [Google Scholar] [CrossRef]
  30. Mafarja, M.; Aljarah, I.; Heidari, A.A.; Hammouri, A.I.; Faris, H.; Al-Zoubi, A.; Mirjalili, S. Evolutionary Population Dynamics and Grasshopper Optimization Approaches for Feature Selection Problems. Knowl. Based Syst. 2017, 145, 25–45. [Google Scholar] [CrossRef] [Green Version]
  31. Zhang, W.; Zhang, S.; Wu, F.; Wang, Y. Path Planning of UAV based on Improved Adaptive Grey Wolf Optimization Algorithm. IEEE Access 2021, 9, 89400–89411. [Google Scholar] [CrossRef]
  32. Wang, J.; Zhang, Y.; Hong, M.; He, H.; Huang, S. A self-adaptive level-based learning artificial bee colony algorithm for feature selection on high-dimensional classification. Soft Comput. 2022, 26, 9665–9687. [Google Scholar] [CrossRef]
  33. Jin, H.C.; Lee, J.; Park, C. Deep-space trajectory optimizations using differential evolution with self-learning. Acta Astronaut. 2022, 191, 258–269. [Google Scholar]
  34. Li, X.; Wang, L.; Jiang, Q.; Li, N. Differential evolution algorithm with multi-population cooperation and multi-strategy integration. Neurocomputing 2021, 421, 285–302. [Google Scholar] [CrossRef]
  35. Pan, J.S.; Liu, N.; Chu, S.C. A competitive mechanism based multi-objective differential evolution algorithm and its application in feature selection. Knowl.-Based Syst. 2022, 245, 108582. [Google Scholar]
  36. Gao, D.; Wang, G.G.; Pedrycz, W. Solving Fuzzy Job-Shop Scheduling Problem Using DE Algorithm Improved by a Selection Mechanism. IEEE Trans. Fuzzy Syst. 2020, 28, 3265–3275. [Google Scholar] [CrossRef]
  37. Li, H.; He, F.; Pan, Y. Multi-objective dynamic distribution adaptation with instance reweighting for transfer feature learning. Knowl. Based Syst. 2023, 263, 110303. [Google Scholar] [CrossRef]
  38. Jamil, S.; Rahman, M.; Tanveer, J.; Haider, A. Energy Efficiency and Throughput Maximization Using Millimeter Waves and ndash;Microwaves HetNets. Electronics 2022, 11, 474. [Google Scholar] [CrossRef]
  39. Nyiam, P.B.; Salhi, A. A Comparison of Benson’s Outer Approximation Algorithm with an Extended Version of Multiobjective Simplex Algorithm. Adv. Oper. Res. 2021, 2021, 1857030. [Google Scholar] [CrossRef]
  40. Wambua, A.W.; Wambugu, G.M. A Comparative Analysis of Bat and Genetic Algorithms for Test Case Prioritization in Regression Testing. Int. J. Intell. Syst. Appl. 2023, 14, 13. [Google Scholar] [CrossRef]
  41. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary Particle Swarm Optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar] [CrossRef]
  42. Qiang, Y.; Chen, W.N.; Deng, J.D.; Yun, L.; Zhang, J. A Level-Based Learning Swarm Optimizer for Large-Scale Optimization. IEEE Trans. Evol. Comput. 2018, 22, 578–594. [Google Scholar]
  43. Wang, Y.J.; Zhang, J.S. Global optimization by an improved differential evolutionary algorithm. Appl. Math. Comput. 2007, 188, 669–680. [Google Scholar]
  44. Li, C.; Deng, L.; Qiao, L.; Zhang, L. An efficient differential evolution algorithm based on orthogonal learning and elites local search mechanisms for numerical optimization. Knowl. Based Syst. 2022, 235, 107636. [Google Scholar] [CrossRef]
  45. Bo, L.; Ding, L.; Li, J. Differential evolution-based parameters optimisation and feature selection for support vector machine. Int. J. Comput. Sci. Eng. 2016, 13, 355. [Google Scholar]
  46. Zhan, Z.H.; Wang, Z.J.; Jin, H.; Zhang, J. Adaptive Distributed Differential Evolution. IEEE Trans. Cybern. 2019, 77, 1–15. [Google Scholar] [CrossRef]
  47. Deng, J.; Hu, P.; Bai, Z.J.; Wang, C.P.; Kang, F.R.; Liu, L. Dynamic behaviours on oxidation heat release of key active groups for coal with different degrees of metamorphism. Fuel 2022, 320, 123967. [Google Scholar] [CrossRef]
  48. Zl, A.; Qz, B. Variable metric evolution strategies by mutation matrix adaptation - ScienceDirect. Inf. Sci. 2020, 541, 136–151. [Google Scholar]
  49. Zt, A.; Kl, A.; Yi, W.B. Differential evolution with adaptive mutation strategy based on fitness landscape analysis. Inf. Sci. 2021, 549, 142–163. [Google Scholar]
  50. Rauf, H.T.; Bangyal, W.; Pervaiz, S.; Ahmad, J. An Improved Particle Swarm Optimization Algorithm with Chi-Square Mutation Strategy. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 481–491. [Google Scholar]
  51. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. Flow diagram of the crow search algorithm.
Figure 1. Flow diagram of the crow search algorithm.
Electronics 12 03123 g001
Figure 2. Trend graph of s-type and v-type transform functions.
Figure 2. Trend graph of s-type and v-type transform functions.
Electronics 12 03123 g002
Figure 3. Diagram of information sharing mechanism.
Figure 3. Diagram of information sharing mechanism.
Electronics 12 03123 g003
Figure 4. Convergence curves of AHL-CSA and other representative wrapper methods on 18 datasets.
Figure 4. Convergence curves of AHL-CSA and other representative wrapper methods on 18 datasets.
Electronics 12 03123 g004aElectronics 12 03123 g004b
Table 1. Summary of Dataset.
Table 1. Summary of Dataset.
DatasetFeaturesClassesSamplesDatasetFeaturesClassesSamples
haberman32306australia142690
ecoli86336zoo167101
balancescale43625vehicle184846
iris43150lymphography182148
glass97214spectheart222267
contraceptive931473breastEW302569
tictactoe92958inosphere332351
breastcancer102683dermatology346366
wine133178sonar602208
Table 2. Parameter setting for AHL-CSA and other comparative algorithms.
Table 2. Parameter setting for AHL-CSA and other comparative algorithms.
AlgorithmParameterValues
BPSOInertia of Particle0.9
c10.5
c20.5
BCSAFlight Length fl2
Awareness Probability AP0.1
BDASeparation s0.1
Alignment a0.1
Cohesion c0.7
Attraction f1
Distraction e1
GACrossover Probability0.8
Mutation Probability0.02
BCSA-TVFLAwareness Probability AP0.1
Flight Length fl[1.5, 2.5]
AHL-CSAAP0.1
F[−1, 1]
σ 0.04
fl[1.5, 2.5]
Table 3. Average accuracy comparison of AHL-CSA with competitors.
Table 3. Average accuracy comparison of AHL-CSA with competitors.
Dataset BCSA1BCSA2BGWOBGOABPSOBDAGABCSA-TVFLAHL-CSA
habermanAvg0.83610.81970.83610.75410.71530.77050.78690.83610.8420
STD0.00560.08100.05720.01290.07240.00000.00000.00000.0000
ecoliAvg0.89060.85070.87910.86520.89320.89450.85070.92550.8955
STD0.00670.00000.00600.00270.10710.00000.00000.00570.0000
balancescaleAvg0.92000.86400.82400.81600.62400.83800.77600.84000.8240
STD0.00400.00320.00000.00360.15070.34500.00000.00000.0000
irisAvg1.00001.00001.00001.00000.87001.00001.00001.00001.0000
STD0.00000.00380.00610.00040.20980.00000.00000.00000.0000
glassAvg0.80790.76590.80560.71280.77980.76860.77600.77210.8095
STD0.01081.04600.01260.02131.32871.29101.43691.54500.0000
contraceptiveAvg0.54750.53060.53610.55250.44260.56120.55440.51800.5442
STD0.00740.00420.00330.03690.20850.29400.00920.00860.0000
tictactoeAvg0.83000.85340.81620.81590.79900.84690.78870.78130.8562
STD0.00000.00000.03000.00480.00090.00540.02340.03030.0000
breastcancerAvg1.00000.98971.00000.98370.86000.97121.00001.00001.0000
STD0.00180.00360.00350.00320.14000.00000.00000.00340.0000
wineAvg1.00001.00000.98381.00000.79521.00001.00001.00001.0000
STD0.00000.00000.01440.00870.09730.00000.00000.01670.0000
australianAvg0.89150.86710.87780.86140.75360.89130.89520.84250.9130
STD0.00840.01770.01260.02120.07820.00000.00370.02230.0000
zooAvg1.00001.00001.00001.00000.86500.95001.00001.00001.0000
STD0.00000.00000.00910.01530.07650.00000.00000.02600.0000
vehicleAvg0.72780.76940.76410.77040.66490.77000.73330.75980.7759
STD0.00710.00650.01290.01150.05240.00590.01500.00940.0104
lymphographyAvg0.91610.92990.87240.87820.87810.89200.88620.86550.9945
STD0.01740.00630.02420.02970.01230.02340.01610.02090.0000
spectheartAvg0.91570.86540.88360.91820.80260.93900.93900.88300.9417
STD0.01080.00650.01120.01940.04430.00810.00810.01520.0095
breastEWAvg0.99260.96460.99060.98260.91060.99000.98170.96111.0000
STD0.00570.00000.00610.00640.02500.00220.00220.00550.0044
inosphereAvg0.96100.93430.89550.87100.95430.93620.96140.92160.9514
STD0.00830.00960.00460.01270.02760.00900.00670.00680.0125
dermatologyAvg1.00001.00000.97121.00001.00001.00000.99780.97531.0000
STD0.00000.00000.01220.00000.00000.00000.00510.00910.0000
sonarAvg0.95260.89190.95690.91630.76670.98540.98050.95630.9890
STD0.01081.18641.25490.02020.04180.84440.00992.08160.0124
RankingW|T|L1|5|120|4|140|3|150|4|140|1|171|3|141|4|131|4|139|5|4
Table 4. Average fitness comparison of AHL-CSA with competitors.
Table 4. Average fitness comparison of AHL-CSA with competitors.
Dataset BCSA1BCSA2BGWOBGOABPSOBDAGABCSA-TVFLAHL-CSA
habermanAvg0.16900.18520.16900.25010.20140.23390.21770.16900.1500
STD0.00550.01310.03160.01010.00000.01200.01260.01300.0000
ecoliAvg0.16970.15350.12560.13820.12540.11260.15350.15780.1106
STD0.00000.00000.00650.00320.00000.00120.00850.00480.0000
balancescaleAvg0.08920.14460.18420.19220.17630.17050.23080.16840.1842
STD0.00430.00820.01550.00750.01100.02300.01560.00860.0000
irisAvg0.03550.00750.01090.00500.00550.00500.00500.03550.0050
STD0.00000.00000.00470.00500.00000.00000.00000.00000.0000
glassAvg0.19570.24120.19670.22280.28400.31090.26260.30580.1930
STD0.00840.01210.01170.01730.01590.02110.01660.01430.0000
contraceptiveAvg0.45360.47040.46420.45080.47590.43990.44890.48340.4759
STD0.00000.01210.00540.00330.00360.01400.01310.000920.0000
tictactoeAvg0.18000.15510.19090.19210.16030.15960.21750.22350.1521
STD0.00000.00000.02780.00390.00090.00630.02130.02870.0044
breastcancerAvg0.00520.01510.01970.02170.01920.03250.00910.02880.0040
STD0.00100.00350.00350.00290.00000.00000.00160.00310.0000
wineAvg0.00240.00440.02050.00730.00150.00320.00160.09660.0039
STD0.01050.00120.01440.00560.00880.00340.00320.00640.0067
australianAvg0.11070.13540.12500.14110.11270.11190.10700.15970.0891
STD0.00860.01800.01310.02100.00170.01100.00330.02260.0000
zooAvg0.00270.00530.00620.01030.05350.05200.00270.00270.0027
STD0.00490.00390.00690.00690.00000.00430.00250.00510.0000
vehicleAvg0.27450.23590.23880.23270.20860.22000.26870.24270.2300
STD0.00700.00670.01310.01110.00660.00570.01540.00930.0116
lymphographyAvg0.08670.07480.13210.12620.12500.11020.11610.13810.0049
STD0.01680.00660.02330.02880.01230.01140.01530.02050.0072
spectheartAvg0.08790.14000.12020.08730.06220.06450.11610.12070.0819
STD0.01000.00640.01090.01900.00000.00780.00890.01440.0091
breastEWAvg0.01200.04190.01510.02310.03210.00420.02160.04330.0033
STD0.00530.00000.00570.00600.00440.00180.00220.00520.0041
inosphereAvg0.04040.07030.10370.13280.05500.06590.04090.06790.0504
STD0.00810.01020.00480.01270.00860.00890.00640.00430.0125
dermatologyAvg0.00310.00550.03560.00350.00340.00270.00500.02960.0036
STD0.00000.00000.01170.00070.00040.00000.00510.00870.0000
sonarAvg0.01040.11320.04840.08900.09490.01720.02170.15680.0141
STD0.00960.01200.01640.00400.00910.05600.00980.01860.0155
RankingW|T|L1|1|160|0|180|0|180|1|173|0|152|1|150|1|170|0|1810|1|7
Table 5. Feature selection ratio of AHL-CSA and competitors.
Table 5. Feature selection ratio of AHL-CSA and competitors.
Dataset BCSA1BCSA2BGWOBGOABPSOBDAGABCSA-TVFLAHL-CSA
haberman 0.670.670.670.670.580.670.670.670.53
ecoli 0.710.570.590.470.510.710.570.760.71
balancescale 1.001.001.001.000.601.001.001.001.00
iris 0.250.750.970.500.470.650.500.250.50
glass 0.550.640.390.510.430.440.330.410.44
contraceptive 0.560.570.490.770.500.550.770.630.44
tictactoe 1.001.000.890.980.610.780.820.701.00
breastcancer 0.330.490.410.550.480.400.400.390.40
wine 0.240.440.440.450.420.320.170.420.39
australian 0.330.390.390.380.460.430.330.380.30
zoo 0.270.530.460.540.490.630.280.400.32
vehicle 0.500.760.530.390.520.460.280.490.43
lymphography 0.360.540.580.550.280.320.350.500.39
spectheart 0.440.670.510.630.470.510.460.490.44
breastEW 0.470.680.570.590.490.540.380.470.37
inosphere 0.350.520.420.500.460.270.270.440.23
dermatology 0.310.550.500.240.240.270.210.520.36
sonar 0.320.610.570.610.480.270.240.470.32
RankingW|T|L3|2|130|0|180|0|181|0|173|0|151|0|174|0|140|1|175|1|12
Table 6. Maximum accuracy obtained in 30 runs by AHL-CSA and competitors.
Table 6. Maximum accuracy obtained in 30 runs by AHL-CSA and competitors.
Dataset BCSA1BCSA2BGWOBGOABPSOBDAGABCSA-TVFLAHL-CSA
haberman 0.83610.81970.83610.75410.80330.77050.78690.83610.8420
ecoli 0.89120.85070.88060.86570.90120.89550.85070.92660.8955
balancescale 0.92000.86400.82400.81600.83200.87200.77600.84000.8240
iris 1.00001.00001.00001.00001.00001.00001.00001.00001.0000
glass 0.80950.76590.80560.78570.77980.76860.78260.78260.8095
contraceptive 0.54760.53060.54080.55440.52040.56120.55440.53060.5442
tictactoe 0.83000.85340.83250.81680.84820.84700.81150.84290.8562
breastcancer 1.00000.99281.00000.98560.97120.97121.00001.00001.0000
wine 1.00001.00001.00001.00001.00001.00001.00001.00001.0000
australian 0.89860.89860.89130.89860.89000.89130.89860.87680.9130
zoo 1.00001.00001.00001.00000.95001.00001.00001.00001.0000
vehicle 0.73960.78000.78700.78000.73960.78000.75560.77510.7866
lymphography 0.93100.93100.89660.96550.89120.89660.89660.89660.9945
spectheart 0.94340.86790.90570.94340.90570.94340.94100.90570.9490
breastEW 1.00000.96461.00001.00000.94691.00000.98230.97351.0000
inosphere 0.97140.95710.91430.90000.95430.95710.97140.94250.9714
dermatology 1.00001.00000.98121.00001.00001.00000.99790.98631.0000
sonar 1.00000.92681.00000.95120.80491.00001.00000.87801.0000
RankingW|T|L1|8|90|4|140|6|120|5|130|3|151|7|100|6|121|4|136|9|3
Table 7. The p-value of the Wilcoxon rank-sum test results.
Table 7. The p-value of the Wilcoxon rank-sum test results.
        Dataset          BCSA1    BCSA2    BGWO    BGOA    BPSO     BDA      GA         BCSA-TVFL       
haberman<0.05<0.05<0.050.092<0.05<0.0556<0.05<0.05
ecoli<0.05<0.05<0.05<0.05<0.05<0.05<0.050.068
balancescale<0.050.135<0.05<0.05<0.05<0.05<0.05<0.05
iris<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
glass<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
contraceptive<0.05<0.05<0.05<0.050.07450.056<0.05<0.05
tictactoe<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
breastcancer<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
wine<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
australian0.073<0.05<0.05<0.05<0.05<0.05<0.050.0657
zoo<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
vehicle<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
inosphere<0.05<0.05<0.050.081<0.05<0.05<0.05<0.05
lymphography<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
spectheart<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
breastEW<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
dermatology<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
sonar<0.05<0.05<0.05<0.05<0.05<0.05<0.05<0.05
Table 8. Classification accuracy results of filter-based methods compared to the AHL-CSA.
Table 8. Classification accuracy results of filter-based methods compared to the AHL-CSA.
Dataset ReliefFInfoGainCFSGainRatioAHL-CSA
haberman 0.74830.74830.74830.74830.8420
ecoli 0.84520.84520.85410.78870.8955
balancescale 0.77120.90400.63520.90400.8240
iris 0.96000.96000.96000.96001.0000
glass 0.49530.50000.49530.47190.8995
contraceptive 0.44600.50100.55470.49150.5612
tictactoe 0.09380.05980.15600.19560.8562
breastcancer 0.59000.59440.60020.59441.0000
wine 0.94440.88890.77780.88781.0000
australian 0.84340.74780.81880.74780.9130
zoo 0.80000.85000.80000.85001.0000
vehicle 0.59170.60520.60950.60590.7866
lymphography 0.70000.72140.69540.75160.9945
spectheart 0.86230.67130.78540.79580.9417
breastEW 0.92410.89230.87560.91231.0000
inosphere 0.84710.84900.92020.85750.9514
dermatology 0.84930.93810.86620.88041.0000
sonar 0.52380.19050.30950.43200.9634
RankingW|T|L0|0|180|1|170|0|180|1|1717|0|1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Ye, Z.; Gao, B.; Wu, Y.; Yan, X.; Liao, X. A Robust Adaptive Hierarchical Learning Crow Search Algorithm for Feature Selection. Electronics 2023, 12, 3123. https://doi.org/10.3390/electronics12143123

AMA Style

Chen Y, Ye Z, Gao B, Wu Y, Yan X, Liao X. A Robust Adaptive Hierarchical Learning Crow Search Algorithm for Feature Selection. Electronics. 2023; 12(14):3123. https://doi.org/10.3390/electronics12143123

Chicago/Turabian Style

Chen, Yilin, Zhi Ye, Bo Gao, Yiqi Wu, Xiaohu Yan, and Xiangyun Liao. 2023. "A Robust Adaptive Hierarchical Learning Crow Search Algorithm for Feature Selection" Electronics 12, no. 14: 3123. https://doi.org/10.3390/electronics12143123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop