Next Article in Journal
Gamifying Community Education for Enhanced Disaster Resilience: An Effectiveness Testing Study from Australia
Previous Article in Journal
Evaluation of Online Teaching Quality Based on Facial Expression Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EBBA: An Enhanced Binary Bat Algorithm Integrated with Chaos Theory and Lévy Flight for Feature Selection

1
Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
2
Academy for Electromechanical, Changchun Polytechnic, Changchun 130033, China
*
Author to whom correspondence should be addressed.
Future Internet 2022, 14(6), 178; https://doi.org/10.3390/fi14060178
Submission received: 10 May 2022 / Revised: 1 June 2022 / Accepted: 3 June 2022 / Published: 9 June 2022
(This article belongs to the Topic Big Data and Artificial Intelligence)

Abstract

:
Feature selection can efficiently improve classification accuracy and reduce the dimension of datasets. However, feature selection is a challenging and complex task that requires a high-performance optimization algorithm. In this paper, we propose an enhanced binary bat algorithm (EBBA) which is originated from the conventional binary bat algorithm (BBA) as the learning algorithm in a wrapper-based feature selection model. First, we model the feature selection problem and then transfer it as a fitness function. Then, we propose an EBBA for solving the feature selection problem. In EBBA, we introduce the Lévy flight-based global search method, population diversity boosting method and chaos-based loudness method to improve the BA and make it more applicable to feature selection problems. Finally, the simulations are conducted to evaluate the proposed EBBA and the simulation results demonstrate that the proposed EBBA outmatches other comparison benchmarks. Moreover, we also illustrate the effectiveness of the proposed improved factors by tests.

1. Introduction

Communication networks, computers, and artificial intelligence technologies provide a vast array of tools and techniques to improve efficiency. With the development of these technologies and tools, huge amounts of data are generated, stored, and utilized [1]. For example, a large number of IoT devices monitor, sense, and generate continuous data from the edge [2]. In addition, operators retain large amounts of historical user data in various transaction and computing platforms. Then again, the generation of rich media data such as short videos and motion pictures also makes the amount of data in the network grow exponentially [3]. Machine learning can effectively use these data and learn rules and patterns from them to help people make predictions and decisions. Machine learning algorithms have been successfully applied to various fields of life, such as medicine, materials science, and physics. Specifically, machine learning algorithms can extract features from various types of data and use the features to train models for classification, regression, and clustering operations [4].
Despite the advantages of the mentioned machine learning algorithms in terms of effectiveness, wide application, and malleability, there are still some challenges and urgent issues that machine learning algorithms need to address. First, the training of machine learning algorithms is a time-consuming, computationally intensive, and energy-intensive process, which can lead to limited applications of machine learning algorithms [5]. Second, the training datasets of machine learning algorithms are derived from the features they extract, which are mostly extracted using automated tools and human experience, and have many repetitive, meaningless, or even misleading features [6]. These features can slow down the training process of machine learning algorithms even more and reduce the effectiveness of classification, clustering, and regression of machine learning algorithms. Therefore, the elimination of these useless and redundant features is important to improve the performance of machine learning algorithms and reduce their training consumption.
Feature selection is an effective means to solve the above problem. Feature selection can eliminate useless and redundant features in the dataset, thus reducing the number of features and improving the classification accuracy of machine learning algorithms. For a dataset with N features, there are 2 N feature selection schemes available, producing a combinatorial explosion. Therefore, selecting a subset of features with high classification accuracy and a low number of features can be regarded as an optimization problem. On the other hand, the feature selection problem is also proved to be an NP-hard problem. It is important to select and propose a suitable algorithm to solve the feature selection problem.
In general, feature selection methods can be classified into three categories, namely filter-based methods, wrapper-based methods, and embedded methods. Specifically, filter-based methods use a statistical measure that gives a score to the relevance of each feature in the dataset, by which the importance of the feature can be quantified. Subsequently, the decision-maker can set a threshold to remove features with scores below the threshold, thus achieving a reduction in the number of features. However, such methods do not consider the complementarity and mutual exclusivity among features, and therefore the classification accuracy obtained by the subset of features selected by such methods is low [7]. Embedded-based methods are a special kind of wrapper-based method, so they are not discussed in this paper. The wrapper-based methods introduce classifiers and learning algorithms. The learning algorithm continuously generates new feature subsets, while the classifier evaluates the generated feature subsets and selects the optimal one in continuous iterations [8]. In this type of method, the classification accuracy of feature selection is high but consumes more time due to the introduction of classifiers. On the other hand, this type of method has a great relationship with the performance of the learning algorithm.
Akin to some previous works [7,8,9], we aim to adopt the swarm intelligence algorithm as the learning algorithm in the wrapper-based feature selection method. Specifically, a swarm intelligence algorithm studies the complex behaviors of a swarm consisting of several simple agents. By iteratively updating the swarm, the agents will have a more powerful performance than before, so that the algorithm can provide a sub-optimal solution. Swarm intelligence has the benefits of high convergence and powerful solving ability. Moreover, swarm intelligence can also handle NP-hard problems such as feature selection. Thus, it can be seen as an effective method to overcome the challenges of feature selection. For instance, some well-known swarm intelligence, i.e., genetic algorithm (GA) [10], particle swarm optimization (PSO) [11], dragonfly algorithm (DA) [12], ant-lion optimizer (ALO) [13], and grey wolf optimizer (GWO) [14] has been applied in feature selection.
Bat algorithm (BA) and binary BA (BBA) are promising forms of swarm intelligence and have been demonstrated to be better than other algorithms in some applications due to their effectiveness and high performance. However, as suggested in no-free lunch (NFL) theory, there are no algorithms that can suitably solve all optimization problems. In addition, BA also has some shortcomings in solving feature selection problems. Thus, we aim to enhance the performance of BA for solving feature selection. The contributions of this work are summarized as follows:
  • We show that the feature selection is a multi-objective optimization problem, and we present the decision variables and the optimization goals of the feature selection problem.
  • We propose an enhanced binary BA (EBBA) for solving the feature selection problem. In EBBA, we propose Lévy flight-based global search method, which enables the algorithm to jump out of the local optimum. Moreover, we propose a population diversity boosting method so that the exploration capability of the algorithm can be further enhanced. In addition, we use a recently proposed chaotic mapping to assign values to the key parameter of the algorithm, thus enhancing the exploitation capability of the algorithm.
  • Simulations are conducted based on open datasets of UC Irvine machine learning repository to verify the solving ability of the proposed EBBA. First, we introduce some benchmark algorithms for comparisons. Then, we show the effectiveness of the proposed improved factors.
The rest of this work is arranged as follows. Section 2 reviews some key related works about swarm intelligence algorithms and feature selection. Section 3 gives the model of feature selection. Section 4 proposes the EBBA and details the improved factors. Section 5 provides simulation results and Section 6 concludes this work.

2. Related Works

In this work, we aim to use one of the swarm intelligence algorithms, i.e., BA, to solve the feature selection problem, and thus some key related works are briefly introduced in this section.

2.1. Swarm Intelligence Algorithms

Swarm intelligence algorithms refer to evolutionary theory and swarm behavior. In the past few years, a large number of researchers have proposed various types of swarm intelligence algorithms to solve optimization problems in different domains.
First, some representative classical swarm intelligence algorithms are presented as follows. PSO is another representative swarm intelligence algorithm, which is inspired by the behavior of bird/fish populations. Moreover, artificial bee colony (ABC) [15], ant colony optimization (ACO) [16], etc., are also well-known swarm intelligence algorithms. Second, swarm intelligence algorithms also contain various types of bio-inspired algorithms. For example, Meng et al. [17] proposed a chicken swarm optimization (CSO) to solve optimization problems by simulating the rank order and the behavior of chickens (including roosters, hens, and chicks) in a flock. Yang et al. [18] proposed a BA and validated the performance of BA using eight nonlinear engineering optimization problems. Third, certain swarm intelligence algorithms were proposed inspired by various natural phenomena of the universe. Jiang et al. [19] proposed a new metaheuristic method, artificial raindrop algorithm (ARA), from natural rainfall phenomena and used it for the identification of unknown parameters of chaotic systems. Kaveh et al. [20] proposed a ray optimization (RO) based on Snell’s law of light refraction and the phenomenon of light refraction.
In summary, researchers have proposed a large number of effective swarm intelligence algorithms and applied them to various optimization problems. However, these algorithms are not necessarily applicable to all engineering fields. Accordingly, proposing an enhanced swarm intelligence algorithm version according to the characteristics of an optimization problem is a major challenge.

2.2. Ways of Feature Selection

There are several existing methods and ways have been proposed for the purpose of feature selection. First, some filter methods are widely used due to their simplicity and relatively high performance. For instance, some methods based on correlation criteria and mutual information are detailed in reference [21]. In this case, several effective filter-based algorithms including correlation-based feature selection (CFS) [22], fast correlation-based filter (FCBF) [23], wavelet power spectrum (Spectrum) [24], Information Gain (IG) [25], ReliefF [26], etc. Second, wrapper-based approaches are key methods in feature selection. This type of method can be categorized by the type of learning algorithms. For instance, exhaustive, random search and metaheuristic search methods. Due to their effectiveness, the metaheuristic search methods including swarm intelligence algorithms can be seen as the most popular methods [27]. Finally, there are several embedded methods. The main approach is to incorporate feature selection as part of the training process, e.g., [21,28].

2.3. Swarm Intelligence-Based Feature Selection

There are many swarm intelligence algorithms have been adopted or proposed as the learning algorithm in wrapper-based feature selection methods, and we review some key algorithms as follows.
Li et al. [29] proposed an improved binary GWO (IBGWO) algorithm for solving feature selection problems, in which an enhanced opposition-based learning (E-OBL) initialization and a local search strategy were proposed for improving the performance of the algorithm. Kale et al. [30] presented four different improved versions of the sine cosine algorithm (SCA), where the updating mechanism of SCA is the improvements and innovations. Ouadfel et al. [31] proposed a hybrid feature selection approach based on the ReliefF filter method and equilibrium optimizer (EO), which is composed of two phases and tested in some open datasets. Abdel-Basset et al. [14] proposed three variants of BGWO in addition to the standard variant, applying different transfer functions to tackle the feature selection problem. In [32], two different wrapper feature selection approaches were proposed based on farmland fertility algorithm (FFA), which denoted as BFFAS and BFFAG, and these methods are effective in solving feature selection problems. On the other hand, BA and some variants have been adopted for solving feature selection problems. Varma et al. [33] proposed a bat optimization algorithm for wrapper-based feature selection and conducted simulations based on the CICInvesAndMal2019 benchmark dataset. Naik et al. [34] proposed a feature selection method to identify the relevant subset of features for the machine-learning task using the wrapper approach via BA. Rodrigues et al. [35] presented a wrapper feature selection approach based on bat algorithm (BA) and optimum-path forest (OPF). In [36], the authors proposed an improved BPSO algorithm as an essential tool of pre-processing for solving classification problem, in which a new updating mechanism for calculating Pbest and Gbest were proposed. Moreover, the authors in [37] proposed a binary DA (BDA) and use it to solve the feature selection problems. Likewise, Nakamura et al. [38] proposed a binary version of the bat algorithm, i.e., BBA, and evaluate its performance in solving the feature selection problems. In addition, in [39], a new hybrid feature selection method was proposed by using the sine cosine algorithm (SCA) and genetic algorithm (GA), and the algorithm is used for solving feature selection problems. Furthermore, Nagpal et al. [40] proposed a feature selection method via binary gravitational search algorithms (BGSA) in medical datasets, in which they can reduce the number of features by an average of 66% and enhance the accuracy of prediction.
The aforementioned methods can solve feature selection problems in various applications. However, according to NFL theory, different swarm intelligence algorithms may have different performances in various applications. Therefore, the existing methods are insufficient to solve all feature selection problems, which motivates us to propose an EBBA to handle more feature selection problems in this work.

3. Feature Selection Model

As shown in [8,29,33], the feature selection problem can be seen as a binary optimization model, and in this section, we introduce it in details. Specifically, the main purpose of feature selection is to reduce the data dimension by retaining the most valuable features through feature selection methods. Thus, there are two possibilities for each feature, i.e., to be selected and to be discarded. Therefore, the feature selection problem can be regarded as an optimization problem with a binary solution space.
It is can be seen from Figure 1, the solution space of the considered feature selection problem is a binary. Each feature is represented by a binary number, and if that binary number is 1, it means that the feature is selected, and conversely, if that binary number is 0, it means that the feature is discarded. Thus, the feature selection of a dataset can be represented by a binary array as follow:
( Decision variables ) X = [ x 1 , x 2 , x 3 , , x N d i m ] ,
where N d i m is the number of features, in other words, the dimension number of the dataset. Under this model, there are two main objectives of the feature selection, i.e., to reduce the classification error rate of the obtained feature subsets, and to reduce the feature number of feature subsets. Thus, the feature selection problem is a multi-objective problem in which the first objective can be expressed as follows:
( Objective 1 ) f 1 = 1 f a c c ,
where f a c c is the classification accuracy of the obtained feature subsets. Note that we introduce the KNN as a classifier to evaluate the feature subsets and the reasons are analyzed in following section. Moreover, the second objective of this work is to reduce the feature number of feature subsets, which can be expressed as follows:
( Objective 2 ) f 2 = N d i m N d i m ,
where N d i m is the feature number of the selected feature subsets. To simultaneously the aforementioned objectives, we introduce the fitness function as follows:
( Fitness function ) f f i t = a × f 1 + b × f 2 ,
where a [ 0 , 1 ] and b = 1 a are constants that denote the weights of the two objectives f 1 and f 2 , respectively. Specifically, we can increase a to obtain a higher classification accuracy or increase b to obtain a smaller dimensional feature subset.

4. Proposed Algorithm

Based on the aforementioned feature selection model, we can optimize the decision variables shown in Equation (1) to obtain a better fitness function shown in Equation (4). Accordingly, we propose an EBBA in this section for solving the feature selection problem.

4.1. Conventional BA

BA is a swarm intelligence algorithm for global optimization, which is inspired by the echolocation behavior of bats. Specifically, bats look for prey by flying at a random velocity V i at a random point X i with a fixed frequency f m i n , changing wavelength l, and loudness A 0 . Depending on the proximity of their target, these bats can autonomously modify the wavelength (in other words, frequency) of their generated pulses as well as the rate of pulse emission r in the range of [ 0 , 1 ] . The corresponding mathematical model of BA can be detailed as follows. In the tth iteration, the frequency f i of the the ith bat is expressed as follows.
f i = f m i n + ( f m a x f m i n ) × β ,
where f m a x and f m i n are upper and lower bounds on the frequencies of all bats, respectively, and β is a random number between [ 0 , 1 ] .
Moreover, the velocity of the ith bat v i can be modeled as follows:
V i t = V i t 1 + ( X i t 1 X * ) f i ,
where X * is the bat with the highest fitness function value of the swarm.
In addition, the update method of the ith bat is shown as follows:
X i t = X i t 1 + V i t ,
where X i t is the position of the ith bat in the tth iteration.
Furthermore, BA also enhances search ability through local random walks. Specifically, BA asks the best bat in the swarm to conduct a local search with a certain probability, which can be expressed as follows:
X N = X * + ϵ × A t ,
where X N is the newly generated bat after the random walk, A t is the loudness of all bats in the t iteration, and ϵ is a random variable that ranges [ 1 , 1 ] .
Additionally, the loudness A i and the rate r i of pulse emission are also updated as the iterations proceed, which is shown as follows:
A i t + 1 = α A i t , r i t + 1 = r i 0 [ 1 exp ( γ t ) ] ,
where α is a parameter that ranges from 0 to 1, and γ > 0 is a parameter.
By using these mathematical models, the main steps of BA can be summarized as follows.
Step 1: Randomly generate population (bat swarm) P = [ X 1 , X 2 , , X N p o p ] , where N p o p is the population size. Moreover, the velocity, pulse emissivity, and loudness of the bats are randomly generated. Then, the fitness values of all bats are calculated.
Step 2: Update the positions and velocities of bats by using Equations (5)–(9).
Step 3: A random number N r a n d between 0 and 1 is firstly generated, and then if N r a n d > r i , a random walk will be performed by using Equation (8) to generate a new individual X N around the current best individual X * .
Step 4: Generate a random number N r a n d of [ 0 , 1 ] again. If N r a n d < A i and f f i t ( X N ) < f f i t ( X i ) , replace X i with X N and then update the loudness and pulse firing rate. If f f i t ( X N ) < f f i t ( X * ) , X N is used to replace X * .
Step 5: Repeat steps 2–4 until the terminal condition is reached.
Step 6: Return X * as the final solution to the problem.

4.2. BBA

To make the BA can handle the binary solution space of the feature selection, Mirjalili et al. [41] introduce a binary operator. Specifically, the authors introduced a v-shape transfer function to map the continuous parameters into binary solution space, which can be expressed as follows:
V v i , j t = 2 π arctan π 2 v i , j t ,
x i , j t + 1 = x i , j t 1 if rand < V v i , j t + 1 x i , j t if rand V v i , j t + 1
where x i , j t and v i , j t indicate the position and velocity of ith individual at tth iteration in jth dimension, and ( x i , j t ) 1 is the complement of x i , j t . As such, the BBA can handle and update the binary decision variable reasonably.

4.3. EBBA

Conventional BA may confront some key challenges in solving the feature selection problems. First, when dealing with the big solution space of feature selection problem, BA may lack exploration ability, which may make the algorithm fall in local optima. Second, the bats of the BA are guided by the best bat of the swarm, i.e., X * , which means that the population diversity is lower for the large scale datasets. Third, the exploration and exploration abilities of BA should be further balanced. Finally, BA is proposed for continuous problems whereas the feature selection problems are with a binary solution space. Thus, these reasons motivate us to enhance BA for better feature selection performance. The main steps of the proposed EBBA are detailed in Algorithm 1, and the correspondingly improved factors are as follows:
Algorithm 1 EBBA
Futureinternet 14 00178 i001

4.3.1. Lévy Flight-Based Global Search Method

Feature selection problems also are large-scale optimization problems since the dimension of some datasets is relatively large. In this case, the exploration ability of the optimization algorithm should be sufficient. However, the update of other bats in the swarm are determined by the best bat, and this mechanism will undoubtedly decrease the exploration ability of the algorithm. Thus, we introduce the Lévy flight to propose a global search method to improve the exploration ability of the algorithm. Specifically, a Lévy flight is a random walk in which the step-lengths have a Lévy distribution, a probability distribution that is heavy-tailed. By using the short-distance or long-distance searching alternately, the search scope can be extended.
First, mathematically, in each iteration, we generate a new bat according to the best bat X * and Lévy flight, which can be expressed as follows:
X n e w = X * + α vy ( λ ) ,
where α is a parameter and its value is often assigned according to applications. Moreover, Lévy flight is taken from the Lévy distribution, which can be expressed as follows:
vy ( λ ) u = t λ ( 1 < λ < 3 ) ,
Second, the newly generated bat X n e w is evaluated to obtain its fitness value, and then we compare the X n e w with the best bat X * based on their fitness function values. If the X n e w outmatches X * , then X * = X n e w . By using this method, the best bat of the swarm is easy to jump out of local optima, thereby enhancing the exploration ability of EBBA.

4.3.2. Population Diversity Boosting Method

In BA, all the bats are guided by the best bat of the swarm, which may decrease the population diversity of the algorithm, thereby affecting the solving performance. In this case, we aim to propose a method for boosting population diversity. In bat swarm, the second best bat also is meaningful and with a strong value for guiding other bats. Thus, we use the second best bat X * * to generate a part of new bats as follows:
X N 2 = X * * + ϵ × A t ,
where X N 2 is the newly generated bat and whether the method used is determined by a parameter r which can be expressed as follows:
r i t = r i t 2 ,
By using this method, the bat swarm is simultaneously guided by the best bat and the second-best bat, so that enhancing the population diversity of the algorithm.

4.3.3. Chaos-Based Loudness Method

The loudness in BA can determine the weights of exploitation and exploration abilities of the BA. However, the loudness update method of conventional BA is linear, which may be unsuitable for the feature selection. Thus, we introduce a novel fractional one dimensional chaotic map to update loudness [42], which can be expressed as follows:
A i t = C t ,
where C t is the tth dimension of the fractional one dimensional chaotic map, which can be expressed as follows:
C t + 1 = f C t = 1 C t 2 + α β C t if C t 0 , 1 α 1 C t 2 + α β C t if C t 1 α , 0 ,
where α and β are two real parameters, and they are assigned as 0.001 and 0.9 in this work, respectively. By using the high chaotic behavior of the method, the exploitation and exploration abilities of EBBA can be balanced.

4.3.4. Complexity Analysis of EBBA

The complexity of EBBA is analyzed in this part. In the proposed feature selection model, the most time-consuming step is the calculation of fitness function value since we introduce classifier, which is several orders of magnitude complex than other steps. In this case, other steps can be omitted. Accordingly, the complexity of EBBA is O ( t m a x · N p o p ) when the maximum number of iteration and population size are denoted as t m a x and N p o p , respectively.

5. Simulations

In this section, we conduct the simulations to evaluate the proposed EBBA. First, the datasets and setups are presented. Second, we compare the EBBA with some benchmark algorithms. Third, we illustrate the effectiveness of the improved factors.

5.1. Datasets and Setups

In this work, we introduce 10 typical UC Irvine Machine Learning Repository datasets. The main information of these datasets are shown in Table 1.
Moreover, the used CPU is 11th Gen Intel(R) Core(TM) i7-11700 @ 2.50 GHz and the RAM is 16 GB. We use python to implement the simulation codes and adopt KNN (k = 5). Note that we use KNN classifier since it is simple, easy and highly accurate, it is also insensitive to outliers and no data entry settings. Moreover, using a simple and relatively cheap classification algorithm in a wrapper approach can obtain a good feature subset that is also suitable for complex classification algorithms. In contrast, if an advanced classification algorithm is used for wrapper-based feature selection, the obtained feature subset will be failed for simple classification algorithms. The reason is that when using advanced classification algorithms, the learning algorithm of the wrapper approach (e.g., the proposed EBBA) will capture the characteristics of the classification algorithm instead of the relationship of different features. In addition, a and b in the fitness function are set to 0.99 and 0.01, respectively. Furthermore, In this paper, binary PSO (BPSO) [36], BGWO [43], BDA [37], and BBA [41] are introduced as the benchmarks, and the key parameters of these algorithms are shown in Table 2. Note that the population size and iteration number of EBBA and other benchmarks are set as 24 and 100, respectively. Additionally, to avoid the experiment’s random bias, each algorithm is performed 30 times independently in these selected datasets, as specified by the central limit theorem. What’s more, 80% of the instances are utilized for training, while the remaining 20% are used for testing [7,44,45]. Note that the classifier only provides feedback to the EBBA, which means that the overfitting affected by the division of the dataset will not have too much impact on the feature selection method based on the wrapper and swarm intelligence.

5.2. Simulation Results

Table 3 shows the optimization results of the accuracy rate, number of selected features, fitness function values, and CPU times achieved by various algorithms. Note that the best values among all comparison algorithms are highlighted in bold font. As can be seen, the proposed EBBA achieves the best accuracy rate on 7 datasets and achieves the best of selected feature number on 2 datasets. More intuitively, the proposed EBBA achieves the best fitness function values on 9 datasets, which means that the proposed EBBA is with the best performance among all benchmark algorithms. The reason may be that we enhance the EBBA by improving its exploration ability and balancing its exploration and exploitation abilities, which make the EBBA more suitable for solving feature selection problems.
In addition, Figure 2 shows the convergence rates obtained by different benchmark algorithms during the solving processes. As can be seen, the proposed EBBA achieves the best curves on most datasets, which performs the best convergence ability among all the comparison algorithms.
On the other hand, we also evaluate the effectiveness of the proposed improved factors. Specifically, we combine the proposed Lévy flight-based global search method, population diversity boosting method, and chaos-based loudness method with conventional BBA, namely, EBBA-IF1, EBBA-IF2, and EBBA-IF3, respectively. Table 4 shows the optimization results of the accuracy rate, number of selected features, fitness function values, and CPU times achieved by various algorithms. Moreover, Figure 3 shows the convergence rates obtained by different EBBA versions during the solving processes. As can be seen, the EBBA, EBBA-IF1, EBBA-IF2, and EBBA-IF3 outperform conventional BBA, which means that the proposed improved factors are non-trivial and effective.

5.3. Performance Evaluation under Different Classifiers

In this section, we consider two other classification algorithms which are decision tree and random forest. Specifically, the decision tree is easy to understand and visualized, requires only little data preparation, whereas it may be easy to be overfitting. Likewise, random forest is also extremely accurate and not prone to overfitting, and can run effectively on large datasets with good noise immunity. Moreover, due to the complexity of the classification algorithms, validating the newly introduced two classifiers using all datasets is a huge and time-consuming task. Thus, we only consider using the dataset SpectEW as the experimental dataset since it has the middle dimension number of all datasets, which is representative. In addition, other settings are similar to that of the KNN-based method.
In this case, Table 5 and Table 6 provide the simulation results obtained by different algorithms in terms of accuracy, the number of features and fitness function value under decision tree and random forest, respectively. As can be seen, the proposed EBBA also outperforms other comparison algorithms under other classification algorithms, which shows that the improved factors are effective even if the classification algorithm is changed. Thus, the proposed method has good performance in both maybe overfitting and non-overfitting cases.

6. Conclusions

In this paper, the feature selection problems which can enhance the classification and reduce data dimension are studied. First, we model the feature selection problem and then transfer it as a fitness function. Then, we propose an EBBA for solving the feature selection problem. In EBBA, we introduce Lévy flight-based global search method, population diversity boosting method and chaos-based loudness method to improve the BA and make it more applicable to feature selection problems. Finally, the simulations are conducted to evaluate the proposed EBBA and the simulation results demonstrate that the proposed EBBA outmatches other comparison benchmarks. Moreover, the non-trivial of the proposed improved factors is illustrated. In the future, we intend to use more realistic datasets to evaluate the proposed EBBA.

Author Contributions

Data curation, H.K.; Supervision, L.Z.; Writing—original draft, J.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by National Natural Science Foundation of China under Grant 82090052, and Shanghai Municipal Science and Technology Major Project 2021SHZDZX0103.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xia, Q.; Zhou, L.; Ren, W.; Wang, Y. Proactive and intelligent evaluation of big data queries in edge clouds with materialized views. Comput. Netw. 2022, 203, 108664. [Google Scholar] [CrossRef]
  2. Berger, C.; Eichhammer, P.; Reiser, H.P.; Domaschka, J.; Hauck, F.J.; Habiger, G. A Survey on Resilience in the IoT: Taxonomy, Classification, and Discussion of Resilience Mechanisms. ACM Comput. Surv. 2022, 54, 147. [Google Scholar] [CrossRef]
  3. Xu, L.; Tolmochava, T.; Zhou, X. Search History Visualization for Collaborative Web Searching. Big Data Res. 2021, 23, 100180. [Google Scholar] [CrossRef]
  4. Notash, A.Y.; Bayat, P.; Haghighat, S.; Notash, A.Y. Evolutionary ensemble feature selection learning for image-based assessment of lymphedema arm volume. Concurr. Comput. Pract. Exp. 2022, 34, e6334. [Google Scholar] [CrossRef]
  5. Abdulla, M.; Khasawneh, M.T. Integration of aggressive bound tightening and Mixed Integer Programming for Cost-sensitive feature selection in medical diagnosis. Expert Syst. Appl. 2022, 187, 115902. [Google Scholar] [CrossRef]
  6. Alsahaf, A.; Petkov, N.; Shenoy, V.; Azzopardi, G. A framework for feature selection through boosting. Expert Syst. Appl. 2022, 187, 115895. [Google Scholar] [CrossRef]
  7. Li, J.; Kang, H.; Sun, G.; Feng, T.; Li, W.; Zhang, W.; Ji, B. IBDA: Improved Binary Dragonfly Algorithm With Evolutionary Population Dynamics and Adaptive Crossover for Feature Selection. IEEE Access 2020, 8, 108032–108051. [Google Scholar] [CrossRef]
  8. Ji, B.; Lu, X.; Sun, G.; Zhang, W.; Li, J.; Xiao, Y. Bio-Inspired Feature Selection: An Improved Binary Particle Swarm Optimization Approach. IEEE Access 2020, 8, 85989–86002. [Google Scholar] [CrossRef]
  9. Agrawal, P.; Ganesh, T.; Oliva, D.; Mohamed, A.W. S-shaped and V-shaped gaining-sharing knowledge-based algorithm for feature selection. Appl. Intell. 2022, 52, 81–112. [Google Scholar] [CrossRef]
  10. Lappas, P.Z.; Yannacopoulos, A.N. A machine learning approach combining expert knowledge with genetic algorithms in feature selection for credit risk assessment. Appl. Soft Comput. 2021, 107, 107391. [Google Scholar] [CrossRef]
  11. Li, A.; Xue, B.; Zhang, M. Improved binary particle swarm optimization for feature selection with new initialization and search space reduction strategies. Appl. Soft Comput. 2021, 106, 107302. [Google Scholar] [CrossRef]
  12. Too, J.; Mirjalili, S. A Hyper Learning Binary Dragonfly Algorithm for Feature Selection: A COVID-19 Case Study. Knowl. Based Syst. 2021, 212, 106553. [Google Scholar] [CrossRef]
  13. Wang, M.; Wu, C.; Wang, L.; Xiang, D.; Huang, X. A feature selection approach for hyperspectral image based on modified ant lion optimizer. Knowl. Based Syst. 2019, 168, 39–48. [Google Scholar] [CrossRef]
  14. Abdel-Basset, M.; Sallam, K.M.; Mohamed, R.; Elgendi, I.; Munasinghe, K.S.; Elkomy, O.M. An Improved Binary Grey-Wolf Optimizer With Simulated Annealing for Feature Selection. IEEE Access 2021, 9, 139792–139822. [Google Scholar] [CrossRef]
  15. Bacanin, N.; Bezdan, T.; Venkatachalam, K.; Zivkovic, M.; Strumberger, I.; Abouhawwash, M.; Ahmed, A.B. Artificial Neural Networks Hidden Unit and Weight Connection Optimization by Quasi-Refection-Based Learning Artificial Bee Colony Algorithm. IEEE Access 2021, 9, 169135–169155. [Google Scholar] [CrossRef]
  16. Zhao, H.; Zhang, C.; Zheng, X.; Zhang, C.; Zhang, B. A decomposition-based many-objective ant colony optimization algorithm with adaptive solution construction and selection approaches. Swarm Evol. Comput. 2022, 68, 100977. [Google Scholar] [CrossRef]
  17. Meng, X.; Liu, Y.; Gao, X.Z.; Zhang, H. A New Bio-inspired Algorithm: Chicken Swarm Optimization. In Advances in Swarm Intelligence; Tan, Y., Shi, Y., Coello, C.A.C., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8794, pp. 86–94. [Google Scholar] [CrossRef]
  18. Yang, X.S.; Gandomi, A.H. Bat algorithm: A novel approach for global engineering optimization. Eng. Comput. 2012, 29, 464–483. [Google Scholar] [CrossRef] [Green Version]
  19. Jiang, Q.; Wang, L.; Hei, X. Parameter identification of chaotic systems using artificial raindrop algorithm. J. Comput. Sci. 2015, 8, 20–31. [Google Scholar] [CrossRef]
  20. Kaveh, A.; Khayatazad, M. A new meta-heuristic method: Ray Optimization. Comput. Struct. 2012, 112–113, 283–294. [Google Scholar] [CrossRef]
  21. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  22. Doshi, M. Correlation based feature selection (CFS) technique to predict student Perfromance. Int. J. Comput. Netw. Commun. 2014, 6, 197. [Google Scholar] [CrossRef]
  23. Senliol, B.; Gulgezen, G.; Yu, L.; Cataltepe, Z. Fast Correlation Based Filter (FCBF) with a different search strategy. In Proceedings of the 2008 23rd International Symposium on Computer and Information Sciences, Istanbul, Turkey, 27–29 October 2008; pp. 1–4. [Google Scholar]
  24. Subramani, P.; Sahu, R.; Verma, S. Feature selection using Haar wavelet power spectrum. BMC Bioinform. 2006, 7, 432. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Azhagusundari, B.; Thanamani, A.S. Feature selection based on information gain. Int. J. Innov. Technol. Explor. Eng. (IJITEE) 2013, 2, 18–21. [Google Scholar]
  26. Spolaôr, N.; Cherman, E.A.; Monard, M.C.; Lee, H.D. ReliefF for Multi-label Feature Selection. In Proceedings of the Brazilian Conference on Intelligent Systems, BRACIS 2013, Fortaleza, Brazil, 19–24 October 2013; pp. 6–11. [Google Scholar] [CrossRef]
  27. Rostami, M.; Berahmand, K.; Nasiri, E.; Forouzandeh, S. Review of swarm intelligence-based feature selection methods. Eng. Appl. Artif. Intell. 2021, 100, 104210. [Google Scholar] [CrossRef]
  28. Dhal, P.; Azad, C. A comprehensive survey on feature selection in the various fields of machine learning. Appl. Intell. 2022, 52, 4543–4581. [Google Scholar] [CrossRef]
  29. Li, W.; Kang, H.; Feng, T.; Li, J.; Yue, Z.; Sun, G. Swarm Intelligence-Based Feature Selection: An Improved Binary Grey Wolf Optimization Method. In Knowledge Science, Engineering and Management; Qiu, H., Zhang, C., Fei, Z., Qiu, M., Kung, S., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2021; Volume 12817, pp. 98–110. [Google Scholar] [CrossRef]
  30. Kale, G.A.; Yüzgeç, U. Advanced strategies on update mechanism of Sine Cosine Optimization Algorithm for feature selection in classification problems. Eng. Appl. Artif. Intell. 2022, 107, 104506. [Google Scholar] [CrossRef]
  31. Ouadfel, S.; Elaziz, M.A. Efficient high-dimension feature selection based on enhanced equilibrium optimizer. Expert Syst. Appl. 2022, 187, 115882. [Google Scholar] [CrossRef]
  32. Hosseinalipour, A.; Gharehchopogh, F.S.; Masdari, M.; Khademi, A. A novel binary farmland fertility algorithm for feature selection in analysis of the text psychology. Appl. Intell. 2021, 51, 4824–4859. [Google Scholar] [CrossRef]
  33. Varma, P.R.K.; Mallidi, S.K.R.; Jhansi, S.J.; Dinne, P.L. Bat optimization algorithm for wrapper-based feature selection and performance improvement of android malware detection. IET Netw. 2021, 10, 131–140. [Google Scholar] [CrossRef]
  34. Naik, A.K.; Kuppili, V.; Edla, D.R. Efficient feature selection using one-pass generalized classifier neural network and binary bat algorithm with a novel fitness function. Soft Comput. 2020, 24, 4575–4587. [Google Scholar] [CrossRef]
  35. Rodrigues, D.; Pereira, L.A.M.; Nakamura, R.Y.M.; Costa, K.A.P.; Yang, X.; de Souza, A.N.; Papa, J.P. A wrapper approach for feature selection based on Bat Algorithm and Optimum-Path Forest. Expert Syst. Appl. 2014, 41, 2250–2258. [Google Scholar] [CrossRef]
  36. Huda, R.K.; Banka, H. A group evaluation based binary PSO algorithm for feature selection in high dimensional data. Evol. Intell. 2021, 14, 1949–1963. [Google Scholar] [CrossRef]
  37. Mafarja, M.M.; Eleyan, D.; Jaber, I.; Hammouri, A.; Mirjalili, S. Binary dragonfly algorithm for feature selection. In Proceedings of the 2017 International Conference on New Trends in Computing Sciences (ICTCS), Amman, Jordan, 11–13 October 2017; pp. 12–17. [Google Scholar]
  38. Nakamura, R.Y.M.; Pereira, L.A.M.; Rodrigues, D.; Costa, K.A.P.; Papa, J.P.; Yang, X.S. Binary bat algorithm for feature selection. In Swarm Intelligence and Bio-Inspired Computation; Elsevier: Amsterdam, The Netherlands, 2013; pp. 225–237. [Google Scholar]
  39. Abualigah, L.M.; Dulaimi, A.J. A novel feature selection method for data mining tasks using hybrid Sine Cosine Algorithm and Genetic Algorithm. Clust. Comput. 2021, 24, 2161–2176. [Google Scholar] [CrossRef]
  40. Nagpal, S.; Arora, S.; Dey, S.; Shreya. Feature selection using gravitational search algorithm for biomedical data. Procedia Comput. Sci. 2017, 115, 258–265. [Google Scholar] [CrossRef]
  41. Mirjalili, S.; Mirjalili, S.M.; Yang, X.S. Binary bat algorithm. Neural Comput. Appl. 2014, 25, 663–681. [Google Scholar] [CrossRef]
  42. Talhaoui, M.Z.; Wang, X. A new fractional one dimensional chaotic map and its application in high-speed image encryption. Inf. Sci. 2021, 550, 13–26. [Google Scholar] [CrossRef]
  43. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing 2016, 172, 371–381. [Google Scholar] [CrossRef]
  44. Mafarja, M.M.; Aljarah, I.; Heidari, A.A.; Hammouri, A.I.; Faris, H.; Al-Zoubi, A.M.; Mirjalili, S. Evolutionary Population Dynamics and Grasshopper Optimization approaches for feature selection problems. Knowl. Based Syst. 2018, 145, 25–45. [Google Scholar] [CrossRef] [Green Version]
  45. Tubishat, M.; Ja’afar, S.; Alswaitti, M.; Mirjalili, S.; Idris, N.; Ismail, M.A.; Omar, M.S. Dynamic Salp swarm algorithm for feature selection. Expert Syst. Appl. 2021, 164, 113873. [Google Scholar] [CrossRef]
Figure 1. Feature selection model with binary solution space.
Figure 1. Feature selection model with binary solution space.
Futureinternet 14 00178 g001
Figure 2. Convergence rates achieved by various algorithms.
Figure 2. Convergence rates achieved by various algorithms.
Futureinternet 14 00178 g002
Figure 3. Convergence rates obtained by different EBBA versions.
Figure 3. Convergence rates obtained by different EBBA versions.
Futureinternet 14 00178 g003
Table 1. Datasets.
Table 1. Datasets.
DatasetNumber of FeaturesNumber of Instance
1Breastcancer10699
2BreastEW30569
3Congress16435
4Exactly131000
5Exactly2131000
6HeartEW13270
7SonarEW60208
8SpectEW22267
9tic-tac-toe9958
10Vote16300
Table 2. Key parameters of benchmark algorithms.
Table 2. Key parameters of benchmark algorithms.
AlgorithmKey Parameters
1BPSO c 1 = 2, c 2 = 2
2BGWO α = [2, 0]
3BDAw = [0.9, 0.4], s = [0.2, 0], a = [0.2, 0], c = [0.2, 0], f = [0.2, 0], e = [0, 0.1]
4BBAA = 0.25, Q m a x = 2, Q m i n = 0
5EBBA Q m a x = 2, Q m i n = 0
Table 3. Optimization results achieved by various algorithms. (The best values are highlighted in bold).
Table 3. Optimization results achieved by various algorithms. (The best values are highlighted in bold).
AccuracyFeature #Fitness ValueCPU Time
BreastcancerBBA0.97866.00000.027269.5141
BDA0.97675.80000.028968.7234
BGWO0.97676.50000.029675.3848
BPSO0.97866.00000.027269.2956
EBBA0.97866.00000.027260.9926
BreastEWBBA0.96136.60000.040573.3552
BDA0.95895.16670.042467.3963
BGWO0.953214.66670.051279.7631
BPSO0.96128.50000.041375.4869
EBBA0.96145.43330.040063.4576
CongressBBA0.97936.90000.024858.3419
BDA0.97435.16670.028755.9260
BGWO0.97508.80000.030361.5544
BPSO0.97856.73330.025559.3514
EBBA0.97936.40000.024549.2745
ExactlyBBA1.00006.00000.004691.2861
BDA0.91976.66670.0846100.3068
BGWO0.90408.03330.1012116.4705
BPSO0.99996.03330.0048103.6469
EBBA1.00006.00000.004688.3184
Exactly2BBA0.78542.26670.214286.8956
BDA0.78471.76670.214590.0558
BGWO0.76937.53330.2342117.3080
BPSO0.78901.06670.2097101.6048
EBBA0.78681.96670.212681.4770
HeartEWBBA0.85115.03330.151345.4301
BDA0.83914.93330.163144.0686
BGWO0.83576.66670.167845.2137
BPSO0.85275.00000.149744.8358
EBBA0.85205.06670.150439.0150
SonarEWBBA0.910027.30000.093749.9454
BDA0.901020.50000.101548.1458
BGWO0.898940.66670.106955.5145
BPSO0.904627.73330.099151.0521
EBBA0.911727.06670.091942.8189
SpectEWBBA0.737910.50000.264346.5456
BDA0.71858.96670.282744.8928
BGWO0.721914.36670.281947.5410
BPSO0.733510.50000.268746.3761
EBBA0.740710.70000.261538.4262
tic-tac-toeBBA0.84938.86670.159097.9119
BDA0.80997.16670.196297.8065
BGWO0.84658.80000.161891.4816
BPSO0.85219.00000.156488.2232
EBBA0.85219.00000.156484.4753
VoteBBA0.95195.20000.050946.4947
BDA0.94575.76670.057445.8363
BGWO0.94408.53330.060847.9505
BPSO0.94934.90000.053246.5221
EBBA0.95135.40000.051638.9119
Table 4. Optimization results achieved by various EBBA versions. (The best values are highlighted in bold).
Table 4. Optimization results achieved by various EBBA versions. (The best values are highlighted in bold).
AccuracyFeature #Fitness ValueCPU Time
BreastcancerBBA0.97866.00000.027269.5141
BBA-IF10.97866.00000.027264.4441
BBA-IF20.97866.00000.027267.0577
BBA-IF30.97866.00000.027260.1375
EBBA0.97866.00000.027260.9926
BreastEWBBA0.96136.60000.040573.3552
BBA-IF10.96135.23330.040065.2602
BBA-IF20.96145.90000.040268.0782
BBA-IF30.96115.40000.040363.0563
EBBA0.96145.43330.040063.4576
CongressBBA0.97936.90000.024858.3419
BBA-IF10.97906.66670.024951.4119
BBA-IF20.97836.23330.025354.6810
BBA-IF30.97916.66670.024950.4441
EBBA0.97936.40000.024549.2745
ExactlyBBA1.00006.00000.004691.2861
BBA-IF11.00006.00000.004688.5090
BBA-IF21.00006.00000.004692.2044
BBA-IF31.00006.00000.004688.9417
EBBA1.00006.00000.004688.3184
Exactly2BBA0.78542.26670.214286.8956
BBA-IF10.78592.03330.213681.4199
BBA-IF20.78482.53330.215087.3399
BBA-IF30.78482.43330.214984.0370
EBBA0.78681.96670.212681.4770
HeartEWBBA0.85115.03330.151345.4301
BBA-IF10.85164.90000.150739.6087
BBA-IF20.85215.10000.150340.7423
BBA-IF30.85114.83330.151138.9901
EBBA0.85205.06670.150439.0150
SonarEWBBA0.910027.30000.093749.9454
BBA-IF10.909727.03330.093942.1371
BBA-IF20.910027.30000.093744.0751
BBA-IF30.910526.70000.093141.8531
EBBA0.911727.06670.091942.8189
SpectEWBBA0.737910.50000.264346.5456
BBA-IF10.739310.23330.262838.7934
BBA-IF20.739310.20000.262840.4760
BBA-IF30.738610.70000.263639.2009
EBBA0.740710.70000.261538.4262
tic-tac-toeBBA0.84938.86670.159097.9119
BBA-IF10.84938.86670.159086.5623
BBA-IF20.85219.00000.156489.9898
BBA-IF30.85219.00000.156486.5650
EBBA0.85219.00000.156484.4753
VoteBBA0.95195.20000.050946.4947
BBA-IF10.95185.36670.051139.7080
BBA-IF20.95165.10000.051141.6823
BBA-IF30.95135.43330.051640.1654
EBBA0.95135.40000.051638.9119
Table 5. Optimization results obtained by different algorithms under decision tree. (The best values are highlighted in bold).
Table 5. Optimization results obtained by different algorithms under decision tree. (The best values are highlighted in bold).
AlgorithmsAccuracyThe Number of FeaturesFitness Value
BBA0.73127.93330.2683
BDA0.72117.73330.2783
BGWO0.711512.40000.2886
BPSO0.72819.16670.2726
EBBA0.73258.53330.2682
Table 6. Optimization results obtained by different algorithms under random forest. (The best values are highlighted in bold).
Table 6. Optimization results obtained by different algorithms under random forest. (The best values are highlighted in bold).
AlgorithmsAccuracyThe Number of FeaturesFitness Value
BBA0.72478.93330.2766
BDA0.70585.93330.2940
BGWO0.703512.43330.2992
BPSO0.72359.06670.2779
EBBA0.72529.06670.2762
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Feng, J.; Kuang, H.; Zhang, L. EBBA: An Enhanced Binary Bat Algorithm Integrated with Chaos Theory and Lévy Flight for Feature Selection. Future Internet 2022, 14, 178. https://doi.org/10.3390/fi14060178

AMA Style

Feng J, Kuang H, Zhang L. EBBA: An Enhanced Binary Bat Algorithm Integrated with Chaos Theory and Lévy Flight for Feature Selection. Future Internet. 2022; 14(6):178. https://doi.org/10.3390/fi14060178

Chicago/Turabian Style

Feng, Jinghui, Haopeng Kuang, and Lihua Zhang. 2022. "EBBA: An Enhanced Binary Bat Algorithm Integrated with Chaos Theory and Lévy Flight for Feature Selection" Future Internet 14, no. 6: 178. https://doi.org/10.3390/fi14060178

APA Style

Feng, J., Kuang, H., & Zhang, L. (2022). EBBA: An Enhanced Binary Bat Algorithm Integrated with Chaos Theory and Lévy Flight for Feature Selection. Future Internet, 14(6), 178. https://doi.org/10.3390/fi14060178

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop