Next Article in Journal
Thermodynamics of an Empty Box
Next Article in Special Issue
Automatic Analysis of Transverse Musculoskeletal Ultrasound Images Based on the Multi-Task Learning Model
Previous Article in Journal
Impact of the Global Fear Index (COVID-19 Panic) on the S&P Global Indices Associated with Natural Resources, Agribusiness, Energy, Metals, and Mining: Granger Causality and Shannon and Rényi Transfer Entropy
Previous Article in Special Issue
A Parallel Multiobjective PSO Weighted Average Clustering Algorithm Based on Apache Spark
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Binary Bamboo Forest Growth Optimization Algorithm for Feature Selection Problem

1
College of Computer Science and Engineering, Shandong University of Science and Technology, Qingdao 266590, China
2
Department of Information Management, Chaoyang University of Technology, Taichung 41349, Taiwan
3
College of Electronic and Information Engineering, Shandong University of Science and Technology, Qingdao 266590, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(2), 314; https://doi.org/10.3390/e25020314
Submission received: 9 January 2023 / Revised: 27 January 2023 / Accepted: 5 February 2023 / Published: 8 February 2023
(This article belongs to the Special Issue Swarm Intelligence Optimization: Algorithms and Applications)

Abstract

:
Inspired by the bamboo growth process, Chu et al. proposed the Bamboo Forest Growth Optimization (BFGO) algorithm. It incorporates bamboo whip extension and bamboo shoot growth into the optimization process. It can be applied very well to classical engineering problems. However, binary values can only take 0 or 1, and for some binary optimization problems, the standard BFGO is not applicable. This paper firstly proposes a binary version of BFGO, called BBFGO. By analyzing the search space of BFGO under binary conditions, the new curve V-shaped and Taper-shaped transfer function for converting continuous values into binary BFGO is proposed for the first time. A long-mutation strategy with a new mutation approach is presented to solve the algorithmic stagnation problem. Binary BFGO and the long-mutation strategy with a new mutation are tested on 23 benchmark test functions. The experimental results show that binary BFGO achieves better results in solving the optimal values and convergence speed, and the variation strategy can significantly enhance the algorithm’s performance. In terms of application, 12 data sets derived from the UCI machine learning repository are selected for feature-selection implementation and compared with the transfer functions used by BGWO-a, BPSO-TVMS and BQUATRE, which demonstrates binary BFGO algorithm’s potential to explore the attribute space and choose the most significant features for classification issues.

1. Introduction

Nowadays, with the rapid growth of the computer industry, a wide variety of data has been affected. The high speed of development has led to a discontinuous growth in the dimensionality and sample size of the data collected. Managing these data is becoming increasingly difficult. In the early stages of computing, attempts were made to manage these data sets using manual management, but as features in the data set increased, this approach became impractical [1,2]. With further developments, data mining and machine learning application techniques have been developed. Practical applications such as statistical analysis, neural networks and pattern recognition also have surfaced [3,4,5,6]. However, the data collected are often accompanied by high noise levels, mainly caused by the immaturity of the technology used to collect the data and the provenance of the data themselves. There is no doubt that extracting useful content and patterns from such large and noisy data is an extremely challenging task [7].
Feature selection (FS) is an effective method for reducing dimensionality and removing noisy and unreliable data. The aim is to remove unnecessary features from the whole feature set and finally obtain a representative subset [8]. FS is very important and essential for data scientists and machine learning practitioners. A good feature-selection method can simplify models, improve learning accuracy, reduce runtime and help understand the underlying structure of the data, which can significantly influence further improved models and algorithms. A high-quality sample is key to training a classifier. The performance of a classifier is directly influenced by the presence of redundant or irrelevant features in the sample [9].
A realistic data set is usually represented by a collection of data containing plenty of features, not all of which are useful for classification. Redundant, irrelevant features can reduce classification accuracy. As the dimensionality of the data rises and the search space expands, selecting the best subset of features becomes increasingly challenging. In general, the enumeration method cannot solve the problem of finding the optimal subset of features, so some strategies are needed to find the subset of features, and the popular search strategies are global search, heuristic search and random search [10,11,12]. Although existing search techniques have achieved good results in feature selection, there is still a high probability of slipping into a local optimum. Therefore, to solve the feature selection-problem more effectively, a direct and effective search strategy is needed.
The heuristic optimization algorithm is a common optimization method to solve optimization problems. It has a high search power and search speed for NP problems, which can obtain a better solution in polynomial time [13,14]. It solves feature selection by converting successive optimization algorithms into binary versions using transfer functions (tfs).
Heuristic algorithms are inspired by nature, social behavior or the behavior of groups of organisms [15]. It puts forward feasible solutions to optimization problems by imitating natural phenomena and biological behaviors, but the quality of solutions is very different. The original heuristics suffered from the following problems: they rely too much on information about the organization of the algorithm, have low applicability and can easily slip into a local optimum solution. With the development of heuristics, meta-heuristics have emerged that are different from the original heuristics, adding the idea of random search and possessing generality compared to traditional heuristics. Although the meta-heuristic algorithm is improved compared with the original heuristic algorithm, neither is guaranteed to obtain an optimal global solution, and due to the addition of the idea of random search, repeated executions may converge to a globally optimal solution. There are four main categories of meta-heuristic algorithms based on the type of inspiration: evolution-based algorithms, group intelligence-based algorithms, human-based algorithms, physics and chemistry-based algorithms [16]. The main inspiration for evolution-based algorithms comes from the evolutionary law of survival of the fittest (Darwin’s law).
Genetic algorithms (GA) [17,18,19] are the main representatives, such as differential evolution (DE) [20,21,22] and quasi-affine transformation evolution (QUATRE) [23,24,25]. Population intelligence optimization algorithms simulate group intelligence to reach a globally optimal solution. Each group in the algorithm represents a population of organisms that, through the cooperative behaviour of group members, can accomplish tasks that are impossible for individuals. Examples include Particle Swarm Optimization (PSO) [26], Cat Swarm Optimization (CSO) [27], Fish Migration Optimization (FMO) [28], Whale Optimization Algorithm (WOA) [29,30,31], etc. Human behavior, including teaching, social, learning, emotional and managerial behavior, is a major source of inspiration for human-based algorithms. Examples include teaching and learning-based optimization (TLBO) [32], social learning optimization (SLO) [33], social-based algorithms (SBA) [34], etc. Physical rules and chemical reflections in the universe inspire physics and chemistry-based algorithms. Examples include simulated annealing (SA) [35], gravitational local search (GLSA) [36], etc.
Nonetheless, discrete problems are always popular among optimization problems, such as feature selection and shop floor scheduling problems. Continuous optimization algorithms are not suitable for solving such problems, so there is a need to convert continuous optimization algorithms into discrete versions. So far, scholars have proposed many binary versions of algorithms applied to feature selection, while many scholars have also implemented improvements to existing binary algorithms and achieved better results. For example, the classical PSO, GWO and PIO algorithms have been successfully applied to feature selection. Hu et al. improved BGWO by introducing a new transfer function to replace the S-shaped function. Then, a new parametric equation and an improved transfer function were proposed to improve the quality of the solution [37]. Tian et al. analysed BPIO, introduced four new transfer functions along with an improved velocity update equation, and successfully implemented feature selection with better results [38]. Liu et al. devised an improved multi-swarm PSO (MSPSO) to solve the feature-selection problem while combining SVM with F-score methods to improve generalization [39]. However, many metaheuristics have been redesigned without consideration of the problem of sliding into local optima.
Bamboo Forest Growth Optimization (BFGO) is a meta-heuristic algorithm inspired by the bamboo growth process, recently proposed by Chu et al. It is applied to wireless sensor networks (WSNs) and has been effective in reducing energy consumption and improving network performance [40]. This research aims to propose a BFGO with a binary version for the application of discrete optimization problems such as FS. This paper converts the algorithm to a binary version using transfer functions, the better-known ones being the S−type transfer function family and the V−type transfer function family [41]. A novel type of transfer function is also introduced: the Taper-shaped transfer function [42]. The 23 benchmark functions test and compare the performance of different types of transfer functions. The evaluation of the Binary Bamboo Forest Growth Optimization (BBFGO) algorithm against cutting-edge, sophisticated and efficient algorithms shows that the proposed BBFGO possesses optimal or sub-optimal solutions to the problem of finding optimal values. The main contributions are as follows.
  • The first binary bamboo forest growth optimization algorithm (BBFGO) is proposed.
  • Based on a mathematical analysis approach, the first analysis is carried out for the search space of binary BFGO. Based on the results of this analysis, the V−transfer function is stretched in two ways, two new curvature V−transfer functions for binary BFGO are proposed and the new curvature V−transfer function is successfully verified to have better performance in the test function.
  • The long-mutation strategy is introduced to the original BBFGO to avoid solution stagnation, and a new mutation approach is proposed.
  • BBFGO and BBFGO with the new mutation method are compared in test functions with advanced algorithms, and it is confirmed that the long-mutation strategy of the new mutation method improves the performance of BBFGO. Compared with the advanced algorithm, the new mutation strategy leads BBFGO to complete the reversal.
  • BBFGO is applied to feature selection and compared with cutting-edge algorithms, which performs well in low and high dimensional classification accuracy. In particular, it is more competitive on high-dimensional data sets.
The paper is organized as follows: Section 2 introduces bamboo forest growth optimization, Section 3 presents a concrete implementation of binary bamboo forest growth optimization based on mathematical analysis, proposing three classes of transfer functions, BBFGO-S, BBFGO-V, BBFGO-T, and introducing a novel mutation approach to prevent the optimization process from stalling. Section 4 shows the experimental results of the families of BBFGO transfer functions compared to BPSO-TVMS, BGWO-a and BQUATRE and the effect of ABBFGO on the performance improvement. Section 5, ABBFGO-S, ABBFGO-V and ABBFGO-T algorithms are used for feature selection and compared with the three advanced algorithms in Section 4.

2. Bamboo Forest Growth Optimization

FS is an effective method for dimensionality reduction of data, which is widely used and plays an important role in machine learning and pattern recognition. By reducing the dimensionality of the data set, the computational speed of the model is improved. This section focuses on bamboo forest growth optimization. Bamboo is a fast-growing herb and one of the world’s fastest-growing plants. Bamboo has underground rhizomes, also known as bamboo whips, which grow horizontally and produce roots on the nodes called whip roots. Each node has a shoot that has the opportunity to grow into a new whip or bamboo shoot. The new bamboo whip will continue to spread underground, and the bamboo shoot will break through the soil and develop into a bamboo pole, and then gradually develop into a bamboo forest. Figure 1 shows the specific structure of the bamboo. Bamboo whip plays an important role in the overall growth of bamboo forests, and it expands the living area of bamboo and provides nutrients for the growth of bamboo. According to Guihua Jin [43], bamboo has unique growth characteristics compared to other grasses because the tall stems of bamboo grow rapidly within 2–3 months with a slow and slow growth rhythm. This trait may help them adapt to the environment and stand out from the competition to survive. The growth of a bamboo forest can be divided into two stages: (a) bamboo whip extension; (b) bamboo shoot growth. In addition, a bamboo forest can correspond to more than one bamboo whip, and a bamboo whip can only belong to one bamboo.
Recently, Chu and Feng et al. proposed a novel optimization method inspired by the growth behavior of bamboo forests: bamboo forest growth optimization (BFGO). It views the global extension of the bamboo whip as the development phase of the algorithm and the growth of bamboo shoots as the exploration phase of the algorithm, where the shoots emerge through the soil to become bamboo shoots, and the emerged bamboo shoots have only a small probability of growing into bamboo.
1. Extension of the Bamboo Whip
Based on the relationship between bamboo forest and bamboo whip, the concept of clustering is added to the algorithm. While optimizing the algorithm, individuals are grouped by memes, and dynamic adjustment is made between uniform grouping and random grouping. The uniform grouping is based on individual fitness, and the fitness is arranged in descending order. All the individuals in the initial population are arranged from high to low according to the fitness function value to form a sequence, and then the meme group is classified to divide the bamboo forest into multiple meme groups evenly. Random grouping is used when the renewal of the best individuals of the bamboo whip within each group has all stalled and will re-break up the individuals for random assignment. The idea of meme grouping is shown in Figure 2.
The direction of the extension of the bamboo whip underground is influenced by three factors: group cognition, bamboo whips memory and bamboo forest centre, which correspond to global optimal, intra-group optimal and bamboo forest centre, respectively. The formula for the centre position is shown in Equation (1). The formula for the extension direction is shown in Equations (2)–(4).
C ( k ) = 1 n i = 1 n X ( k i )
cos ( α ) = X t · X G | X t | × | X G |
cos ( β ) = X t · X P ( k ) | X t | × | X P ( k ) |
cos ( γ ) = X t · C ( k ) | X t | × | C ( k ) |
where X ( k i ) represents the ith bamboo shoot position on the k bamboo whip, X t is the current bamboo shoot position, X G is the global optimal bamboo shoot position, X p ( k ) is the optimal bamboo shoot position on the k bamboo whip and C ( k ) is the central position of the bamboo forest. Moreover, c o s ( α ) , c o s ( β ) and c o s ( γ ) represent the degree of extension of the current bamboo shoot position to X G , X p ( k ) and C ( k ) , respectively. The formula for the update is shown in Equation (5).
X t + 1 = X G + Q × ( c 1 × X G X t ) × cos ( α ) , r a n d 0.4 X p ( k ) + Q × ( c 1 × X p ( k ) X t ) × cos ( β ) , 0.4 < r a n d 0.8 C ( k ) + Q × ( c 1 × C ( k ) X t ) × cos ( γ ) , e l s e
Q = 2 × ( 1 t T ) ,
where Q is a crucial parameter impacting the step size of the algorithm development and steadily reduces from 2 to 0 as the number of iterations grows, t is the current iteration, and T indicates the maximum number of iterations; c 1 is a random number from 0 to 1. Taking a random number and comparing it with 0.4 and 0.8 to determine the direction of extension of the next generation of solutions ensures the diversity of solutions and enhances the algorithm’s ability to find the best.
2. Shoot Growth of the Bamboo
As we all know, the growth of trees is inevitably affected by many random factors. During the whole growth process, these factors have a large and small impact on them, both individually and comprehensively. At present, it is not possible to accurately determine all of them. Even if they can be measured, the relationship between the factors is also random. Therefore, when describing the tree growth process, the tree-measuring factors are generally regarded as random variables, and the tree growth process is described as a random process. Due to the interference of random factors and the different site conditions, the cumulative growth amount of different bamboo at a specific time t is randomly changed, which shows that the growth process of bamboo is random. Based on the characteristics of bamboo shoot growth, Shi et al. constructed a stochastic process model of bamboo shoot growth by using stochastic process theory and the Sloboda growth equation [44]. There are two stages to the growth process: the slow growth stage and the fast growth stage. Combined with this model, the bamboo shoot growth stages grow as shown in Equation (7).
X ( t ) = X G × e b ω × e ω
The shape of the bamboo growth increment model is shown in Figure 3. The high growth of the bamboo shoot stage is completed in about 55 days, and the growth of the bamboo shoot stage can be divided into 2 stages around the 25th day: 1–25 days is the first stage, in which the growth of bamboo is relatively gentle, 25–55 is the second stage, in which the bamboo shows explosive growth. Moreover, X G represents the maximum bamboo height in a particular growth environment, varying with the growth environment; b is the bamboo measurement factor, a random variable; and ω is the shape parameter of the model, independent of the environmental conditions.
According to the incremental calculation of changes at different times, the calculation equation is shown in Equation (8). In the multi-dimensional case, the result of Equation (8) is a vector.
Δ H = X ( t ) X ( t 1 ) X G C ( k ) + 1 ,
where Δ H represents the relationship between the increment between the two generations and X G and C ( k ) , the denominator represents the distance between X G and C ( k ) , and X ( t ) indicates the cumulative length of the bamboo growth within the t-th generation.
The individual renewal of bamboo shoots at this stage is shown in Equations (9) and (10). In the multi-dimensional case, the result of Equation (10) is a vector.
X t e m p = X t + X D × Δ H , r a n d < 0.5 X t X D × Δ H , e l s e
X D = 1 | X t C ( k ) X G C ( k ) + 1 | ,
where X D represents the ratio relationship between C ( k ) and the distance between X t and X G , which varies to a large extent in the early stages of exploration and stabilizes or even remains constant in the later stages of exploration as X t converges extremely closely to X G and C ( k ) . This results in a more extensive exploration in the early stages and a slow growth towards convergence in the later stages.
In Equation (9) ‘+’ means the distance increases and ‘−’ means the distance decreases, increasing the capacity to search for the optimal solution by expanding the search range and balancing global exploitation and local exploration. The pseudocode of BFGO is shown in Algorithm 1.
Algorithm 1 Pseudocode of BFGO
  1:
Initialize the parameters of BFGO: N (number of search agents), T (maximum iteration), n (number of bamboo shoots) and K (number of bamboo whips)
  2:
Initialize the positions of the search agents
  3:
Calculate the fitness of each search agent
  4:
The search agents are sorted in descending order of fitness and divided into K groups according to the uniform grouping in Figure 2
  5:
Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
  6:
t = 2
  7:
while t < T + 1 do
  8:
   for each bamboo whip do
  9:
     Update C ( k ) using Equation (1)
10:
     Update the X t + 1 using Equations (2)–(6)
11:
     Calculate f ( X t + 1 )
12:
   end for
13:
   Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
14:
   for each bamboo whip do
15:
     Update C ( k ) using Equation (1)
16:
     Update the X t e m p using Equations (7)–(10)
17:
     Calculate f ( X t e m p )
18:
   end for
19:
   Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
20:
    t = t + 1
21:
   Repeat step 4
22:
end while

3. Binary Bamboo Forest Growth Optimization Algorithms

In bamboo forest optimization, bamboo constantly changes its position in space. In some special problems, such as feature selection, the solution is limited to binary 0, 1 values, which inspires a special version of BBFGO. BFGO is a novel algorithm for population evolution by updating positional information through optimal global directions, intra-group optimal directions and central direction guides. BFGO incorporates the idea of clustering to achieve co-competition between multiple groups and has a stronger merit-seeking capability compared to other meta-heuristics. The mechanism for converting continuous BFGO to binary BFGO is explained in Section 3.1. The advanced binary BFGO framework with integrated long-mutations is presented in Section 3.2.

3.1. Binary Bamboo Forest Growth Optimization (BBFGO)

The standard bamboo forest growth optimization algorithm has continuous solutions and can update the equations without restriction constraints, but for feature selection, the search space needs to be set up as a hypercube, which means that the elements of each solution need to be integrated as 0 or 1. The graphical interpretation of BBFGO is shown in Figure 4.

3.1.1. Mathematical Analysis

Under the constraints of the binary condition, the positions of the bamboo whips and shoots cannot be moved arbitrarily in space, so it is necessary to consider the position structures that belong only to the binary BFGO. For the sake of a simple description of the mathematical model, only the one-dimensional case is considered after analyzing the range of values of the individual parameters. From Equation (6), Q ∈ (0, 2), and from Equations (1)–(3), cos ( α ) , cos ( β ) and cos ( γ ) are all ∈ (0, 1), and c 1 is a random number between 0 and 1, so c 1 ∈ (0, 1).
In the bamboo whip extension stage, the globally optimal extension direction in Equation (4) is first analysed. Since X t and X G only take 0 or 1, there are four occurrences of calculating the next generation position, and the value of X t + 1 is calculated as follows.
(1) if X G = 0 and X t = 0
X t + 1 = X G + Q × ( c 1 × X G X t ) × cos ( α ) = 0
(2) if X G = 0 and X t = 1
X t + 1 = X G + Q × ( c 1 × X G X t ) × cos ( α ) = Q × (−1) × cos ( α )
since Q ∈ (0, 2), cos ( α ) ∈ (0, 1), then X t + 1 = (−Q × cos ( α ) ) ∈ (−2, 0).
(3) if X G = 1 and X t = 0
X t + 1 = X G + Q × ( c 1 × X G X t ) × cos ( α ) = 1 + (Q × c 1 × cos ( α ) )
since Q ∈ (0, 2), cos ( α ) ∈ (0, 1), c 1 ∈ (0, 1), then X t + 1 = 1 + (Q × c 1 × cos ( α ) ) ∈ (1, 3).
(4) if X G = 1 and X t = 1
X t + 1 = X G + Q × ( c 1 × X G X t ) × cos ( α ) = 1 + (Q × ( c 1 − 1) × cos ( α ) )
since Q ∈ (0, 2), cos ( α ) ∈ (0, 1), c 1 ∈ (0, 1), then X t + 1 = 1 + (Q × ( c 1 − 1) × cos ( α ) ) ∈ (−1, 1).
From the above analysis, it can be obtained that X t + 1 ∈ (−2, 3). Similarly, when calculating the optimal extension direction within the group, as the constituent elements of the X p ( k ) solution are only 0 or 1, then the analysis method is the same as above, and X t + 1 ∈ (−2, 3) is obtained. When calculating the central extension direction, the formula for C(k) is given by Equation (1), where X ( k i ) is composed of 0 or 1, then the final result of C(k) ∈ (0, 1), which is brought into the analysis process, the final result is X t + 1 ∈ (−2, 3). In summary, it can be concluded that X t + 1 ∈ (−2, 3).
To avoid the search agent missing better solutions due to too large a step, the step size of the exploration in the bamboo shoot growth phase is therefore limited. The accumulation at a specific moment is restricted in Equation (7) to the interval [0, 1].
Since X t takes values only 0 and 1 and Δ H ∈ [0, 1] and X D ∈ (0, 1), in the one-dimensional case there are two occurrences and the value of X t e m p is calculated as follows:
(1) if X t = 0
X t e m p = X t ± X D × Δ H = 0 ± X D × Δ H
since Δ H ∈ [0, 1], X D ∈ (0, 1), then X t e m p ∈ (−1, 1).
(2) if X t = 1
X t e m p = X t ± X D × Δ H = 1 ± X D × Δ H
since Δ H ∈ [0, 1], X D ∈ (0, 1), then X t e m p ∈ (−2, 2).
Through the analysis of the two stages, it can be concluded that X t + 1 ∈ (−2, 3) and X t e m p ∈ (−2, 2). The final results will be used for further discussion of the transfer function.

3.1.2. BBFGO with Transfer Functions

In BFGO, the solutions represented by the bamboo whip and the bamboo shoot are both continuous values, however, FS is a binary optimization problem where the continuous solution needs to be converted to a discrete solution. The transfer function is one of the dominant methods for solving this type of problem. Transfer functions can map continuous values to the interval 0–1 and then update to binary values of 0 or 1 depending on the probability.
In this study, seven tfs are used for this conversion task. Of these seven tfs, two belong to the S-shaped, three to the V-shaped and two to the Taper-shaped classes. The key job of these tfs is to determine the probability of updating the value of an element to 1 or 0. In FS, the solution consists of either 0 or 1. The pseudocode of BBFGO is shown in Algorithm 2. The main three classes of tfs are described as follows:
S-shaped transfer function (S-tf):
The curves of the original sigmoid function and another variant are shown in Figure 5. The transfer vector (S) is calculated according to S-shaped using Equations (11)–(13).
S 1 X m , j t = 1 1 + e X m , j t j = ( 1 , 2 , , d )
S 2 X m , j t = 1 1 + e 10 ( X m , j t 0.5 )
After obtaining X t + 1 by Equation (5), it is updated through Equations (11)–(13). Similarly, X t e m p , which is obtained from Equation (9), is updated by Equations (11)–(13).
X m , j t = 1 rand < S i X m , j t , 0 rand S i X m , j t . i = ( 1 , 2 )
where X m , j t is the element of the dth dimension in the ith solution and S i X m , j t is the probability value of the S-tf based on the mapping of the element of the dth dimension in the ith solution. Whether the final element takes 0 or 1 is determined by Equation (13), with rand being a random value between 0 and 1.
V-shaped transfer function (V-tf):
The curves of the original V-tf and the two variants are shown in Figure 6. The transfer vector (V) is calculated according to V-shaped using Equations (14)–(16):
From the curve V−1 in Figure 6, it can be seen that in the binary BFGO, the bounded maxima of V 1 X m , j t in the interval [−2, 3] are 0.8038, and 0.8669, respectively, indicating that, even when the bounded values of −2 or 3 of the search space of the search agent are reached, there are still probabilities of 0.1962 and 0.1331, respectively, that 1 is not reached. This contradicts our initial aim to make the search agent’s value large when its large probability of its becoming 1 contradicts this. To resolve this situation, a method of stretching the transfer function is adopted. Since the interval is not symmetric, two approaches to stretching are taken.
V 1 X m , j t = | 2 π a t a n ( π 2 X m , j t ) |
V 2 X m , j t = | a t a n ( π 2 X m , j t ) a t a n ( π ) |
V 3 X m , j t = | a t a n ( π 2 X m , j t ) a t a n ( π ) | X m , j t < 0 | a t a n ( π 2 X m , j t ) a t a n ( 3 π 2 ) | X m , j t 0
X m , j t = 1 rand < V i X m , j t , 0 rand V i X m , j t . i = ( 1 , 2 , 3 )
where V i X m , j t is the probability value of the V-tf based on the mapping of the element of the dth dimension in the ith solution. The elements take the value of 0 or 1 as determined by Equation (17).
Taper-shaped transfer function(T-tf):
The T-tf is a novel transfer function, a primary function constructed from a power function. Its uniform formula is shown in the Equation (18).
T ( x ) = | x A | 1 n , x [ A , A ] , n 1
where A is a positive real number and n can determine the curvature of the function.
Since the function curve resembles the taper’s tip, it is called the Taper-shaped transfer function. The T-tf has a beneficial effect on the execution time of the elemental discretization process compared to the S-tf consisting of an exponential function and the V-tf consisting of a trigonometric or inverse trigonometric function.
Since the search space of the T-tf is determined by A in Equation (18), however, the analysis in Section 3 shows that the search space of the binary BFGO is a non-symmetric interval, being [−2, 3]. Therefore two different curvature Taper-shaped transfer functions (T-tfs) are proposed. The curves are shown in Figure 7. The transfer vector (T) is calculated according to Taper-shaped using Equations (19) and (20):
T 1 ( X m , j t ) = | X m , j t | 4 2 4
T 2 X m , j t = | X m , j t | 4 2 4 X m , j t < 0 | X m , j t | 4 3 4 X m , j t 0
X m , j t = 1 rand < T i X m , j t , 0 rand T i X m , j t . i = ( 1 , 2 )
where T i X m , j t is the probability value of the T-tf based on the mapping of the element of the dth dimension in the ith solution. The elements take the value of 0 or 1 as determined by Equation (21).
Algorithm 2 Pseudocode of BBFGO
  1:
Initialize the parameters of BFGO: N (number of search agents), T (maximum iteration), n (number of bamboo shoots) and K (number of bamboo whips)
  2:
Initialize the positions of the search agents
  3:
Calculate the fitness of each search agent
  4:
The search agents are sorted in descending order of fitness and divided into K groups according to the uniform grouping in Figure 2
  5:
Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
  6:
t = 2
  7:
while t < T + 1 do
  8:
   for each bamboo whip do
  9:
     Update C ( k ) using Equation (1)
10:
     Update the X t + 1 using Equations (2)–(6)
11:
     Calculate the position under binary conditions of each search agent by Equations (11)–(21)
12:
     Calculate f ( X t + 1 )
13:
   end for
14:
   Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
15:
   for each bamboo whip do
16:
     Update C ( k ) using Equation (1)
17:
     Update the X t e m p using Equations (7)–(10)
18:
     Calculate the position under binary conditions of each search agent by Equations (11)–(21)
19:
     Calculate f ( X t e m p )
20:
   end for
21:
   Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
22:
    t = t + 1
23:
   Repeat step 4
24:
end while

3.2. Advanced Binary Bamboo Forest Growth Optimization (ABBFGO)

When a solution is good enough, it will attract other search agents, who will quickly converge towards that solution position. However, there are many local optima in the optimization process. Once BBFGO is stuck in a local trap, all search agents are exploited within a narrow region. The group’s diversity for this feature is discarded, and the best solution is not updated for a while. To break out of the local trap, advanced BFGO (ABBFGO) uses a long-mutation strategy to avoid the algorithm entering stagnation.
Long-mutations, similar to variations in genetic algorithms, are added to improve the global search. The mutation is an important component of evolutionary algorithms because it prevents populations from losing diversity and ensures that wide search space is covered. The long-mutation differs from the short-mutation in that the short-mutation randomly selects a dimension of the solution to mutate, whereas the long-mutation mutates every dimension of the solution.
In this, a new mutation strategy is proposed, which uses a strategy influenced by elites to change the position of the bamboo. The elites are divided into historical elites and contemporary elites, where the historical elites consist of the best solutions within the historical group, and the best solutions within the group are saved to the historical elites when the algorithm iterates to update the global optimum, and the contemporary elites consist of the global optimum and the best solutions within the group in the most recent iteration. When the number of iterations exceeds 10 and the global optimum does not change within three generations, the long-mutation strategy is used to try to escape the trap.
In feature selection, 1 indicates that the feature is used and 0 indicates that the feature is not used. Feature selection aims to achieve high accuracy in classification while selecting as few features as possible. So the solution consists of either 0 or 1. The pseudocode of ABBFGO is shown in Algorithm 3. The long-mutation process considers the influence of contemporary and historical elites on mutation, so each elite type produces one solution, and the following strategy is used when mutating each dimension.
X A 1 ( d ) = 0 if rand < d 1 else
where X A 1 ( d ) indicates the element in the d-dimension of the long-mutation generating solution. d is the rate at which the feature is not selected in the d-dimension of the historical elite pool.
X A 2 ( d ) = ϕ i d if ϕ i d = ϕ j d ϕ i d else if r a n d < 0.5 ϕ j d else
where X A 2 ( d ) indicates the element in the d-dimension of the long-mutation generating solution. ϕ i d , ϕ j d is two randomly selected elements in dimension d of the contemporary elite pool whose values are either 0 or 1.
Algorithm 3 Pseudocode of ABBFGO
  1:
Initialize the parameters of BFGO: N (number of search agents), T (maximum iteration), n (number of bamboo shoots) and K (number of bamboo whips)
  2:
Initialize the positions of the search agents
  3:
Calculate the fitness of each search agent
  4:
The search agents are sorted in descending order of fitness and divided into K groups according to the uniform grouping in Figure 2
  5:
Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
  6:
t = 2
  7:
while t < T + 1 do
  8:
   if  t > 10 &&Convergence(t)==Convergence(t-3) then
  9:
     Use the long-mutation strategy to generate new solutions
10:
     If the fitness value of the new solution is greater than the global optimal value, replace it
11:
     Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
12:
   end if
13:
   for i = 1;i ≤ k;i++ do
14:
     if  X P ( k ) not updated then
15:
        P++
16:
     end if
17:
   end for
18:
   if P = k then
19:
     Do Random grouping
20:
   end if
21:
   for each bamboo whip do
22:
     Update C ( k ) using Equation (1)
23:
     Update the X t + 1 using Equations (2)–(6)
24:
     Calculate the position under binary conditions of each search agent by Equations (11)–(21)
25:
     Calculate f ( X t + 1 )
26:
   end for
27:
   Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
28:
   for each bamboo whip do
29:
     Update C ( k ) using Equation (1)
30:
     Update the X t e m p using Equations (7)–(10)
31:
     Calculate the position under binary conditions of each search agent by Equations (11)–(21)
32:
     Calculate f ( X t e m p )
33:
   end for
34:
   Update X G , X P ( k ) , f ( X G ) and f ( X P ( k ) )
35:
   Convergence(t) = f ( X G )
36:
    t = t + 1
37:
   Repeat step 4
38:
end while

4. Experimental Results and Analysis

In this section, the simulation experiment process of BBFGO is mainly introduced. The main purpose is to reveal the effects of various transfer functions and stretched transfer functions on the performance of binary BFGO through 23 benchmark test functions. Table 1 describes the basic information of these 23 benchmark functions. Single-peak functions (1–7), multi-peak functions (8–13) and fixed-dimensional functions (14–23) are examples of reference functions. Furthermore, opt is the minimum value that the test function can reach in theory; parameter space is the search space of the search agent; Dim is the dimension of the function.
To validate the results, BBFGO is compared with the original V-tf, the V-tf after two different ways of stretching and the T-tf. Table 2 shows details of the comparison methods. Table 3 shows information about the parameters of the optimization algorithm recommended and used in solving the examples. Each algorithm has a population size of 30. The maximum iteration is 500 times, and the experiment is run 30 times.

4.1. Experimental Analysis of the Transfer Functions and ABBFGO

Table 4 shows the results using different transfer function family algorithms and ABBFMO. A total of 23 benchmark test functions can achieve the minimum values shown in the Opt column of Table 1 for a given continuous solution space, however when testing the BBFGO family of algorithms, the solution space is hypercubic, meaning that the solutions all consist of binary values 0 and 1, so the optimum values that can be achieved under these conditions are not the same. The experimental data in Table 2 are marked in red if an algorithm achieves the minimum of the two-value condition or performs best in the test function. Experimental data where ABBFGO is more effective than BBFGO are marked in blue.
The single-peak test function without local traps is used to test the convergence performance of the different algorithms. If an algorithm performs well in the single-peak test function, it can be shown to have a strong convergence exploitation ability. The first seven functions in Table 4 show that the traditional sigmoid transformation function performs the least well in terms of convergence capability. In F 1 , F 2 , F 3 and F 6 , all except BBFGO reach the theoretical minimum in the binary condition. For F 4 , BBFGO-S and BBFGO-V1 reach the theoretical minimum, and the two stretched V2 and V3 do not outperform V1 in terms of optimal values, but this difference is not significant. In F 5 , all algorithms do not reach the theoretical optimum and are some distance away from the optimum. BBFGO obtains the worst result at 124.4, but V2 and V3 are stronger than V1 in terms of effect. T1 is closest to the optimum at 1.9333. In F 7 , the best result is obtained by the stretched V2, which shows that stretching the transfer function to the interval [0, 1] helps to improve the BFGO-V convergence development capability.
As the multi-peak function has many local optima, it can be used to test the performance of different algorithms to jump out of the local trap. For the F 8 F 13 test functions, BBFGO performs the worst, and other algorithms achieve theoretical minima in F 9 , F 10 , F 11 and F 12 . It indicates that BBFGO* has better performance in jumping out of local traps. For F 8 and F 13 , it can be seen that the stretched V2 and V3 outperform V1. It shows that stretching V-tf also decreases the likelihood of the binary BFGO slipping into a local optimum. BBFGO* performs very well in all the fixed-dimensional functions, except T2, which does not search for a theoretical minimum in F 23 . It illustrates two things: firstly, it reflects the excellent performance of BBFGO*. The second is that there is a limit to how much the transfer function can improve BBFGO performance, especially in functions with fewer local minima or lower dimensionality.
ABBFGO beats BBFGO in 12 out of 23 test functions, which are focused on single and multi-peak test functions. The long-mutation method using elite learning can help change the search space in case the algorithm becomes stuck. ABBFGO outperforms BBFGO significantly, demonstrating the effectiveness of the new mutation method.

4.2. Experimental Results for Cutting-Edge Algorithms

BQUATRE is a novel binary algorithm inspired by matrix iteration, Binary QUATRE (BQUATRE) is a binary version that can be used to solve binary application problems [46]. Binary Grey Wolf Optimizer (BGWO) extends the application of the GWO algorithm and is applied to binary optimization issues. In BGWO-a, new a-parameters are used to control the values of A and D, the ability to balance global and local search and the use of a new transfer function to improve the quality of the solution [37]. BPSO-TVMS introduces a new time-varying mirror S-shaped transfer function to enhance global exploration and local exploitation in the algorithm [45]. In Table 5, BBFGO and ABBFGO are compared with these novel improved algorithms, where red font indicates that BBFGO and ABBFGO were defeated, green font indicates that BBFGO is defeated but not ABBFGO, and blue font indicates that both BBFGO and ABBFGO won.
As can be seen in Table 5, in the single-peak test function, all the data colors are green except for F 4 , where BBFGO, which had a poor effect, reverses under the long-mutation strategy, further demonstrating the effectiveness of the strategy. The effect of the new mutation in the multi-peak test function is also unquestionable, with only the F 10 function failing to beat BQUATRE. BBFGO and ABBFGO outperform BPSO-TVMS in F 8 and F 13 . ABBFGO takes advantage of the fact that, if the strategy falls into a local trap in the update iteration, it will take the direction of the overall elite to learn to jump out of that trap and search for another better value. Although it helps BBFGO improve its capacity to escape from local optima, this strategy has a limit to the improvement. For example, in F 10 and F 11 , it does not help BBFGO to achieve the search for the theoretical minimum. It also illustrates the importance of the transfer function in another way. It is the reason why this paper focuses on both the transfer function and the strategy at the same time. All of the above algorithms perform well in the fixed-dimensional test functions, with only BGWO-a performing poorly in F 19 and F 20 .

5. Apply to Feature Selection

Dealing with enormous data owing to their size is extremely challenging due to the abundance of noise and unnecessary aspects in data mining, and which features apply to a learning algorithm is unknown, it is essential to pick the pertinent features from the set of features that the learning algorithm will find useful. As a result, the data set’s characteristics must be reduced. The majority of studies focus on techniques with great accuracy and few characteristics. This section uses the wrapper approach to feature selection.

5.1. Datasets Description

The data sets were taken from the UCI machine learning repository [47], and the details of these data sets are shown in Table 6. Table 6 shows the main characteristics of these data sets in terms of the number of features, number of instances and number of classes. The selected data sets were categorized by dimensionality as low-dimensional and high-dimensional data sets, which varied in the number of features and instances and could be used as a sample of the many problems tested. The more novel data set details are as follows: the data set of Turkish Music Emotion in Table 6 is designed as a discrete model, and there are four classes in the data set: happy, sad, angry and relaxed; the data set of LSVT Voice Rehabilitation includes standard perturbation analysis methods, wavelet-based features, fundamental frequency-based features and tools used to mine nonlinear time-series; the data set of Musk (Version 1) describes a set of 92 molecules of which 47 are judged by human experts to be musks, and the remaining 45 molecules are judged to be non-musks. The goal is to learn to predict whether new molecules will be musks or non-musks; the data set of Dermatology contains 34 attributes, 33 of which are linear-valued and one of them is nominal, the diseases in this group are psoriasis, seboreic dermatitis, lichen planus, pityriasis rosea, chronic dermatitis and pityriasis rubra pilaris; instances in the data set of Ionosphere are described by two attributes per pulse number, corresponding to the complex values returned by the function resulting from the complex electromagnetic signal.

5.2. Simulation Results

Several cutting-edge algorithms are chosen for comparison tests with the proposed binary BFGO algorithms to confirm the performance in various dimensions, and deeply examine the application. By examining the results of the experiments, such as the accuracy and the number of features selected, and fitness function values for specific evaluation criteria, it is possible to compare the merits of the BBFGO* with other algorithms. The benefits of the proposed binary BFGO algorithm and other algorithms may be compared by examining several outcomes from the feature-selection experiment, such as accuracy, number of feature selections and fitness function values for specific evaluation criteria.

5.2.1. KNN and K-Fold Validation

A fundamental and straightforward machine learning method called the K-Nearest Neighbor (KNN) algorithm classifies data by calculating the separation between various eigenvalues. In KNN classification, a classification population is produced from an input learning instance. The category of an object’s neighbours determines its classification. The category given to the object is determined by the K nearest neighbours’ most common classification (K is a positive number, typically smaller). The calculation method is shown in Equation (23).
D 1 ( x , x ) = k = 1 m | x ( k ) x ( k ) | l 1 l
where l has two values, 1 and 2. When l = 1 D 1 ( x , x ) denotes the Manhattan distance and when l = 2 D 2 ( x , x ) denotes the Euclidean distance. x and x are the two input instances for calculating the distance.
In machine learning modelling, to reduce the probability of overfitting problems without adjusting the model parameters in the test data. The original data set is randomly divided into K parts for K-fold cross validation. One of the K parts is utilized as test data, while the remaining K-1 parts are used as training data. The experiment is run K times, and in the end, the average value of the K experimental outcomes is calculated.

5.2.2. Evaluation Criteria

There are many metrics to judge the merits of an algorithm in feature selection, for example, classification accuracy and the number of feature subsets. However, if only classification accuracy is chosen as an evaluation metric, there is no guarantee that the number of subsets is small, so one influence cannot be considered alone. Suitable evaluation metrics need to be constructed to reconcile the balance of factors. In the simulations, the following criteria were used for evaluation.
Fitness = α × K-Fold ( error ) + β × | NoSF | | NoAF |
where K-Fold ( error ) is the classification error after completing cross-validation. NoSF is the subset feature after feature selection and NoAF is the number of features for the data set. α and β balancing the classification accuracy and the number of subsets, with α being 0.99 and β being 0.01.

5.2.3. Result Analysis

In the analysis of the results of the binary BFGO algorithm, ABBFGO, ABBFGO-S, ABBFGO-V2, ABBFGO-V3, ABBFGO-T1, ABBFGO-T2, BPSO-TVMS, BQUATRE and BGWO-a were compared. They were run 15 times on each selected data set, with 100 iterations each. The population size in each population is 30. The value of K-Fold parameter in cross validation is 10. The K value in KNN is 5, and the error rates corresponding to individuals were calculated using the five-nearest neighbour approach.
As can be seen from Table 7, ABBFGO and ABBFGO-V3 beat all other algorithms once, ABBFGO-V2 beat all other algorithms six times, and ABBFGO-T2 beat all other algorithms four times in completing the classification correctly. ABBFGO-V2 is also at a good level for feature subsets and meets the feature-selection requirements. In Table 8, ABBFGO-V2 has an overall ranking of 20. Collectively ABBFGO-V2 performs best, with the best classification accuracy and feature subsets. A comparison of the classification accuracy of the Advanced binary BFGO algorithm family (ABBFGO*) with that of BPSO-TVMS, BQUATRE, and BGWO-a shows that the difference between the two is not significant on some low-dimensional data sets, with the former only slightly ahead of the latter, e.g., ABBFGO, the best performer in Cancer, is 0.0017 ahead of BGWO-a. It is partly due to the fact that there are only two categories in Cancer, or it may be that the small dimensionality of the data set results in a transfer function with a mutation strategy that does not give the better performance of binary BFGO. Glass has the same number of features as Cancer, but Glass has six categories, resulting in a difference in the accuracy of 0.0138 between ABBFGO-T1 and BPSO-TVMS. However, the difference is more pronounced in the high-dimensional data set, where ABBFGO* is superior, with the best ABBFGO-V2 being 0.0516 ahead of BGWO-a in the Turkish Music Emotion, as well as in the Musk (Version 1), Dnatest, LSVT Voice Rehabilitation and Sonar data sets. The differences were also more pronounced. For the tapered-shaped transfer function, it can be seen in Table 7 that T2 with lower curvature has more advantages than T1 on the high-dimensional data sets of Turkish Music Emotion, Musk (Version 1), Dnatest and LSVT Voice Rehabilitation. From Table 9, it is known that the P-value of the Feldman test is greater than 5% in all 12 data sets, so it can be concluded that there is no significant difference between the algorithms, and the data are considered plausible. Based on the above analysis, it is believed that the transfer function and the long-mutation strategy of the new mutation mode give the binary BFGO stronger performance, which makes it more competitive compared with other advanced algorithms in the high-dimensional multi-type data sets.

6. Conclusions

The bamboo forest growth optimization algorithm, inspired by the growth process of bamboo forests, successfully solves many optimization problems. This paper focuses on the analysis of tfs and mutation strategies. Based on the analysed search space and the characteristics of the transfer functions, two different curvatures, V-tfs and T-tfs, are proposed. To avoid the stagnation of the binary BFGO algorithm, the long-mutation strategy with a novel mutation approach is introduced. The newly constructed tfs and the new mutation strategy are tested in 23 benchmark test functions. The experiments show that the newly constructed transfer function has better performance and that the binary BFGO with the long-mutation strategy with a novel mutation has a significant advantage in solving the optimization problem. Feature selection is an important optimization problem and ABBFGO, ABBFGO-S, ABBFGO-V2, ABBFGO-V3, ABBFGO-T1, and ABBFGO-T2 are selected to complete the feature selection and compared with three cutting-edge algorithms, BPSO-TVMS, BQUATRE and BGWO-a. The experiments show that the long-mutation strategy of the transfer function with the new mutation method gives a stronger performance of binary BFGO, and the extent of this improvement is particularly striking on high-dimensional data sets. As this is the first application of BFGO to the discrete domain, much about the capabilities of BFGO has yet to be fully explored and it has more room for development.

Author Contributions

Conceptualization, L.Y.; methodology, P.H.; validation, B.Y.; formal analysis, H.Y.; investigation, L.Y.; resources, S.-C.C.; data curation, P.H.; writing—original draft, L.Y.; writing—review and editing, J.-S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data involved in this study are all public data, which can be found here: http://archive.ics.uci.edu/ml/index.php (accessed on 5 August 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tang, J.; Alelyani, S.; Liu, H. Feature Selection for Classification: A Review. Data Classification: Algorithms and Applications. 2014, p. 37. Available online: https://www.cvs.edu.in/upload/feature_selection_for_classification.pdf (accessed on 8 September 2022).
  2. Abualigah, L.; Diabat, A. Chaotic binary group search optimizer for feature selection. Expert Syst. Appl. 2022, 192, 116368. [Google Scholar]
  3. Yang, L.; Xu, Z. Feature extraction by PCA and diagnosis of breast tumors using SVM with DE-based parameter tuning. Int. J. Mach. Learn. Cybern. 2019, 10, 591–601. [Google Scholar] [CrossRef]
  4. Zeng, A.; Li, T.; Liu, D.; Zhang, J.; Chen, H. A fuzzy rough set approach for incremental feature selection on hybrid information systems. Fuzzy Sets Syst. 2015, 258, 39–60. [Google Scholar]
  5. Li, G.; Zhao, J.; Murray, V.; Song, C.; Zhang, L. Gap analysis on open data interconnectivity for disaster risk research. Geo-Spat. Inf. Sci. 2019, 22, 45–58. [Google Scholar]
  6. Abualigah, L.M.; Khader, A.T.; Hanandeh, E.S. A combination of objective functions and hybrid krill herd algorithm for text document clustering analysis. Eng. Appl. Artif. Intell. 2018, 73, 111–125. [Google Scholar]
  7. Arafat, H.; Elawady, R.M.; Barakat, S.; Elrashidy, N.M. Different feature selection for sentiment classification. Int. J. Inf. Sci. Intell. Syst. 2014, 1, 137–150. [Google Scholar]
  8. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 1–45. [Google Scholar]
  9. Chandrashekar, G.; Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 2014, 40, 16–28. [Google Scholar] [CrossRef]
  10. Dash, M.; Liu, H. Feature selection for classification. Intell. Data Anal. 1997, 1, 131–156. [Google Scholar] [CrossRef]
  11. Liu, Y.; Tang, F.; Zeng, Z. Feature selection based on dependency margin. IEEE Trans. Cybern. 2014, 45, 1209–1221. [Google Scholar] [CrossRef]
  12. Liu, H.; Yu, L. Toward integrating feature selection algorithms for classification and clustering. IEEE Trans. Knowl. Data Eng. 2005, 17, 491–502. [Google Scholar]
  13. Sharma, M.; Kaur, P. A comprehensive analysis of nature-inspired meta-heuristic techniques for feature selection problem. Arch. Comput. Methods Eng. 2021, 28, 1103–1127. [Google Scholar]
  14. SS, V.C.; HS, A. Nature inspired meta heuristic algorithms for optimization problems. Computing 2022, 104, 251–269. [Google Scholar] [CrossRef]
  15. Beheshti, Z.; Shamsuddin, S.M.H. A review of population-based meta-heuristic algorithms. Int. J. Adv. Soft Comput. Appl 2013, 5, 1–35. [Google Scholar]
  16. Osman, I.H.; Kelly, J.P. Meta-heuristics: An overview. In Meta-Heuristics; Springer: Berlin/Heidelberg, Germany, 1996; pp. 1–21. [Google Scholar]
  17. Holland, J.H. Genetic algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  18. Sayed, S.; Nassef, M.; Badr, A.; Farag, I. A nested genetic algorithm for feature selection in high-dimensional cancer microarray datasets. Expert Syst. Appl. 2019, 121, 233–243. [Google Scholar] [CrossRef]
  19. Mellit, A.; Kalogirou, S.A.; Drif, M. Application of neural networks and genetic algorithms for sizing of photovoltaic systems. Renew. Energy 2010, 35, 2881–2893. [Google Scholar] [CrossRef]
  20. Ilonen, J.; Kamarainen, J.K.; Lampinen, J. Differential evolution training algorithm for feed-forward neural networks. Neural Process. Lett. 2003, 17, 93–105. [Google Scholar] [CrossRef]
  21. Hancer, E.; Xue, B.; Zhang, M. Differential evolution for filter feature selection based on information theory and feature ranking. Knowl.-Based Syst. 2018, 140, 103–119. [Google Scholar]
  22. Zhang, Y.; Gong, D.W.; Gao, X.Z.; Tian, T.; Sun, X.Y. Binary differential evolution with self-learning for multi-objective feature selection. Inf. Sci. 2020, 507, 67–85. [Google Scholar] [CrossRef]
  23. Pan, J.S.; Meng, Z.; Xu, H.; Li, X. QUasi-Affine TRansformation Evolution (QUATRE) algorithm: A new simple and accurate structure for global optimization. In Proceedings of the International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems, Morioka, Japan, 2–4 August 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 657–667. [Google Scholar]
  24. Meng, Z.; Pan, J.S.; Xu, H. QUasi-Affine TRansformation Evolutionary (QUATRE) algorithm: A cooperative swarm based algorithm for global optimization. Knowl.-Based Syst. 2016, 109, 104–121. [Google Scholar]
  25. Liu, N.; Pan, J.S.; Nguyen, T.T. A bi-population QUasi-Affine TRansformation Evolution algorithm for global optimization and its application to dynamic deployment in wireless sensor networks. EURASIP J. Wirel. Commun. Netw. 2019, 2019, 175. [Google Scholar]
  26. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  27. Chu, S.C.; Tsai, P.W.; Pan, J.S. Cat swarm optimization. In Proceedings of the Pacific Rim International Conference on Artificial Intelligence, Guilin, China, 7–11 August 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 854–858. [Google Scholar]
  28. Pan, J.S.; Tsai, P.W.; Liao, Y.B. Fish migration optimization based on the fishy biology. In Proceedings of the 2010 IEEE Fourth International Conference on Genetic and Evolutionary Computing, Shenzhen, China, 13–15 December 2010; pp. 783–786. [Google Scholar]
  29. Xing, J.; Zhao, H.; Chen, H.; Deng, R.; Xiao, L. Boosting Whale Optimizer with Quasi-Oppositional Learning and Gaussian Barebone for Feature Selection and COVID-19 Image Segmentation. J. Bionic Eng. 2022, 1–22. [Google Scholar] [CrossRef]
  30. Jiang, F.; Wang, L.; Bai, L. An improved whale algorithm and its application in truss optimization. J. Bionic Eng. 2021, 18, 721–732. [Google Scholar] [CrossRef]
  31. Fang, L.; Liang, X. A Novel Method Based on Nonlinear Binary Grasshopper Whale Optimization Algorithm for Feature Selection. J. Bionic Eng. 2022, 20, 237–252. [Google Scholar] [CrossRef]
  32. Rao, R.V.; Savsani, V.J.; Vakharia, D. Teaching–learning-based optimization: A novel method for constrained mechanical design optimization problems. Comput.-Aided Des. 2011, 43, 303–315. [Google Scholar]
  33. Liu, Z.Z.; Chu, D.H.; Song, C.; Xue, X.; Lu, B.Y. Social learning optimization (SLO) algorithm paradigm and its application in QoS-aware cloud service composition. Inf. Sci. 2016, 326, 315–333. [Google Scholar]
  34. Ramezani, F.; Lotfi, S. Social-based algorithm (SBA). Appl. Soft Comput. 2013, 13, 2837–2856. [Google Scholar]
  35. Kirkpatrick, S.; Gelatt, C.D., Jr.; Vecchi, M.P. Optimization by simulated annealing. Science 1983, 220, 671–680. [Google Scholar]
  36. Webster, B.; Bernhard, P.J. A Local Search Optimization Algorithm Based on Natural Principles of Gravitation. Technical Report. 2003. Available online: http://hdl.handle.net/11141/117 (accessed on 8 September 2022).
  37. Hu, P.; Pan, J.S.; Chu, S.C. Improved binary grey wolf optimizer and its application for feature selection. Knowl.-Based Syst. 2020, 195, 105746. [Google Scholar]
  38. Pan, J.S.; Tian, A.Q.; Chu, S.C.; Li, J.B. Improved binary pigeon-inspired optimization and its application for feature selection. Appl. Intell. 2021, 51, 8661–8679. [Google Scholar]
  39. Liu, Y.; Wang, G.; Chen, H.; Dong, H.; Zhu, X.; Wang, S. An improved particle swarm optimization for feature selection. J. Bionic Eng. 2011, 8, 191–200. [Google Scholar] [CrossRef]
  40. Feng, Q.; Chu, S.C.; Pan, J.S.; Wu, J.; Pan, T.S. Energy-Efficient Clustering Mechanism of Routing Protocol for Heterogeneous Wireless Sensor Network Based on Bamboo Forest Growth Optimizer. Entropy 2022, 24, 980. [Google Scholar]
  41. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar]
  42. He, Y.; Zhang, F.; Mirjalili, S.; Zhang, T. Novel binary differential evolution algorithm based on Taper-shaped transfer functions for binary optimization problems. Swarm Evol. Comput. 2022, 69, 101022. [Google Scholar]
  43. Jin, G.; Ma, P.F.; Wu, X.; Gu, L.; Long, M.; Zhang, C.; Li, D.Z. New Genes Interacted with Recent Whole-Genome Duplicates in the Fast Stem Growth of Bamboos. Mol. Biol. Evol. 2021, 38, 5752–5768. [Google Scholar] [CrossRef]
  44. Shi, Y.; Liu, E.; Zhou, G.; Shen, Z.; Yu, S. Bamboo shoot growth model based on the stochastic process and its application. Sci. Silvae Sin. 2013, 49, 89–93. [Google Scholar]
  45. Beheshti, Z. A time-varying mirrored S-shaped transfer function for binary particle swarm optimization. Inf. Sci. 2020, 512, 1503–1542. [Google Scholar] [CrossRef]
  46. Liu, F.F.; Chu, S.C.; Wang, X.; Pan, J.S. A Novel Binary QUasi-Affine TRansformation Evolution (QUATRE) Algorithm and Its Application for Feature Selection. In Proceedings of the Advances in Intelligent Systems and Computing, Hangzhou, China, 29–31 May 2021; Springer Nature: Singapore, 2022; pp. 305–315. [Google Scholar]
  47. Asuncion, A.; Newman, D. UCI Machine Learning Repository. 2007. Available online: https://www.semanticscholar.org/paper/%5B7%5D-A.-Asuncion-and-D.-J.-Newman.-UCI-Machine-Aggarwal-Han/ea2be4c9913e781e7930cc2d4a0b2021a6a91a44 (accessed on 8 September 2022).
Figure 1. The structure of bamboo.
Figure 1. The structure of bamboo.
Entropy 25 00314 g001
Figure 2. The idea of meme grouping.
Figure 2. The idea of meme grouping.
Entropy 25 00314 g002
Figure 3. Growth curve of bamboo shoots.
Figure 3. Growth curve of bamboo shoots.
Entropy 25 00314 g003
Figure 4. The graphical interpretation of BBFGO.
Figure 4. The graphical interpretation of BBFGO.
Entropy 25 00314 g004
Figure 5. The curves of S-shaped transfer functions.
Figure 5. The curves of S-shaped transfer functions.
Entropy 25 00314 g005
Figure 6. The curves of V-shaped transfer functions.
Figure 6. The curves of V-shaped transfer functions.
Entropy 25 00314 g006
Figure 7. The curves of Taper-shaped transfer functions.
Figure 7. The curves of Taper-shaped transfer functions.
Entropy 25 00314 g007
Table 1. The details of the benchmark functions.
Table 1. The details of the benchmark functions.
NumNameParameter SpaceDimOpt
F 1 Sphere[−100, 100]300
F 2 Schwefel’s function 2.21[−10, 10]300
F 3 Schwefel’s function 1.2[−100, 100]300
F 4 Schwefel’s function 2.221[−100, 100]300
F 5 Rosenbrock[−30, 30]300
F 6 Step[−100, 100]300
F 7 Dejong’s noisy[−1.28, 1.28]300
F 8 Schwefel[−500, 500]30−12,569
F 9 Rastringin[−5.12, 5.12]300
F 10 Ackley[−32, 32]300
F 11 Griewank[−600, 600]300
F 12 Generalized penalized 1[−50, 50]300
F 13 Generalized penalized 2[−50, 50]300
F 14 Fifth of Dejong[−65, 65]21
F 15 Kowalik[−5, 5]40.0003
F 16 Six-hump camel back[−5, 5]2−1.0316
F 17 Branins[−5, 5]20.398
F 18 Goldstein–Price[−2, 2]23
F 19 Hartman 1[1, 3]3−3.86
F 20 Hartman 2[0, 1]6−3.32
F 21 Shekel 1[0, 10]4−10.1532
F 22 Shekel 2[0, 10]4−10.4028
F 23 Shekel 3[0, 10]4−10.5363
Table 2. The details of the comparison methods.
Table 2. The details of the comparison methods.
AlgorithmTransfer Function
BBFGOTransfer Function from Equation (11)
BBFGO-STransfer Function from Equation (12)
BBFGO-V1Transfer Function from Equation (14)
BBFGO-V2Transfer Function from Equation (15)
BBFGO-V3Transfer Function from Equation (16)
BBFGO-T1Transfer Function from Equation (18)
BBFGO-T2Transfer Function from Equation (19)
BGWOTransfer Function from [37]
BPSO-STransfer Function from [45]
BQUATRETransfer Function from [46]
Table 3. Parameter settings of the algorithms used for comparison in the current study.
Table 3. Parameter settings of the algorithms used for comparison in the current study.
AlgorithmParameterValues
BFGO(BBFGO)Number of individuals30
Maximum number of iterations500
Number of bamboo shoots6
Number of bamboo whips5
GWO(BGWO)Number of wolves30
Maximum number of iterations500
The a parameter2 * it/MAX_IT
PSO(BPSO-S)Number of particles30
Maximum number of iterations500
Inertia weight w m a x 0.9
Inertia weight w m i n 0.2
C 1 2
C 2 2
V m i n , V m a x [−6, 6]
QUATRE(BQUATRE)Number of individuals30
Maximum number of iterations500
Matrix control factor0.7
Table 4. The statistical results of the BBFGO* family algorithms.
Table 4. The statistical results of the BBFGO* family algorithms.
FunctionBBFGO-SBBFGO-V1BBFGO-V2BBFGO-V3BBFGO-T1BBFGO-T2BBFGOABBFGO
AVGSTDAVGSTDAVGSTDAVGSTDAVGSTDAVGSTDAVGSTDAVGSTD
F 1 0000000000004.46670.973200
F 2 0000000000004.96670.850300
F 3 000000000000207.133356.89050.10.5477
F 4 00000.10.30510.03330.1826100.93330.25371010
F 5 29029019.333313.904428.03335.29471.93337.357514.933331.1392124.482.043100
F 6 7.507.507.507.507.507.5017.63331.56987.50
F 7 3.68 × 10 5 3.75 × 10 5 3.39 × 10 5 3.10 × 10 5 3.23 × 10 5 2.89 × 10 5 3.42 × 10 5 3.13 × 10 5 3.45 × 10 5 3.63 × 10 5 3.56 × 10 5 4.13 × 10 5 55.93779.57404.15 × 10 4 3.60 × 10 4
F 8 −25.24411.08 × 10 14 −21.06480.84099−25.24411.08 × 10 14 −24.65510.54802−25.24411.08 × 10 14 −25.13192.91 × 10 1 −24.62710.4908−25.24411.08 × 10 14
F 9 0000000000004.86670.819300
F 10 8.88 × 10 16 08.88 × 10 16 08.88 × 10 16 08.88 × 10 16 08.88 × 10 16 08.88 × 10 16 01.56650.13830.02390.1309
F 11 0000000000000.15330.02620.00170.0066
F 12 1.66901.13 × 10 15 1.66901.13 × 10 15 1.66901.13 × 10 15 1.66901.13 × 10 15 1.66901.13 × 10 15 1.66901.13 × 10 15 2.42230.13661.66901.13 × 10 15
F 13 1.35 × 10 32 5.57 × 10 48 0.48670.15481.35 × 10 32 5.57 × 10 48 0.05670.05681.35 × 10 32 5.57 × 10 48 0.04004.98 × 10 2 0.06670.05471.35 × 10 32 5.57 × 10 48
F 14 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15
F 15 0.148400.148400.148400.148400.148400.148400.148400.14840
F 16 0000000000000000
F 17 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14
F 18 60006000600060006000600060006000
F 19 −0.33481.69 × 10 16 −0.33481.69 × 10 16 −0.33481.69 × 10 16 −0.33481.69 × 10 16 −0.33481.69 × 10 16 −0.33481.69 × 10 16 −0.33481.69 × 10 16 −0.33481.69 × 10 16
F 20 −0.16572.82 × 10 17 −0.16572.82 × 10 17 −0.16572.82 × 10 17 −0.16572.82 × 10 17 −0.16572.82 × 10 17 −0.16572.82 × 10 17 −0.16572.82 × 10 17 −0.16572.82 × 10 17
F 21 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15
F 22 −5.08770−5.08770−5.08770−5.08770−5.08770−5.08770−5.08770−5.08770
F 23 −5.12852.71 × 10 15 −5.12852.71 × 10 15 −5.12852.71 × 10 15 −5.12852.71 × 10 15 −5.12852.71 × 10 15 −4.98917.63 × 10 1 −5.12852.71 × 10 15 −5.12852.71 × 10 15
Table 5. The statistical results of BQUATRE, BGWO-a and BPSO-TVMS.
Table 5. The statistical results of BQUATRE, BGWO-a and BPSO-TVMS.
FunctionBQUATREBGWO-aBPSO-TVMS
AVGSTDAVGSTDAVGSTD
F 1 002.83331.20581.16670.5921
F 2 0.03330.18263.16671.55551.43330.6789
F 3 0.46671.105983.766767.517417.133313.9599
F 4 101010
F 5 000093.566790.9607
F 6 7.5013.76672.558710.43331.4606
F 7 2.60013.654039.033421.05579.13074.5008
F 8 −25.24411.08 × 10 14 −25.24411.08 × 10 14 −24.01000.4808
F 9 003.51.33261.43330.5040
F 10 8.88 × 10 16 01.19640.36190.80400.1349
F 11 0.01440.01870.13900.05720.02930.0154
F 12 1.67251.33 × 10 2 2.14090.21061.86530.0981
F 13 1.35 × 10 32 5.57 × 10 48 1.35 × 10 32 5.57 × 10 48 0.15000.0509
F 14 12.67053.61 × 10 15 12.67053.61 × 10 15 12.67053.61 × 10 15
F 15 0.148400.148400.14840
F 16 000000
F 17 27.70291.81 × 10 14 27.70291.81 × 10 14 27.70291.81 × 10 14
F 18 600060006000
F 19 −0.33481.69 × 10 16 −0.33370.0063−0.33481.69 × 10 16
F 20 −0.16572.82 × 10 17 −0.13970.0479−0.16572.82 × 10 17
F 21 −5.05524.52 × 10 15 −5.05524.52 × 10 15 −5.05524.52 × 10 15
F 22 −5.08770−5.08770−5.08770
F 23 −5.12852.71 × 10 15 −5.12852.71 × 10 15 −5.12852.71 × 10 15
Table 6. The details of the testing data sets.
Table 6. The details of the testing data sets.
DatasetNo. of FeaturesNo. of InstancesNo. of Classes
Turkish Music Emotion504004
Musk (Version 1)1664762
Cancer96832
Dermatology343666
Dnatest18011863
German2410004
Glass92146
Heartstatlog132702
Ionosphere343512
LSVT Voice Rehabilitation3101262
Sonar602082
WDBC305692
Table 7. The accuracy and number of the compared algorithms.
Table 7. The accuracy and number of the compared algorithms.
DatasetABBFGOABBFGO-SABBFGO-V2ABBFGO-V3ABBFGO-T1ABBFGO-T2BPSO-TVMSBQUATREBGWO-a
AccNumAccNumAccNumAccNumAccNumAccNumAccNumAccNumAccNum
Turkish Music Emotion0.761725.670.774811.800.790217.530.787916.730.778320.400.788115.870.709822.330.700538.270.738627.40
Cancer0.97875.670.97754.800.97664.270.97664.330.97684.470.97704.600.97665.400.97206.070.97705.80
Musk (Version 1)0.914985.930.931846.270.942447.200.933143.530.928658.000.942049.330.878681.730.8627136.670.907386.87
Dermatology0.990419.130.984715.600.990418.270.990918.000.991220.130.991019.530.980219.400.978925.870.985421.53
Dnatest0.854891.070.904127.530.909216.800.907417.670.892439.330.905720.530.819185.400.8044132.800.856490.47
German0.507612.930.50386.800.51219.600.50707.330.513412.670.510710.130.488611.070.475216.530.498013.53
Glass0.45593.600.45212.930.46103.730.46093.800.46164.070.45593.470.44783.800.37484.800.44134.73
Heartstatlog0.85625.200.85693.870.85884.330.85805.130.86085.870.85624.800.84944.600.82927.800.84407.13
Ionosphere0.929110.130.93875.200.94154.470.93774.330.93857.730.94094.270.906611.270.902315.800.916813.33
LSVT Voice Rehabilitation0.9192154.870.942150.670.946243.270.948229.400.931082.870.940545.870.8717151.200.8234231.470.9073149.73
Sonar0.908127.930.912514.670.932116.530.925715.270.929023.000.923817.400.860027.400.843142.070.902330.60
WDBC0.985114.800.98437.600.98539.130.98468.130.984910.330.98488.530.983113.870.973921.730.983814.47
If an algorithm has the best classification accuracy on the data set, its data are marked as bold.
Table 8. The fitness of the compared algorithms.
Table 8. The fitness of the compared algorithms.
DatasetABBFGOABBFGO-SABBFGO-V2ABBFGO-V3ABBFGO-T1ABBFGO-T2BPSO-TVMSBQUATREBGWO-a
FitnessRankFitnessRankFitnessRankFitnessRankFitnessRankFitnessRankFitnessRankFitnessRankFitnessRank
Turkish Music Emotion0.241260.225450.211210.213430.223640.213020.291990.254670.26448
Cancer0.027410.027620.027940.028050.028050.027830.029270.028460.02927
Musk (Version 1)0.089560.070340.059910.068830.074250.060420.125190.115180.09707
Dermatology0.015140.019750.014830.014310.014720.014720.025370.015140.02086
Dnatest0.148870.096540.090910.092720.108750.094530.183890.176980.14726
German0.492850.494160.487010.491230.487010.488720.510980.492640.50267
Glass0.542750.545770.537820.537930.537510.542540.550980.543360.55849
Heartstatlog0.146360.144740.143120.144630.142310.146050.152780.148470.16009
Ionosphere0.073260.062230.059310.062940.063250.059820.095890.079070.08638
LSVT Voice Rehabilitation0.085060.058930.054720.052210.071050.060440.131990.129180.09667
Sonar0.095760.089150.070010.076230.074120.078340.143190.115880.10197
WDBC0.019760.018140.017610.017930.018450.017920.021490.020670.02098
Total64 52 20 34 41 35 101 80 89
Table 9. The result of Friedman test on feature selection.
Table 9. The result of Friedman test on feature selection.
DatasetSum of SquaresDegree of FreedomMean Squaresp-Value
Turkish Music Emotion2845.4114020.32440.6354
Cancer2876.3614020.54550.9315
Musk (Version 1)2798.7314019.99090.4450
Dermatology2751.5514019.65390.3303
Dnatest2741.4514019.58180.2625
German2797.0914019.97920.4440
Glass2743.9114019.59940.2792
Heartstatlog2939.4514020.99610.9499
Ionosphere2772.2314019.80160.3581
LSVT Voice Rehabilitation2852.5014020.37500.6619
Sonar2851.4514020.36750.6540
WDBC2812.0014020.08570.5548
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pan, J.-S.; Yue, L.; Chu, S.-C.; Hu, P.; Yan, B.; Yang, H. Binary Bamboo Forest Growth Optimization Algorithm for Feature Selection Problem. Entropy 2023, 25, 314. https://doi.org/10.3390/e25020314

AMA Style

Pan J-S, Yue L, Chu S-C, Hu P, Yan B, Yang H. Binary Bamboo Forest Growth Optimization Algorithm for Feature Selection Problem. Entropy. 2023; 25(2):314. https://doi.org/10.3390/e25020314

Chicago/Turabian Style

Pan, Jeng-Shyang, Longkang Yue, Shu-Chuan Chu, Pei Hu, Bin Yan, and Hongmei Yang. 2023. "Binary Bamboo Forest Growth Optimization Algorithm for Feature Selection Problem" Entropy 25, no. 2: 314. https://doi.org/10.3390/e25020314

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop