Next Article in Journal
Vertical Jump Height Estimation Using Low-Sampling IMU in Countermovement Jumps: A Feasible Alternative to Motion Capture and Force Platforms
Previous Article in Journal
A Signal-On Microelectrode Electrochemical Aptamer Sensor Based on AuNPs–MXene for Alpha-Fetoprotein Determination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multipopulation Whale Optimization-Based Feature Selection Algorithm and Its Application in Human Fall Detection Using Inertial Measurement Unit Sensors

School of Mechanical Electrical and Information Engineering, Shandong University, Weihai 264209, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 7879; https://doi.org/10.3390/s24247879
Submission received: 22 October 2024 / Revised: 24 November 2024 / Accepted: 7 December 2024 / Published: 10 December 2024
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Feature selection (FS) is a key process in many pattern-recognition tasks, which reduces dimensionality by eliminating redundant or irrelevant features. However, for complex high-dimensional issues, traditional FS methods cannot find the ideal feature combination. To overcome this disadvantage, this paper presents a multispiral whale optimization algorithm (MSWOA) for feature selection. First, an Adaptive Multipopulation merging Strategy (AMS) is presented, which uses exponential variation and individual location information to divide the population, thus avoiding the premature aggregation of subpopulations and increasing candidate feature subsets. Second, a Double Spiral updating Strategy (DSS) is devised to break out of search stagnations by discovering new individual positions continuously. Last, to facilitate the convergence speed, a Baleen neighborhood Exploitation Strategy (BES) which mimics the behavior of whale tentacles is proposed. The presented algorithm is thoroughly compared with six state-of-the-art meta-heuristic methods and six promising WOA-based algorithms on 20 UCI datasets. Experimental results indicate that the proposed method is superior to other well-known competitors in most cases. In addition, the proposed method is utilized to perform feature selection in human fall-detection tasks, and extensive real experimental results further illustrate the superior ability of the proposed method in addressing practical problems.

1. Introduction

With the continuous development of science and technology, the volume of data is increasing exponentially, which creates “the curse of dimensionality” [1]. In the domain of feature selection, researchers have introduced numerous approaches to address the diversity and complexity of data characteristics. However, many features of the vast databases are nonessential or even redundant. To cope with this issue, researchers have turned to feature selection (FS) as a common tool. By removing irrelevant and redundant features, FS obtains an optimal subset with very few features and high classification accuracies to achieve dimensionality reduction [2]. There are applications of FS techniques in many fields, including text classification [3], bioinformatics [4], spectral-spatial classification [5], and others.
FS has four main operations, generating feature subsets, evaluating feature subsets, stopping criterion, and evaluation [6]. The search methods and evaluation criterion are the most essential parts of the FS algorithms. There are three main search strategies, a complete search strategy, sequence search strategy, and heuristic search strategy. A complete search necessitates traversing all the potential feature subset combinations, such as an exhaustive search [7]. While this strategy assures that the optimal subsets can be obtained, it comes with a considerable computational cost for large datasets. Based on model assessment feedback, the sequence search continuously adds or removes features from the candidate feature subsets to form the best feature subset. Common approaches are sequential backward selection (SBS) [8], sequential forward selection (SFS) [9], sequential floating backward selection (SFBS) [10], and sequential floating forward selection (SFFS) [11]. Lakshmi and Mohanaiah proposed a Multi-Support Vector Neural Network (MultiSVNN) method for FER based on a hybrid Whale Optimization Algorithm-Teaching Learning Based Optimization (WOA-TLBO) [12]. The sequence search algorithms reduce computing costs effectively, but it is easy to fall into the local optimum. A heuristic search gradually employs metaheuristics to approximate the global best subset. The metaheuristics can thoroughly consider the feature interactions during exploration and find the position closer to the target. Hence, a heuristic search is better suited than the first two methods for dealing with complicated and high-dimensional real-world problems. These heuristic algorithms include adaptively balanced grey wolf optimization [13], bare-bones particle swarm optimization [14], and a cultural algorithm and differential evolution [15], etc.
The Whale Optimization Algorithm (WOA) is one of the typical heuristic algorithms that is designed based on the feeding process of humpback whales [16]. The WOA is known for its efficient local exploitation ability and only needs to tune two main internal parameters. With these benefits, WOA has been implemented in a range of areas, including facial emotion recognition [17], medical diagnosis problems [18], and feature selection [19]. To achieve superior feature combinations, researchers frequently improve WOA from many perspectives. To tackle the challenge of feature selection in imbalanced classification, Lin Sun and colleagues proposed a two-stage feature subset selection scheme that integrates the Fuzzy Multi-Granularity Rough Set (FMRS) and a Binary Whale Optimization Algorithm (BWOA) [20], for instance, maintaining the balance of exploration and exploitation [21], getting rid of the local optimum [22,23], improving the global search ability [24,25,26], speeding up convergence [27], and promoting population diversity [28].
This study focuses on three aspects of the challenges in WOA-based feature selection. First, the global best individual of a canonical WOA is the primary directing factor in the search process. Individuals may likely cluster together early in this instance, resulting in limited global exploration capabilities and a premature convergence. Second, as the distance between individuals gradually decreases, the original search approach cannot guarantee that individuals continue to explore alternative locations, which could cause the search to stall. Finally, the population explores everywhere, and individual quality is poor in the initial stage, leading to low search efficiency.
In the context of fall detection, our proposed algorithm demonstrates the effectiveness of the feature selection (FS) method while primarily contributing to advancing FS research. Koo et al. introduced an artificial neural network (ANN) approach integrating ranking algorithms for post-fall detection, which enhances classification performance and balances accuracy with computational efficiency [29]. Their innovative use of ranking algorithms significantly improves dimensionality reduction, offering valuable insights for FS optimization. Similarly, Le et al. compared machine learning classifiers in fall detection, showing that FS techniques improve model performance, reduce computational complexity, and refine feature sets [30]. This study highlighted the synergistic effects of classifiers and FS methods, providing critical guidance for addressing fall-detection challenges. Moreover, widely adopted datasets, such as those in [31] and the influential dataset by Jang et al. [32], have significantly supported research efforts. These works underscore the importance of effective FS and robust datasets in advancing fall-detection applications and FS research.
Therefore, a multispiral whale optimization algorithm (MSWOA) with three improvements is presented to enhance the global search capability and jump out of the local optima trap. The following are the key theoretical contributions of this paper:
  • A new Adaptive Multipopulation merging strategy (AMS) separates a population into several subpopulations based on a particular exponential function, which can generate more candidate solutions and expand the exploration space. Each subpopulation selects individuals according to a fitness-distance criterion, which enhances global exploration capability and prevents premature convergence.
  • A Double Spiral updating Strategy (DSS) based on the golden spiral is designed to update the position of individuals along a specific spiral path. The DSS and the bubble-net attacking method use different spiral properties to achieve inward exploitation and outward exploration, which can address the issue of stalling to the local optimum solutions.
  • A Baleen neighborhood Exploitation Strategy (BES) combines quadratic interpolation (QI) with baleen-assisted predation behavior to enhance the candidate solutions and accelerate convergence. This strategy explores more prospective areas by focusing on the area near the best location, which provides a good direction for the population in the early optimization process.
  • The performance of the MSWOA-based feature selection algorithm is tested on 20 UCI classification datasets and a realistic classification optimization problem.
  • The MSWOA is tested against six variants of the WOA and six other advanced swarm intelligence optimization approaches. Extensive feature selection experiments indicate that MSWOA obtains a more promising overall performance compared with the comparison algorithms on most datasets.
The remaining organization of this paper is as follows: Section 2 provides the various works related to feature selection. Section 3 describes an introduction to the original WOA. Section 4 details the improvement points of the MSWOA approach. Section 5 compares MSWOA with other algorithms and analyzes the reported results from the experiments on various benchmark datasets and real-world problems. Section 6 summarizes this paper and suggests the subsequent research options.

2. Related Work

The feature subset chosen using feature selection approaches contains valuable information, which reduces the training time and facilitates the classification accuracy. Thus, FS can cope with the challenges of high-dimensional data with large quantities of information and complex feature interaction. FS approaches are broadly categorized into filter-based, wrapper-based, and embedding-based methods [33].
Filter-based methods sort features according to different criteria and then select features that meet the threshold and generate a feature subset. This method is independent of the classification algorithm, which gives it the advantage of being computationally efficient. There are several commonly used filtering methods. Eiras-Franco et al. [34] described a new version of ReliefF that simplifies the most expensive phase by approximating the nearest neighbor graph with locality-sensitive hashing (LSH). Song et al. [35] presented a new feature ranking approach on the basis of Fisher discriminant analysis (FDA) and an F-score, which obtained a suitable subset by maximizing the proportion of the average between-class distance to the relative within-class scatter. Other improved approaches in this category include minimal redundancy maximal new classification information (MR-MNCI) [36] and information gain (IG) [37]. However, the selected feature subset using filters has low prediction accuracies due to the lack of a specific learning algorithm guiding the feature selection phase [38].
Wrapper-based methods introduce the learning algorithm to evaluate the quality of feature subsets and consider the complex interactions between the features. A learning algorithm is used in each iteration to assess whether the candidate feature subset is superior to the current best subset [39]. As learning algorithms for evaluating features, the decision tree (DT), K-nearest neighbor (KNN), naïve Bayes (NB) and support vector machine (SVM) have been frequently utilized. In this case, wrappers have better classification accuracy than filters. Recursive feature elimination (RFE) is a traditional wrapper FS method that employs a machine learning model for training and constantly removes features based on the weight value. Han et al. [40] presented a dynamic recursive feature elimination (dRFE) framework. Another typical wrapper feature selection method is the Las Vegas wrapper (LVW) [41]. However, the adoption of learning algorithms incurs a considerable computational cost if the search space is vast [42]. The conventional optimization techniques that search for the optimal feature subset becomes extremely difficult. Therefore, with their powerful global search ability, metaheuristic approaches have been successfully employed in feature selection, these include the ant colony optimization (ACO) and the grasshopper optimization algorithm (GOA).
Embedded-based approaches combine the model construction of the learning algorithm with the feature selection process. The regularization method is one of the most common embedded methods. Based on multimodal neuroimaging, Zhang et al. [43] introduced a new multiclass categorization framework for Alzheimer’s disease (AD) with the introduction of feature selection and fusion, which used a l2,1-norm regularization term and a lp-norm (1 < p < ∞) regularization term. Other algorithms also include the particle swarm optimization-based multiobjective memetic algorithm (PSOMMFS) [44], classification and regression trees (CARTs) [45], decision tree ID3 algorithm [46], etc. Although embedded methods have good classification performance and efficiency, their approach structure is more complicated and ignores feature interactions. Therefore, this paper utilizes a wrapper method based on a metaheuristic algorithm to address the feature-selection problem.
Among the existing approaches, metaheuristics are employed in feature selection in various forms, with hybrid algorithms being one of the most used. Amini and Hu [47] presented a novel hybrid two-level FS method, which combines a GA and EN to construct an appropriate subset of predictors to improve computational efficiencies and prediction performances. Ibrahim et al. [48] integrated the SSA with PSO to boost exploitation abilities and diversity, helping to swiftly obtain the optimal feature subset. This paper presented a novel algorithm for finding high-quality solutions for feature-selection tasks by incorporating the SSA into the Harris hawks optimization (HHO) algorithm [49]. Additionally, academics frequently introduce different strategies in metaheuristics to improve classification performance. Li et al. [50] suggested an enhanced PSO algorithm for feature selection. The personal best position was updated with genetic operators in this technique to avoid the local optimum. Chen et al. [51] presented a novel PSO-based feature selection approach that uses information about the swarm to produce more potential solutions. An improved discretization-based PSO for FS (IDPSO-FS) in [52] was developed to identify each selected feature index accurately. The authors of [53] suggested an improved version of the salp swarm algorithm (ISSA) to solve feature selection problems and to select the optimal subset of the features in wrapper mode. In [54], an improved binary global harmony search algorithm (IBGHS) with a modified improvisation step was presented. Furthermore, several recent articles have combined filter approaches and metaheuristic-based wrapper approaches. Ghosh et al. [55] developed a wrapper-filter combination on the basis of ACO, in which a filter approach is used to evaluate a subset to reduce the computing complexity. In the same year, Moslehi and Haeri [56] presented a hybrid filter-wrapper method that integrates evolutionary-based GA and PSO; an artificial neural network was included into the fitness function.
The WOA is an evolutionary algorithm with simple principles and an easy implementation. Moreover, the WOA exhibits superior performance in solving FS problems. Tubishat et al. [57] incorporated IG filter-feature-reduction techniques with the WOA to capitalize on its advantages and compensate for its shortcomings. The IWOA used the EOBL technique to improve the diversity of the initial solutions and DE evolutionary operators to improve the solutions after each iteration. The IWOA outperformed the other approaches (i.e., WOA, DE, PSO, GA, ALO) on four datasets. To achieve automatic lung tumor detection, a hybrid bioinspired algorithm (WOA_APSO) was developed that combined the benefits of the WOA and APSO to select the optimized feature subset [58]. The experimental results revealed that it outperformed the other advanced methods and achieved 97.18%, 97%, and 98.66% accuracy, sensitivity, and specificity, respectively. In [59], a wrapper-based FS method that hybridized GWO and WOA was developed, aggregating the advantages of their exploration and exploitation. The results showed that HSGW outperforms several well-known feature selection methods (i.e., BGOA, BGSA, GA, and PSO). Mafarja et al. [60] presented a new WOA-based FS by introducing both V-shaped and S-shaped transfer functions, which were adopted in IoT attack detections. This paper integrated five natural selection operators into WOA to enhance the search efficacy in dealing with FS tasks [61]. The proposed wrapper approach was used to determine the most valuable features from the software fault prediction (SFP) datasets. To find the best subset, two binary variants of the WOA were presented [62], which used the sigmoid transfer function and the sigmoid transfer function, respectively. There is also a substantial body of research describing the different WOA variants used in feature selection [63,64,65].
Based on the no-free lunch (NFL) theorem [66], no single strategy can solve all the WOA issues. Thus, there is still some room for further improvement. Focusing on the problem of insufficient global search capability and frequent search stagnation, this paper proposes an innovative wrapper approach called the multispiral whale optimization algorithm (MSWOA). We divide the population into many subpopulations using the novel multipopulation technique, which may enlarge the search space and yield more candidate subsets. In addition, a new stochastic exploration strategy is provided to assist individuals in breaking free from local optima. Furthermore, to maximize the quality of the optimal individual, we apply an exploitation method based on humpback whale behaviors. The MSWOA is tested on 20 UCI datasets and in two group contrast trials. The results show that MSWOA can obtain feature subsets with higher classification accuracies and fewer features on most datasets. Section 4 will go through the MSWOA in further detail.

3. The Whale Optimization Algorithm (WOA)

The WOA is inspired by the distinctive hunting activity of humpback whales. Humpback whales engage in an interesting hunting method called bubble-net feeding, which involves blowing bubbles around the prey in a spiral or ‘9’-shaped route and pushing it toward the center of the net. In the WOA, the target prey symbolizes the current best solution, and humpback whales are the search agents. Based on these characteristics, the mathematical model of WOA is divided into three components: encircling prey, the bubble-net attacking method, and searching for the prey.

3.1. Encircling Prey

When humpback whales swim around their prey, they shrink the circle while swimming in a spiral pattern. After targeting the prey, the search agents keep coming closer to the best solution. The mathematical model of how the present search agent adjusts its location based on the best record position is shown in Equations (1) and (2).
D = C X * ( t ) X ( t )
X ( t + 1 ) = X * ( t ) A D
where t indicates the current iteration, X* indicates the position of the global optimal individual, and X denotes the position of the current individual. | | indicates the absolute value, and · indicates an element-by-element multiplication. The values of vectors A and C control the movement of the current individual around the best solution, and their updating equations are as follows:
A = 2 a r a
C = 2 r
where r is a randomly chosen vector in [0, 1]. The range of a is reduced from 2 to 0 by Equation (5).
a = 2 t 2 M a x I t e r
where t denotes the current number of iterations and MaxIter denotes the number of maximum iterations.

3.2. Bubble-Net Attacking Method

The spiral updating position simulates the unique mechanism of bubble-net foraging to move toward the current optimal solution. The spiral equation is as follows:
X ( t + 1 ) = D e b l cos ( 2 π l ) + X * ( t )
D = X * ( t ) X ( t )
where D′ is the distance of the ith search agent from the optimal individual. b is a constant, and l is a randomly chosen number in [−1, 1].
Above all, the model assumes a parameter p to choose between either encircling prey or using bubble-net attacking to update the position of the whales as in Equation (8).
X ( t + 1 ) = X * ( t ) A D i f   p < 0.5 D e b l cos ( 2 π l ) + X * ( t ) i f   p 0.5
where p is a random number in [0, 1].

3.3. Search for Prey

The exploration phase imitates the behavior of humpback whales when searching randomly. The mathematical formulations of this mechanism are shown in Equations (9) and (10).
D = C X r a n d X ( t )
X ( t + 1 ) = X r a n d A D
where Xrand is a randomly selected individual from the current swarm. The absolute parameter value |A| determines whether the algorithm emphasizes exploitation or exploration. When |A| ≥ 1, Equation (10) is selected to force a solution to explore the potential space. When |A| < 1, Equation (2) is selected.
The pseudocode of the WOA algorithm (Algorithm 1) is as follows:
Algorithm 1. WOA
1 Initializing the population Xi (i = 1, 2…, n)
2 Calculate the individual fitness values in turn
3 X* = the best individual
4 while (t < MaxIter)
5  for each individual
6    Update a, l, p, A, and C
7    If (p < 0.5)
8       If (|A| < 1)
9         Calculate the new position of the current individual according to Equation (2)
10      else if (|A| > 1)
11        Randomly select a individual (Xrand)
12        Calculate the new position of the current individual according to Equation (10)
13      end
14    else if (p > 0.5)
15      Calculate the new position of the current individual according to Equation (6)
16    end
17  end
18  If a search agent exceeds the search range, correct it
19  Calculate the individual fitness values in turn
20  Update X* if there are better results
21  t = t + 1
22 end
23 return X*

4. Proposed MSWOA

Although the WOA achieves good performance in finding reasonable solutions for feature-selection tasks, it still has some drawbacks, including limited global exploration capability and a proclivity to be trapped in the local optimum. To address the drawbacks of WOA, this paper proposes a multispiral whale optimization algorithm, namely, the MSWOA algorithm, which mainly contains three major modifications, an Adaptive Multi-population merging Strategy (AMS), a Double Spiral updating Strategy (DSS), and a Baleen neighborhood Exploitation Strategy (BES). The contents of the three improvement strategies are described below.

4.1. Adaptive Multipopulation Merging Strategy (AMS)

The leader of the population in the original WOA algorithm is single, which might lead to poor global exploration capability and premature convergence. According to Equation (8), the population individuals update their positions primarily according to X* with the best fitness. Therefore, the search range is relatively limited in comparison to the entire space. For this problem, a multipopulation mechanism is a workable approach, in which the population is divided into several subpopulations to enhance the possibility of exploring more space [67]. Each swarm follows its own leader to search for different areas in the search space, which effectively improves the global exploration abilities. Due to the benefits listed above, we introduce the multipopulation concept into the WOA and redesign a more reasonable subpopulation-allocation process with new strategies.
The first procedure is to count the number of subpopulations. In the initial stages of evolution, subpopulations with smaller sizes are advantageous to the diversity of the swarm [68]. In contrast, larger subpopulations are more favorable for producing high-quality solutions in the later stages of evolution. Hence, the number of subpopulations needs to be smoothly merged from K to one. The calculation formula uses exponential characteristics, as shown in Equation (11).
K = f l o o r ( N 2 exp ( ln ( N 2 ) t M a x I t e r ) + γ )
where K is the number of subpopulations, floor represents a function that rounds toward minus infinity, N is the population size, t is the current number of iterations, and MaxIter is the number of maximum iterations. γ is a parameter in [0, 1] that is chosen based on the size of the population. Therefore, each subpopulation size is determined as M = floor(N/K) according to this selection scheme.
For example, when N is 10 and MaxIter is 100, it is reasonable to set γ to 0.8, and the change curve of the number of subpopulations is illustrated in Figure 1. The convergence rate of the number of subpopulations is faster in the early iterations and then gradually become slower, ensuring enough space for exploration in the early stages and finally merging together for deeper exploitation. In the initial phase, the number of subpopulations is five, with each subpopulation containing two individuals. Each subpopulation explores different areas under the guidance of its own optimal individual. Soon, the number of subpopulations aggregates to four, after which they progressively merge into three or two populations at a modest rate. Finally, all individuals aggregate into a single population for a deeper exploitation. This kind of regulation effectively balances exploration and exploitation while providing continuous information sharing.
The second procedure is the selection of the subpopulation individuals. Generally, individuals with different fitness levels have different exploration and exploitation abilities [69]. In addition, taking into account the distance between individuals contributes to enhancing diversity. Therefore, a new selection mechanism is defined in Equation (12), which combines the fitness value and the distance between individuals. According to the comprehensive calculation results from high to low, all the individuals are categorized into different subpopulations.
F D i = ( 1 f i t i ) + d i s i ,   i = 1   ,   2 ,   N
where fit is the normalized fitness, and dis is the normalized value of the distance between the search agent and the global optimal search agent.
In addition, the AMS strategy also introduces a randomized regrouping schedule as another selection mechanism. Every R iterations, individuals are randomly selected for each subpopulation. This mechanism exchanges information among the swarms to stimulate more vitality in the individuals and achieves a wider diversity. The flow of the proposed AMS is shown in Algorithm 2.
Algorithm 2. AMS
Input: N, R, t
1 Calculate the number of subpopulations K by Equation (11) and the size of subpopulations M = floor(N/K)
2 Calculate the individual fitness values in turn
3 X* = the best individual
4 Update FD by the Equation (12) and reorder the population
5 If mod (t, R) = 0
6   Randomly select individuals to form subpopulations
7 end
8 Each subpopulation selects M particles in order
Output: subpopulations Xm(1mK), M, K

4.2. Double Spiral Searching Strategy (DSS)

In the exploration process of the original WOA, a search agent mainly relies on randomly selected individuals to update its location. As shown in Equation (10), when the distance D is small, it is difficult to obtain a competitive new solution. The search is very likely to stop, especially in the later stage of the iteration process. Therefore, the search strategy is further improved by using the golden spiral in this scenario. A golden spiral, such as WOA’s bubble-net attacking method, can spiral inward indefinitely and has a strong exploitation ability. In addition, it can also spiral continuously outward. According to this characteristic, the search agent starts from its current place and then spirals randomly to any location nearby or far away. The position update formulas are defined as follows:
X m ( t + 1 ) = r e θ cos ( 2 π θ ) + X m ( t )
r = W X r a n d X m ( t )
where Xm indicates the current position in the mth subswarm, Xrand indicates a random agent of another subswarm, θ controls the radian of the helix, and W is a random vector in [0, 1].
After implementing the AMS and DSS techniques, the search path has many interacting helices. Part of the population during optimization is depicted in Figure 2. The current agent is located at (Xm, Ym), which is one of the mth subpopulation. Randomly select an agent (Xrand, Yrand) in the population that belongs to the nth sub-population. r represents the helix radius of the particle update. According to Equation (13), the agent can explore any place in the population spiral gap, which increases the randomness and exploration ability considerably.

4.3. Baleen Neighborhood Exploitation Strategy (BES)

Humpback whales have dozens of baleens to assist them in discerning if there is prey nearby, which could be thought of as a local search. Inspired by this idea, a novel Baleen neighborhood Exploitation Strategy (BES) is presented, which imitates baleen behavior with a quadratic interpolation (QI) to speed up convergence. QI is a local search operator that explores the extremum using quadratic approximation [69]. The new method can help the optimal search agent purposefully search for better positions inside the neighborhood. It is beneficial in the early stages of iteration to overcome the poor fitness of individuals. The calculation formula necessitates the determination of three individuals. In this paper, select the optimal individual Xm* (g1, g2…, gD) and an individual Xm (m1, m2…, mD) in the mth subswarm, and the last one Xr (r1, r2…, rD) is chosen at random from the population.
q j = ( m j 2 r j 2 ) F ( X m * ) + ( r j 2 g j 2 ) F ( X m ) + ( g j 2 m j 2 ) F ( X r ) 2 ( ( m j r j ) F ( X m * ) + ( r j g j ) F ( X m ) + ( g j m j ) F ( X r ) )   j = 1 ,   2 ,   D
where qj is the jth dimension of the new solution Xq (q1, q2…, qD). F(Xm*), F(Xm), and F(Xr) are the fitnesses of Xm*, Xm, and Xr, respectively. Then, the crossover operation is employed to obtain an intermediate solution Xc, as shown in Equation (16).
c j = q j i f ( p 0.5 ) g j i f ( p < 0.5 )
where cj represents the jth dimension of the intermediate solution Xc (c1, c2…, cD). Finally, the fitness of two different solutions Xc and Xm* is calculated and compared. The subgroup global optimum will be replaced with the new individual if the current best fitness is inferior to the new fitness. The pseudocode below describes the details of the process of BES (shown in Algorithm 3). This step is only used to help optimize the results in the early stage, and the threshold is set to 0.2.
Algorithm 3. BES
Input: the best solution for subpopulation Xm*, MaxIter, t
1 if t/MaxIter < 0.2
2 for each solution in subpopulation
3   Randomly select individual R in the population
4   Use Equation (15) to calculate the new solution Xq
5   Perform crossover between Xq and Xm* by Equation (16) to get Xc
6   if the fitness of Xc is a better solution, update Xm*
7 end
8 end
Output: Xm*

5. Framework of MSWOA

The pseudocode and flow chart of MSWOA are shown in Algorithm 4 and Figure 3, respectively. The optimization process mainly includes three stages: dividing the population, updating the positions and enhancing the optimal value of the subpopulation. The MSWOA starts with an initial population and then divides it into K subpopulations using the AMS approach. Each subswarm has an optimal individual Xm*. Accordingly, Equations (2) and (6) are revised to Equations (17) and (18).
Algorithm 4. MSWOA
1 Initializing the population Xi (i = 1, 2…, n)
2 while (t < MaxIter)
3  Divide population using AMS
4  for each subpopulation
5    for each solution of subpopulation
6       Update a, l, p, A, and C
7       If (p < 0.5)
8        If (|A| < 1)
9          Calculate the new position of the current individual according to Equation (17)
10         else if (|A| > 1)
11          Randomly select a solution (Xrand)
12          Calculate the new position of the current individual according to Equation (13)
13        end
14      else if (p > 0.5)
15        Calculate the new position of the current individual according to Equation (18)
16      end
17    end
18    Update Xm* by BES
19 end
20 Obtain the global optimal solution X* by comparing Xm* (m = 1, 2…, K)
21 t = t + 1
22 end
23 return X*
X m ( t + 1 ) = X m * ( t ) A D
X m ( t + 1 ) = D e b l cos ( 2 π l ) + X m * ( t )
In the second step, according to the values of the random parameters p and A, individuals are updated using one formula from Equations (13), (17), or (18). After updating each subpopulation, the BES strategy is conducted to purposefully explore the optimal value. If the fitness value of the obtained solution is better, Xm* will be updated. The optimal values of K subpopulations are compared at the end of each iteration, and the global optimal solution is the search agent with the best fitness value.
The computational complexity of MSWOA depends on the maximum number of iterations (MaxIter), subpopulation size (M), subpopulation number (K), dimension size (d), and population size (N). The time complexity of the improved algorithm mainly relies on four parts: initialization, Adaptive Multipopulation merging Strategy (AMS), Double Spiral searching Strategy (DSS), and Baleen neighborhood Exploitation Strategy (BES). The initial stage is O(N × d). The complexity under AMS is O(N). The adaptability of the DSS is assessed to be O(K × M × d). The solution update mechanism of BES is O(K × M × d + K × M). The total time complexity is O(MSWOA) = O(N × d) +O(N + K × M × d + K × M).

5.1. Fitness Function

The evaluation of feature subsets is a crucial aspect of wrapper-based approaches. Generally, the evaluation metrics are classification accuracy and the number of features. In this work, the classification accuracy is calculated by the k-nearest Neighbor (k-NN) classifier (where k = 5) [70]. Since the classification accuracy needs to be maximized while the number of features needs to be minimized, the classification error rate is chosen as an evaluation criterion. In this paper, the fitness function combines the classification error rate and the number of selected features, as shown in Equation (19). Therefore, the criterion is to minimize the fitness value [71].
F i t n e s s = α E r r o r R a t e + β R N
where ErrorRate is the classification error rate, R is the selected features, and N is the full set of features. α and β are the parameters that regulate the ratio of that classification rate and length of a subset, and β = ( 1 α ) and α [ 0 ,   1 ] . In this work, we set α = 0.99 [72].

5.2. Experimental Results and Discussion

Each experiment is completed using MATLAB2017. The machine used to conduct the experiments is configured with an Intel Core i5-10400 2.90 GHz CPU and 16.0 GB RAM. The evaluator is the k-NN classifier, and each dataset is randomly partitioned into a training set and a test set according to the ratio of 80% and 20% [73].
Cross-validation is a more comprehensive model evaluation method that makes effective use of the dataset to enhance the stability and reliability of the model [74]. In comparison to a single train–test split, 5-fold cross-validation divides the entire dataset into five equal parts. In each iteration, one part is used as the validation set, while the remaining four parts are used for training. This process is repeated five times, with a different part serving as the validation set in each iteration. The final model performance is determined by aggregating the results from these five validation rounds. This method helps mitigate the potential evaluation bias introduced by a single validation set and provides a more stable and thorough assessment of model performance [75].
Moreover, cross-validation improves the generalization ability of the model, ensuring that it performs well across different data subsets, thereby reducing the risk of overfitting [76,77,78]. Additionally, it maximizes the utility of each data sample, making it particularly beneficial in cases where the dataset is limited in size. Finally, cross-validation reduces the variability in evaluation results due to the randomness of data splits by repeatedly validating the model on different data partitions, leading to more reliable performance metrics.
For all trials, 5-fold cross-validation is utilized to verify the performance of the selected feature subset to reduce the overfitting issue.

5.3. Datasets and Evaluation Metrics

Twenty different benchmark datasets are gathered from the University of California at Irvine (UCI) [79] to assess the performance of the proposed MSWOA. The details of all the datasets are presented in Table 1. The dimensions of these datasets range from 60 to 22,283, and the datasets were collected from various fields, such as biological/gene expression (Lung_Cancer, Leukemia, Prostate_Tumor), physic (Sonar), speech recognition (Isolet), and face images (Yale, orlraws10P, warpPIE10P). Hence, the performance of the MSWOA can be thoroughly investigated.
The experiments use the following evaluation metrics: average fitness values, average classification accuracy, the number of selected features, convergence curves, and run time. The primary performance indicator for assessing the algorithms is the classification accuracy. In addition, we conduct a significance test and use the symbols ‘+’ ‘−’ ‘≈’ to show various differences in the table, where ‘+’ indicates the MSWOA outperforms the competitor algorithm significantly, ‘−’ indicates the competing algorithm greatly outperforms the MSWOA, and ‘≈’ indicates there is no significant difference. The population size is set according to the corresponding papers [72,80,81,82]. The maximum number of iterations is 100 and the experiments are executed 30 times on each dataset. The feature selection threshold is set to 0.5, which means that if the value is more than 0.5, the feature is chosen; otherwise, the feature is not chosen.

5.4. Comparison Between the WOA-Based Approaches

To verify the superiority of the proposed MSWOA, it is compared with the original WOA and five recent WOA-based methods, including the RDWOA [83], SBWOA [84], QWOA [85], WOASA [86], and WOACM [87]. Table 2 displays several main parameters of the algorithms that have been well tuned.
The average classification accuracy and standard deviations of seven algorithms for the test sets are reported in Table 3, with the numbers in bold indicating the best results. As seen from Table 3, MSWOA outperforms the other variants on 17 out of the 20 datasets. On the PDC, Nci9, and GLIOMA datasets, WOASA exceeds the other approaches and MSWOA achieves the second-highest classification accuracy. On the SRBCT and Leukemia datasets, QWOA, WOACM, and MSWOA all attain 100% classification accuracy. For the Lung_discrete dataset, MSWOA achieves an accuracy of 94.29%, outperforming the second-best result of 92.38%. Furthermore, significance testing indicates that MSWOA holds significant advantages over competing algorithms on most datasets, including Yale, Lung, and Lung_Cancer. In addition, the significance test results indicate that the MSWOA has apparent advantages over the competitive algorithms on most datasets, such as Yale, Lung, and Lung_Cancer. The excellent performance of MSWOA is mainly due to the proposed AMS technique, which boosts the global search capability to explore the high-performance space. Simultaneously, the BES strategy enhances individual quality in the first 20 iterations, providing a better exploration direction for the population.
The average number of selected features of seven methods on 20 datasets is provided in Table 4. The experiment indicates that MSWOA selects fewer features than other algorithms in 15 of the 20 datasets. RDWOA performs second only to MSWOA, with the best results on 3 datasets (i.e., PDC, warpPIE10P, and Arcene). SBWOA performs best on 2 datasets (i.e., Semion and Isolet). Moreover, the average number of selected features of MSWOA is 364.3767 and it ranks first, followed by RDWOA, SBWOA, WOASA, WOA, WOACM, and QWOA. The AMS strategy is the primary rationale for selecting the smallest number of features. Using the proposed AMS, we can evaluate numerous candidate feature subsets rather than just one in each loop, increasing the probability of finding the best answer. To some extent, the AMS also helps to enhance the classification accuracy. For example, on the GLI_85 dataset, MSWOA selects the fewest features compared to the other algorithms and ranks first with a classification accuracy of 100%. In conclusion, the suggested MSWOA determines the best feature combination with high classification accuracy and very few features.
The fitness value can be used to deeply investigate the robustness of the MSWOA. Table 5 shows the standard deviations and average fitness of the seven methods, from which we can see that the MSWOA achieves a better performance than other six methods on 17 out of the 20 datasets. WOASA obtains the smallest fitness values in the remaining three datasets, and MSWOA is slightly worse. WOACM comes in second out of the 6 datasets and is quite close to MSWOA. In addition, the significance test results demonstrate that MSWOA performs better than other methods on most datasets, particularly when compared with WOA, QWOA, and WOACM. MSWOA performs best on the Semion, Lung, 9_Tumors, and Lung_Cancer datasets, with the other algorithms trailing far behind. From the overall results, MSWOA has outstanding performance on different dimensional datasets, which verifies the excellent robustness of MSWOA.
To demonstrate the convergence capability of the proposed MSWOA, the variation curves of the fitness of the seven algorithms on 17 datasets are depicted in Figure 4, from which we can see that the fitness of MSWOA converges fast and achieves the best results in most cases. To explain the change in the mean fitness values, several datasets are presented as examples. For instance, on the Lung_discrete, the fitness value of MSWOA drops to approximately 0.09 after around 25 iterations, whereas the best fitness value among the other six algorithms remains at approximately 0.12. On the 11_Tumors, Isolet, and Lung_discrete datasets, MSWOA converges much faster than the other methods in the entire procedure. On the 9_Tumors, Orlraws10P, and Semion datasets, some algorithms (e.g., WOA, RDWOA, QWOA, SBWOA) perform slightly better than MSWOA in the early phase, but MSWOA can seek better positions and quickly exceeds the others by using the BES strategy. On the Colon, Lung_Cancer, and Yale datasets, MSWOA continuously eliminates search stalls, where the other algorithms are no longer updated. The Leukemia dataset shows that although the other algorithms can also handle the stagnant dilemma, their final fitness values are worse than MSWOA. This is due to the AMS strategy’s enhanced exploration ability, which discoveries additional potential locations.
Training time is a crucial metric for assessing the efficiency of FS algorithms. The average training time for all competing approaches is shown in Table 6. According to the experimental results, the execution time of MSWOA increases slightly when compared with WOA except for the Isolet dataset, because the BES strategy requires one additional estimation of the fitness value. The WOA and RDWOA attain the lowest average training time on nine datasets, respectively. The SBWOA consumes less time on two datasets. According to the average training time, WOA has the fastest calculation speed, followed by SBWOA, RDWOA, MSWOA, WOACM, QWOA, and WOASA. Although RDWOA has the shortest training time on nine datasets, it performs poorly on high-dimensional datasets. In contrast, MSWOA is competitive with the other algorithms on high-dimensional datasets. For example, on the orlraws10P dataset, MSWOA is less than RDWOA, QWOA, WOASA, and WOACM and comparable to SBWOA. Furthermore, it is far superior to the comparison methods in the accuracy, fitness, and screening feature. As a result, MSWOA obtains a more comprehensive feature subset at a reduced cost, especially for high-dimensional datasets.

5.5. Comparisons with Other Well-Known Optimization Algorithms

In this subsection, MSWOA is compared with six different metaheuristic methods to further validate its effectiveness, including MPSO [88], HGWOP [89], HGSO [81], DSSA [90], BGOA [80], and HBBOG [91]. Table 7 displays several main parameters of the methods that have been well tuned.
The average classification accuracy rates, standard deviations, and rankings of MPSO, HGWOP, HGSO, DSSA, BGOA, HBBOG, and MSWOA are presented in Table 8, where the numbers in bold indicate the optimal results. As shown in Table 8, MSWOA has the highest classification accuracy on 15 out of the 20 datasets. On the Sonar, Semion, and Isolet datasets, HGWOP takes first place and the significance test results show little difference between HGWOP and MSWOA. HGSO slightly outperforms MSWOA on the Lung_Cancer datasets. BGOA rounds out the top on the Lung dataset, while the proposed MSWOA ranks second. On the SRBCT and GLI_85 datasets, HGSO and BGOA reach the same level of accuracy as MSWOA, respectively. On the Arcene dataset, MSWOA, HGSO, BGOA, and HBBOG achieve close classification accuracies. Furthermore, the results of the significance tests reveal that MSWOA performs better than other comparison methods on most datasets. For example, when compared to DSSA, MSWOA remained significantly different on 17 datasets and performed similarly on 3 datasets.
According to Table 9, MSWOA produces a smaller number of features than other approaches in 15 out of the 20 datasets. HGSO performs slightly better than MSWOA on the warpPIE10P, 9_Tumors, and Prostate_Tumor datasets. On the PDC and GLIOMA datasets, BGOA is superior to MSWOA. The average number of features selected by MSWOA is 362.2713, which is less than that of any other algorithm. Observing the final ranking, the MSWOA achieves the best results, followed by HGSO, BGOA, MPSO, HBBOG, HGWOP, and DSSA. For high-dimensional datasets, the effect of MSWOA is more pronounced. For example, on the Arcene dataset, the average number of features selected by MSWOA is 600.7333, barely 6% of the total features. As a whole, MSWOA can achieve a higher classification accuracy while using fewer features on high-dimensional and complex datasets. The result also demonstrates that the MSWOA has a significant advantage over the other algorithms in terms of the convergence speed. Considering all the results above, MSWOA is a very competitive feature-selection algorithm. Although MSWOA demonstrates advantages in various aspects, such as achieving higher classification accuracy on 15 out of 20 datasets and significantly reducing the number of selected features, its limitations warrant deeper exploration.
Firstly, while the algorithm’s performance generally surpasses other methods, it did not show substantial superiority over HGSO and BGOA on specific datasets like Lung, SRBCT, and GLI_85. This suggests that MSWOA may have room for further optimization when handling high-dimensional data and complex feature distributions.

5.6. MSWOA for Human Fall Recognition

5.6.1. Problem Description

In this section, the proposed MSWOA is employed as an FS method in a human fall-detection system to enhance the classification recognition. The fall-detection system allows the elderly to obtain the necessary assistance in time after an accidental fall, reducing injuries caused by falls. Human activities are complex composites composed of various simple motions, influenced by diverse contexts, environments, and individual differences, which contribute to variations in behavior [92,93,94]. Consequently, machine learning models trained on incomplete samples and limited features typically exhibit poor generalization performance, making it difficult to accurately differentiate between fall incidents and routine daily activities. To ensure an adequate dataset and diverse feature types, this chapter outlines the design of a data-collection scheme for fall detection.
The current literature divides fall-detection systems into three broad classes, i.e., wearable device, computer vision, and environmental sensor systems [95]. Nevertheless, the latter two systems have the disadvantage of a limited monitoring range, which results in fall detection failing when the user is not within the sensor coverage region. In addition, a vision-based fall-detection system will bring users a huge risk of privacy disclosure. Therefore, the wearable system is selected for a fall-detection experiment because of the benefits of a low cost, simplicity, and low privacy disclosure risk.
Figure 5 illustrates the overall flow of the human fall-detection scheme, including the hardware circuitry, data collection, feature extraction and selection, classification, and results transmission. The hardware consists of a microprocessor, a battery, an inertial sensor module (i.e., MPU6050), a wireless serial module, and an alarm module (i.e., buzzer). MPU6050 has an accelerometer and a gyroscope, which can capture acceleration and angular velocity values and automatically calculate three angles (i.e., pitch, yaw and roll). The sensor’s accelerometer has a measurement range of ±16 g, while the gyroscope’s range is ±2000°/s. The angular range for the X and Z axes is ±180°, and for the Y axis, it is ±90°. The measurement accuracy of the posture angles is 0.05° when static and can reach 0.1° in dynamic conditions, ensuring high measurement precision. To form the original dataset, 13 time-domain features, 6 frequency-domain features, and 8 time-frequency features are extracted from the raw data. Furthermore, the feature subset containing the most information is selected as the input data for the classifier (SVM) using a feature selection method. If the recognition result is a fall, an alert signal is transmitted to both the guardian’s phone and the wearer’s device.

5.6.2. Results and Discussion

The dataset used in the experiment is called SFall, which consists of activities of daily living (ADLs) and falls performed by volunteers. The experiment included seven participants—two females and five males. Due to the inherent risk of fall activities, elderly volunteers were not involved, and all participants were students with an average age of 22 years, an average height of 174.7 cm, and an average weight of 65.2 kg. The fall detection system was configured with a sampling frequency of 20 Hz, and each sampling session was set to a duration of 10 s. The SFall dataset comprises 157 collected samples, each with 200 data points and 243 features derived from acceleration, angular velocity, and angular displacement data. Table 10 shows six falls and eight ADLs. These features include 13 time-domain, 6 frequency-domain, and 8 time-frequency domain metrics, extracted across three axes (x, y, z) from MPU6050 sensor data. The volunteers fix the equipment at the waist and perform a set of activities to collect data. Figure 6 depicts the experimental scenario of some of the actions. The sampling frequency for the data-collection experiment is set to 20 Hz. The acquisition time is 10 s for all falls, and recalibration was performed after every session. As shown in Figure 7, the data signals of the Y-axis acceleration, Y-axis angular velocity, and pitch angle for the three behaviors (walk, squat, and fall) have different magnitudes of variation, which indicates that the collected data can extract effective features to separate falls from ADLs.
On the SFall dataset, MSWOA is experimented against three advanced methods to evaluate their performance, including SASS [96], COLSHADE [97], and sCMAgES [98]. To obtain stable results, the test is run 30 times, and the classifier is Support Vector Machine (SVM). The evaluation metrics are the F1-score (FS) and confusion matrix. As shown in Table 11, TP is the number of falls correctly predicted as falls, FP is the number of ADLs incorrectly predicted as falls, FN is the number of falls incorrectly predicted as ADLs, and TN is the number of ADLs correctly predicted as ADLs. According to the experimental results in Table 12, MSWOA has obtained the best performance. Based on the F1-score analysis, MSWOA achieved a score of 89.6567%, which is 1.5 percentage points higher than the second-ranked SASS algorithm. According to the table, the accuracy of MSWOA is the highest, which is 20% higher than sCMAgE. Based on the confusion matrix, MSWOA predicts the highest number of correct falls with 12.2 and the least number of overall false predictions (FN+FP) with 2.8, indicating that the algorithm possesses robust recognition capabilities for both fall and non-fall behaviors. Thus, the MSWOA-enabled human fall detection system has more robust recognition performance, which proves that the MSWOA method has promising applications in realistic optimization problems.

6. Conclusions

Because of the insufficient randomness of the exploration formula and the single leader, WOA-based feature-selection techniques often suffer from premature convergence and entrapment in local optimal solutions. Therefore, a multispiral WOA (MSWOA) for FS is proposed in this paper to address these dilemmas. The MSWOA includes three major improved modifications. Initially, we utilized an Adaptive Multipopulation merging Strategy (AMS) with a novel allocation method for population delimitation. Each subpopulation chooses its best individual to be the leader, which expands the exploration space and diversity. Additionally, using the Double Spiral updating Strategy (DSS), individuals take their current position as a starting point and randomly explore outward along the spiral trajectory. In this case, the search process is not limited by the distance between the individuals, which prevents search stagnation in the later iterations. Finally, a Baleen neighborhood Exploitation Strategy (BES) with quadratic interpolation was incorporated into the early search phase to obtain better positions and facilitate convergence by searching for regions near the optimal search agent. The proposed MSWOA approach was evaluated against the original WOA and 11 other recent algorithms, including 5 WOA-based and 6 metaheuristic variants on 20 UCI datasets. The results demonstrated that the MSWOA has advantages in classification accuracy, the number of selected features, and the fitness values on most datasets. In addition, MSWOA exhibited a faster convergence speed when compared to the competitive methods. Moreover, the MSWOA is employed for feature selection in a human fall-detection task, and experimental results indicated that it is capable of solving real-world problems.
There is further work that can be performed in the future, for example, combining the multipopulation technique with other updating mechanisms or utilizing different classifiers (e.g., neural networks) to further evaluate the algorithm’s effectiveness. Moreover, the temporal complexity of the algorithm needs to be further reduced to achieve greater computational efficiency. The new MSWOA algorithm can also be applied to other engineering fields, such as fault diagnosis, image segmentation, and path planning.

Author Contributions

Conceptualization, H.C.; methodology, H.C., X.Y. and L.D.; software, H.C.; validation, H.C., X.Y. and B.Y.; investigation, X.Y.; resources, H.C. and L.D.; writing—original draft preparation, H.C. and X.Y.; writing—review and editing, X.Y.; visualization, H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by The National Natural Science Foundation of China (Grant No. 61803227), Young Scholars Program of Shandong University, Weihai (Grant No. 20820211010), China postdoctoral science foundation (Grant No. 2023M743530).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data are not publicly available due to privacy. The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Jia, W.; Sun, M.; Lian, J.; Hou, S. Feature dimensionality reduction: A review. Complex Intell. Syst. 2022, 8, 2663–2693. [Google Scholar] [CrossRef]
  2. Rostami, M.; Berahmand, K.; Nasiri, E.; Forouzandeh, S. Review of swarm intelligence-based feature selection methods. Eng. Appl. Artif. Intell. 2021, 100, 104210. [Google Scholar] [CrossRef]
  3. Babushkin, V.; Alsuradi, H.; Al-Khalil, M.O.; Eid, M. Analyzing Arabic Handwriting Style through Hand Kinematics. Sensors 2024, 24, 6357. [Google Scholar] [CrossRef] [PubMed]
  4. Khalil, R.M.; Shulman, L.M.; Gruber-Baldini, A.L.; Shakya, S.; Fenderson, R.; Van Hoven, M.; Hausdorff, J.M.; von Coelln, R.; Cummings, M.P. Simplification of Mobility Tests and Data Processing to Increase Applicability of Wearable Sensors as Diagnostic Tools for Parkinson’s Disease. Sensors 2024, 24, 4983. [Google Scholar] [CrossRef]
  5. Rasti, B.; Hong, D.; Hang, R.; Ghamisi, P.; Kang, X.; Chanussot, J.; Benediktsson, J.A. Feature extraction for hyperspectral imagery: The evolution from shallow to deep: Overview and toolbox. IEEE Geosci. Remote Sens. Mag. 2020, 8, 60–88. [Google Scholar] [CrossRef]
  6. Wang, P.; Xue, B.; Liang, J.; Zhang, M. Differential evolution-based feature selection: A niching-based multiobjective approach. IEEE Trans. Evol. Comput. 2022, 27, 296–310. [Google Scholar] [CrossRef]
  7. Nssibi, M.; Manita, G.; Korbaa, O. Advances in nature-inspired metaheuristic optimization for feature selection problem: A comprehensive survey. Comput. Sci. Rev. 2023, 49, 100559. [Google Scholar] [CrossRef]
  8. Kannangara, K.P.M.; Zhou, W.; Ding, Z.; Hong, Z. Investigation of feature contribution to shield tunneling-induced settlement using Shapley additive explanations method. J. Rock Mech. Geotech. Eng. 2022, 14, 1052–1063. [Google Scholar] [CrossRef]
  9. Zhang, C.; Xue, Y. Estimation of Biochemical Pigment Content in Poplar Leaves Using Proximal Multispectral Imaging and Regression Modeling Combined with Feature Selection. Sensors 2023, 24, 217. [Google Scholar] [CrossRef]
  10. Tang, C.; Gao, T.; Li, Y.; Chen, B. EEG channel selection based on sequential backward floating search for motor imagery classification. Front. Neurosci. 2022, 16, 1045851. [Google Scholar] [CrossRef]
  11. Molino-Minero-Re, E.; Aguileta, A.A.; Brena, R.F.; Garcia-Ceja, E. Improved accuracy in predicting the best sensor fusion architecture for multiple domains. Sensors 2021, 21, 7007. [Google Scholar] [CrossRef] [PubMed]
  12. Lakshmi, A.V.; Mohanaiah, P. WOA-TLBO: Whale optimization algorithm with Teaching-learning-based optimization for global optimization and facial emotion recognition. Appl. Soft Comput. 2021, 110, 107623. [Google Scholar] [CrossRef]
  13. Wang, J.; Lin, D.; Zhang, Y.; Huang, S. An adaptively balanced grey wolf optimization algorithm for feature selection on high-dimensional classification. Eng. Appl. Artif. Intell. 2022, 114, 105088. [Google Scholar] [CrossRef]
  14. Gad, A.G. Particle swarm optimization algorithm and its applications: A systematic review. Arch. Comput. Methods Eng. 2022, 29, 2531–2561. [Google Scholar] [CrossRef]
  15. Deng, W.; Ni, H.; Liu, Y.; Chen, H.; Zhao, H. An adaptive differential evolution algorithm based on belief space and generalized opposition-based learning for resource allocation. Appl. Soft Comput. 2022, 127, 109419. [Google Scholar] [CrossRef]
  16. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  17. Ashok Kumar, P.M.; Maddala, J.B.; Martin Sagayam, K. Enhanced facial emotion recognition by optimal descriptor selection with neural network. IETE J. Res. 2023, 69, 2595–2614. [Google Scholar] [CrossRef]
  18. Zhang, Y.-P.; Zhang, X.-Y.; Cheng, Y.-T.; Li, B.; Teng, X.-Z.; Zhang, J.; Lam, S.; Zhou, T.; Ma, Z.-R.; Sheng, J.-B.; et al. Artificial intelligence-driven radiomics study in cancer: The role of feature engineering and modeling. Mil. Med. Res. 2023, 10, 22. [Google Scholar] [CrossRef] [PubMed]
  19. Lin, Y.; Liu, H.; Zhao, H.; Hu, Q.; Zhu, X.; Wu, X. Hierarchical feature selection based on label distribution learning. IEEE Trans. Knowl. Data Eng. 2022, 35, 5964–5976. [Google Scholar] [CrossRef]
  20. Sun, L.; Si, S.; Ding, W.; Wang, X.; Xu, J. TFSFB: Two-stage feature selection via fusing fuzzy multi-neighborhood rough set with binary whale optimization for imbalanced data. Inf. Fusion 2023, 95, 91–108. [Google Scholar] [CrossRef]
  21. Tu, J.; Chen, H.; Liu, J.; Heidari, A.A.; Zhang, X.; Wang, M.; Ruby, R.; Pham, Q.V. Evolutionary biogeography-based whale optimization methods with communication structure: Towards measuring the balance. Knowl.-Based Syst. 2021, 212, 106642. [Google Scholar] [CrossRef]
  22. Ye, X.; Liu, W.; Li, H.; Wang, M.; Chi, C.; Liang, G.; Chen, H.; Huang, H. Modified whale optimization algorithm for solar cell and PV module parameter identification. Complexity 2021, 2021, 8878686. [Google Scholar] [CrossRef]
  23. Elaziz, M.A.; Lu, S.; He, S. A multi-leader whale optimization algorithm for global optimization and image segmentation. Expert Syst. Appl. 2021, 175, 114841. [Google Scholar] [CrossRef]
  24. Li, M.W.; Xu, D.Y.; Geng, J.; Hong, W.C. A hybrid approach for forecasting ship motion using CNN–GRU–AM and GCWOA. Appl. Soft Comput. 2022, 114, 108084. [Google Scholar] [CrossRef]
  25. Abdel-Basset, M.; Mohamed, R.; Jameel, M.; Abouhawwash, M. Nutcracker optimizer: A novel nature-inspired metaheuristic algorithm for global optimization and engineering design problems. Knowl.-Based Syst. 2023, 262, 110248. [Google Scholar] [CrossRef]
  26. Zhang, X.; Wen, S. Hybrid whale optimization algorithm with gathering strategies for high-dimensional problems. Expert Syst. Appl. 2021, 179, 115032. [Google Scholar] [CrossRef]
  27. Hussien, A.G.; Heidari, A.A.; Ye, X.; Liang, G.; Chen, H.; Pan, Z. Boosting whale optimization with evolution strategy and Gaussian random walks: An image segmentation method. Eng. Comput. 2023, 39, 1935–1979. [Google Scholar] [CrossRef]
  28. Zhao, F.; Bao, H.; Wang, L.; Cao, J.; Tang, J. A multipopulation cooperative coevolutionary whale optimization algorithm with a two-stage orthogonal learning mechanism. Knowl.-Based Syst. 2022, 246, 108664. [Google Scholar] [CrossRef]
  29. Koo, B.; Kim, J.; Kim, T.; Jung, H.; Nam, Y.; Kim, Y. Post-fall detection using ANN based on ranking algorithms. Int. J. Precis. Eng. Manuf. 2020, 21, 1985–1995. [Google Scholar] [CrossRef]
  30. Le, T.M.; Van Tran, L.; Dao, S.V.T. A feature selection approach for fall detection using various machine learning classifiers. IEEE Access 2021, 9, 115895–115908. [Google Scholar] [CrossRef]
  31. Sucerquia, A.; López, J.D.; Vargas-Bonilla, J.F. SisFall: A fall and movement dataset. Sensors 2017, 17, 198. [Google Scholar] [CrossRef] [PubMed]
  32. Yu, X.; Jang, J.; Xiong, S. A large-scale open motion dataset (KFall) and benchmark algorithms for detecting pre-impact fall of the elderly using wearable inertial sensors. Front. Aging Neurosci. 2021, 13, 692865. [Google Scholar] [CrossRef] [PubMed]
  33. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  34. Eiras-Franco, C.; Guijarro-Berdiñas, B.; Alonso-Betanzos, A.; Bahamonde, A. Scalable feature selection using ReliefF aided by locality-sensitive hashing. Int. J. Intell. Syst. 2021, 36, 6161–6179. [Google Scholar] [CrossRef]
  35. Song, Q.; Jiang, H.; Liu, J. Feature selection based on FDA and F-score for multi-class classification. Expert Syst. Appl. 2017, 81, 22–27. [Google Scholar] [CrossRef]
  36. Gao, W.; Hu, L.; Zhang, P.; Wang, F. Feature selection by integrating two groups of feature evaluation criteria. Expert Syst. Appl. 2018, 110, 11–19. [Google Scholar] [CrossRef]
  37. Yin, Y.; Jang-Jaccard, J.; Xu, W.; Singh, A.; Zhu, J.; Sabrina, F.; Kwak, J. IGRF-RFE: A hybrid feature selection method for MLP-based network intrusion detection on UNSW-NB15 dataset. J. Big Data 2023, 10, 15. [Google Scholar] [CrossRef]
  38. Li, J.; Cheng, K.; Wang, S.; Morstatter, F.; Trevino, R.P.; Tang, J.; Liu, H. Feature selection: A data perspective. ACM Comput. Surv. (CSUR) 2017, 50, 1–45. [Google Scholar] [CrossRef]
  39. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput. 2015, 20, 606–626. [Google Scholar] [CrossRef]
  40. Han, Y.; Huang, L.; Zhou, F. A dynamic recursive feature elimination framework (dRFE) to further refine a set of OMIC biomarkers. Bioinformatics 2021, 37, 2183–2189. [Google Scholar] [CrossRef]
  41. Zhang, B.; Ni-Jia-Ti, M.-Y.; Yan, R.; An, N.; Chen, L.; Liu, S.; Chen, L.; Chen, Q.; Li, M.; Chen, Z.; et al. CT-based radiomics for predicting the rapid progression of coronavirus disease 2019 (COVID-19) pneumonia lesions. Br. J. Radiol. 2021, 94, 20201007. [Google Scholar] [CrossRef] [PubMed]
  42. Cilia, N.D.; De Stefano, C.; Fontanella, F.; Scotto di Freca, A. Variable-length representation for EC-based feature selection in high-dimensional data. In Applications of Evolutionary Computation, Proceedings of the 22nd International Conference, EvoApplications 2019, Held as Part of EvoStar 2019, Leipzig, Germany, 24–26 April 2019; Proceedings 22; Springer International Publishing: Cham, Switzerland, 2019; pp. 325–340. [Google Scholar]
  43. Zhang, Y.; Wang, S.; Xia, K.; Jiang, Y.; Qian, P. Alzheimer’s Disease Neuroimaging Initiative. Alzheimer’s disease multiclass diagnosis via multimodal neuroimaging embedding feature selection and fusion. Inf. Fusion 2021, 66, 170–183. [Google Scholar] [CrossRef]
  44. Luo, J.; Zhou, D.; Jiang, L.; Ma, H. A particle swarm optimization based multiobjective memetic algorithm for high-dimensional feature selection. Memetic Comput. 2022, 14, 77–93. [Google Scholar] [CrossRef]
  45. Ozcan, M.; Peker, S. A classification and regression tree algorithm for heart disease modeling and prediction. Healthc. Anal. 2023, 3, 100130. [Google Scholar] [CrossRef]
  46. An, Y.; Zhou, H. Short term effect evaluation model of rural energy construction revitalization based on ID3 decision tree algorithm. Energy Rep. 2022, 8, 1004–1012. [Google Scholar] [CrossRef]
  47. Amini, F.; Hu, G. A two-layer feature selection method using genetic algorithm and elastic net. Expert Syst. Appl. 2021, 166, 114072. [Google Scholar] [CrossRef]
  48. Ibrahim, R.A.; Ewees, A.A.; Oliva, D.; Elaziz, M.A.; Lu, S. Improved salp swarm algorithm based on particle swarm optimization for feature selection. J. Ambient. Intell. Humaniz. Comput. 2019, 10, 3155–3169. [Google Scholar] [CrossRef]
  49. Zhang, Y.; Liu, R.; Wang, X.; Chen, H.; Li, C. Boosted binary Harris hawks optimizer and feature selection. Eng. Comput. 2021, 37, 3741–3770. [Google Scholar] [CrossRef]
  50. Li, A.-D.; Xue, B.; Zhang, M. Improved binary particle swarm optimization for feature selection with new initialization and search space reduction strategies. Appl. Soft Comput. 2021, 106, 107302. [Google Scholar] [CrossRef]
  51. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. Correlation-guided updating strategy for feature selection in classification with surrogate-assisted particle swarm optimization. IEEE Trans. Evol. Comput. 2021, 26, 1015–1029. [Google Scholar] [CrossRef]
  52. Kang, Y.; Peng, L.; Guo, J.; Lu, Y.; Yang, Y.; Fan, B.; Pu, B. A Fast Hybrid Feature Selection Method Based on Dynamic Clustering and Improved Particle Swarm Optimization for High-Dimensional Health Care Data. IEEE Trans. Consum. Electron. 2023, 70, 2447–2459. [Google Scholar] [CrossRef]
  53. Tubishat, M.; Idris, N.; Shuib, L.; Abushariah, M.A.; Mirjalili, S. Improved Salp Swarm Algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst. Appl. 2020, 145, 113122. [Google Scholar] [CrossRef]
  54. Gholami, J.; Pourpanah, F.; Wang, X. Feature selection based on improved binary global harmony search for data classification. Appl. Soft Comput. 2020, 93, 106402. [Google Scholar] [CrossRef]
  55. Ghosh, M.; Guha, R.; Sarkar, R.; Abraham, A. A wrapper-filter feature selection technique based on ant colony optimization. Neural Comput. Appl. 2020, 32, 7839–7857. [Google Scholar] [CrossRef]
  56. Moslehi, F.; Haeri, A. A novel hybrid wrapper–filter approach based on genetic algorithm, particle swarm optimization for feature subset selection. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 1105–1127. [Google Scholar] [CrossRef]
  57. Tubishat, M.; Abushariah, M.A.; Idris, N.; Aljarah, I. Improved whale optimization algorithm for feature selection in Arabic sentiment analysis. Appl. Intell. 2019, 49, 1688–1707. [Google Scholar] [CrossRef]
  58. Vijh, S.; Gaurav, P.; Pandey, H.M. Hybrid bio-inspired algorithm and convolutional neural network for automatic lung tumor detection. Neural Comput. Appl. 2023, 35, 23711–23724. [Google Scholar] [CrossRef]
  59. Mafarja, M.; Qasem, A.; Heidari, A.A.; Aljarah, I.; Faris, H.; Mirjalili, S. Efficient hybrid nature-inspired binary optimizers for feature selection. Cogn. Comput. 2020, 12, 150–175. [Google Scholar] [CrossRef]
  60. Mafarja, M.; Heidari, A.A.; Habib, M.; Faris, H.; Thaher, T.; Aljarah, I. Augmented whale feature selection for IoT attacks: Structure, analysis and applications. Future Gener. Comput. Syst. 2020, 112, 18–40. [Google Scholar] [CrossRef]
  61. Hassouneh, Y.; Turabieh, H.; Thaher, T.; Tumar, I.; Chantar, H.; Too, J. Boosted whale optimization algorithm with natural selection operators for software fault prediction. IEEE Access 2021, 9, 14239–14258. [Google Scholar] [CrossRef]
  62. Hussien, A.G.; Oliva, D.; Houssein, E.H.; Juan, A.A.; Yu, X. Binary whale optimization algorithm for dimensionality reduction. Mathematics 2020, 8, 1821. [Google Scholar] [CrossRef]
  63. Vo, H.H.-P.; Nguyen, T.M.; Bui, K.A.; Yoo, M. Traffic Flow Prediction in 5G-Enabled Intelligent Transportation Systems Using Parameter Optimization and Adaptive Model Selection. Sensors 2024, 24, 6529. [Google Scholar] [CrossRef] [PubMed]
  64. Si, Q.; Li, C. Indoor robot path planning using an improved whale optimization algorithm. Sensors 2023, 23, 3988. [Google Scholar] [CrossRef] [PubMed]
  65. Wang, F.; Zhang, X.; Li, X.; Gao, G. Advancing Slim-Hole Drilling Accuracy: A CI-WOA-CNN Approach for Temperature-Compensated Pressure Measurements. Sensors 2024, 24, 2162. [Google Scholar] [CrossRef]
  66. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  67. Liang, J.J.; Suganthan, P.N. Dynamic multi-swarm particle swarm optimizer with local search. In Proceedings of the 2005 IEEE Congress on Evolutionary Computation, Edinburgh, UK, 2–5 September 2005; IEEE: Piscataway, NJ, USA, 2005; Volume 1, pp. 522–528. [Google Scholar]
  68. Liu, C.; Li, Q.; Tian, X.; Li, C. Optimal virtual inertial-based power system frequency regulation through multi-cluster wind turbines using BWOA. Front. Energy Res. 2022, 10, 848905. [Google Scholar] [CrossRef]
  69. Sun, Y.; Chen, Y. Multi-population improved whale optimization algorithm for high dimensional optimization. Appl. Soft Comput. 2021, 112, 107854. [Google Scholar] [CrossRef]
  70. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 1992, 46, 175–185. [Google Scholar] [CrossRef]
  71. Peng, L.; Cai, Z.; Heidari, A.A.; Zhang, L.; Chen, H. Hierarchical Harris hawks optimizer for feature selection. J. Adv. Res. 2023, 53, 261–278. [Google Scholar] [CrossRef]
  72. Hussain, K.; Neggaz, N.; Zhu, W.; Houssein, E.H. An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Syst. Appl. 2021, 176, 114778. [Google Scholar] [CrossRef]
  73. Alsaafin, A.; Elnagar, A. A minimal subset of features using feature selection for handwritten digit recognition. J. Intell. Learn. Syst. Appl. 2017, 9, 55–68. [Google Scholar] [CrossRef]
  74. Yates, L.A.; Aandahl, Z.; Richards, S.A.; Brook, B.W. Cross validation for model selection: A review with examples from ecology. Ecol. Monogr. 2023, 93, e1557. [Google Scholar] [CrossRef]
  75. Zhang, X.; Liu, C.A. Model averaging prediction by K-fold cross-validation. J. Econom. 2023, 235, 280–301. [Google Scholar] [CrossRef]
  76. Yang, J.; Wu, Z.; Wang, Z.; Zhang, D.; Wang, W.; Wen, Q.; Zhang, W. Enhanced anisotropic radius basis function metamodel based on recursive evolution Latin hypercube design and fast K-fold cross-validation. Struct. Multidiscip. Optim. 2023, 66, 169. [Google Scholar] [CrossRef]
  77. Huang, K.W.; Wu, Z.X.; Jiang, C.L.; Huang, Z.H.; Lee, S.H. WPO: A whale particle optimization algorithm. Int. J. Comput. Intell. Syst. 2023, 16, 115. [Google Scholar] [CrossRef]
  78. Han, Y.; Zhang, M.; Nan, P.; Cao, J.; Huang, Q.; Ye, X.; He, Z. Two-stage heuristic algorithm for vehicle-drone collaborative delivery and pickup based on medical supplies resource allocation. J. King Saud Univ.-Comput. Inf. Sci. 2023, 35, 101811. [Google Scholar] [CrossRef]
  79. Dua, D.; Graff, C. UCI Machine Learning Repository. University of California, Irvine, School of Information and Computer Sciences. 2017. Available online: http://archive.ics.uci.edu/ml (accessed on 1 June 2024).
  80. Mafarja, M.; Aljarah, I.; Faris, H.; Hammouri, A.I.; Al-Zoubi, A.M.; Mirjalili, S. Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst. Appl. 2019, 117, 267–286. [Google Scholar] [CrossRef]
  81. Neggaz, N.; Houssein, E.H.; Hussain, K. An efficient henry gas solubility optimization for feature selection. Expert Syst. Appl. 2020, 152, 113364. [Google Scholar] [CrossRef]
  82. Hammouri, A.I.; Mafarja, M.; Al-Betar, M.A.; Awadallah, M.A.; Abu-Doush, I. An improved dragonfly algorithm for feature selection. Knowl.-Based Syst. 2020, 203, 106131. [Google Scholar] [CrossRef]
  83. Chen, H.; Yang, C.; Heidari, A.A.; Zhao, X. An efficient double adaptive random spare reinforced whale optimization algorithm. Expert Syst. Appl. 2020, 154, 113018. [Google Scholar] [CrossRef]
  84. Too, J.; Mafarja, M.; Mirjalili, S. Spatial bound whale optimization algorithm: An efficient high-dimensional feature selection approach. Neural Comput. Appl. 2021, 33, 16229–16250. [Google Scholar] [CrossRef]
  85. Agrawal, R.K.; Kaur, B.; Sharma, S. Quantum based whale optimization algorithm for wrapper feature selection. Appl. Soft Comput. 2020, 89, 106092. [Google Scholar] [CrossRef]
  86. Mafarja, M.M.; Mirjalili, S. Hybrid whale optimization algorithm with simulated annealing for feature selection. Neurocomputing 2017, 260, 302–312. [Google Scholar] [CrossRef]
  87. Mafarja, M.; Mirjalili, S. Whale optimization approaches for wrapper feature selection. Appl. Soft Comput. 2018, 62, 441–453. [Google Scholar] [CrossRef]
  88. Liu, H.; Zhang, X.-W.; Tu, L.-P. A modified particle swarm optimization using adaptive strategy. Expert Syst. Appl. 2020, 152, 113353. [Google Scholar] [CrossRef]
  89. Zhang, X.; Lin, Q.; Mao, W.; Liu, S.; Dou, Z.; Liu, G. Hybrid Particle Swarm and Grey Wolf Optimizer and its application to clustering optimization. Appl. Soft Comput. 2021, 101, 107061. [Google Scholar] [CrossRef]
  90. Tubishat, M.; Ja’Afar, S.; Alswaitti, M.; Mirjalili, S.; Idris, N.; Ismail, M.A.; Omar, M.S. Dynamic Salp swarm algorithm for feature selection. Expert Syst. Appl. 2021, 164, 113873. [Google Scholar] [CrossRef]
  91. Zhang, X.; Kang, Q.; Cheng, J.; Wang, X. A novel hybrid algorithm based on biogeography-based optimization and grey wolf optimizer. Appl. Soft Comput. 2018, 67, 197–214. [Google Scholar] [CrossRef]
  92. Yu, M.; Rhuma, A.; Naqvi, S.M.; Wang, L.; Chambers, J. A posture recognition based fall detection system for monitoring an elderly person in a smart home environment. IEEE Trans. Inf. Technol. Biomed. 2012, 16, 1274–1286. [Google Scholar] [CrossRef]
  93. Lord, S.R.; Ward, J.A.; Williams, P.; Anstey, K.J. An epidemiological study of falls in older community-dwelling women: The randwick falls and fractures study. Aust. J. Public Health 1993, 17, 240–245. [Google Scholar] [CrossRef]
  94. Gianoglio, C.; Mohanna, A.; Rizik, A.; Moroney, L.; Valle, M. On Edge Human Action Recognition Using Radar-Based Sensing and Deep Learning. IEEE Trans. Ind. Inform. 2023, 20, 4160–4172. [Google Scholar] [CrossRef]
  95. Han, S.; Li, S.; Fu, X.; Han, S.; Chen, H.; Zhang, L.; Wang, J.; Sun, G. Research Progress of Flexible Piezoresistive Sensors Based on Polymer Porous Materials. ACS Sens. 2024, 9, 3848–3863. [Google Scholar] [CrossRef] [PubMed]
  96. Kumar, A.; Das, S.; Kong, L.; Snášel, V. Self-adaptive spherical search with a low-precision projection matrix for real-world optimization. IEEE Trans. Cybern. 2021, 53, 4107–4121. [Google Scholar] [CrossRef] [PubMed]
  97. Gurrola-Ramos, J.; Hernàndez-Aguirre, A.; Dalmau-Cedeño, O. COLSHADE for real-world single-objective constrained optimization problems. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
  98. Kumar, A.; Das, S.; Zelinka, I. A modified covariance matrix adaptation evolution strategy for real-world constrained optimization problems. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference Companion, Cancún, Mexico, 8–12 July 2020; pp. 11–12. [Google Scholar]
Figure 1. Variation curve of the number of subpopulations.
Figure 1. Variation curve of the number of subpopulations.
Sensors 24 07879 g001
Figure 2. Particle spiral trajectory curve.
Figure 2. Particle spiral trajectory curve.
Sensors 24 07879 g002
Figure 3. Flowchart of MSWOA.
Figure 3. Flowchart of MSWOA.
Sensors 24 07879 g003
Figure 4. Seven methods’ convergence curves on 17 datasets.
Figure 4. Seven methods’ convergence curves on 17 datasets.
Sensors 24 07879 g004aSensors 24 07879 g004b
Figure 5. Flowchart of the fall detection system.
Figure 5. Flowchart of the fall detection system.
Sensors 24 07879 g005
Figure 6. The data collection details of the SFall dataset.
Figure 6. The data collection details of the SFall dataset.
Sensors 24 07879 g006
Figure 7. Signal changes in acceleration, angular velocity, and angle for three types of activities.
Figure 7. Signal changes in acceleration, angular velocity, and angle for three types of activities.
Sensors 24 07879 g007
Table 1. List of datasets used in the experiments.
Table 1. List of datasets used in the experiments.
No.DatasetClassesNumber of FeaturesNumber of Instances
1Sonar260208
2Semion102651593
3Lung_discrete732573
4Isolet266171560
5PDC2756753
6Yale151024165
7Colon2200062
8SRBCT4230883
9warpPIE10P102420210
10Lung53312203
11GLIOMA4443450
129_Tumors9572660
13Leukemia2707072
14Nci99971260
15Arcene210,000200
16orlraws10P1010,304100
17Prostate_Tumor210,509102
1811_Tumors1112,533174
19Lung_Cancer512,600203
20GLI_85222,28385
Table 2. Each optimization algorithm’s parameter settings.
Table 2. Each optimization algorithm’s parameter settings.
AlgorithmParameter Settings
WOAThe value of a ranges from 2 to 0, b is 1
RDWOAThe value of a ranges from 2 to 0, b is 1, w1 and w2 are the random numbers in (0, 1) and (0.5, 1)
SBWOAThe value of a1 ranges from 2 to 0, b is 1
QWOAq = 2 − t × 2/maxGen, p = 1
WOASAThe value of a ranges from 2 to 0, b is 1, T0 = 2 ×|N|
WOACMThe value of a ranges from 2 to 0, b is 1, the value of r ranges from 0.9 to 0
MSWOAThe value of a ranges from 2 to 0, b is 1, γ = 0.8
Table 3. The seven algorithms’ average classification accuracy on the 20 datasets.
Table 3. The seven algorithms’ average classification accuracy on the 20 datasets.
DatasetCriteriaWOARDWOASBWOAQWOAWOASAWOACMMSWOA
SonarMean92.1138 (+)91.7886 (+)92.8485 (+)90.4878 (+)93.5772 (≈)93.9024 (≈)94.3902
Std2.05842.91892.94801.31352.63322.15871.9049
SemionMean97.9769 (+)97.9036 (+)98.1447 (+)97.6520 (+)97.9874 (+)98.3962 (≈)98.4801
Std0.36930.48940.40070.23990.35950.28530.4226
Lung_discreteMean87.1429 (+)89.2857 (+)90.2381 (+)88.8095 (+)92.3810 (≈)91.6667 (+)94.2857
Std3.86864.01955.04513.53954.85624.15824.2857
IsoletMean90.8440 (+)90.7799 (+)91.7628 (≈)90.4380 (+)91.2607 (+)91.8483 (≈)92.1368
Std0.83861.11240.78140.62530.64070.73870.8097
PDCMean83.6424 (+)84.4592 (≈)83.9294 (+)82.3620 (+)85.9603 (≈)83.5320 (+)85.2318
Std2.16392.07132.51272.30992.01163.54882.0953
YaleMean68.7879 (+)67.2727 (+)69.2929 (≈)67.4747 (+)69.2929 (+)68.9899 (+)70.5051
Std2.94323.35622.56331.34012.31222.43472.7028
ColonMean98.3333 (+)98.6111 (≈)99.1667 (≈)99.7222 (≈)100 (≈)99.7222 (≈)100
Std3.33333.10562.50001.495901.49590
SRBCTMean99.7917 (≈)99.5833 (≈)99.7917 (≈)100 (≈)99.7917 (≈)100 (≈)100
Std1.12191.55901.121901.121900
warpPIE10PMean96.2698 (≈)95.8730 (+)85.4762 (+)95.3968 (+)96.8254 (≈)96.2698 (≈)96.9841
Std1.81151.93755.00761.05291.54711.46561.2192
LungMean94.5833 (+)95.1667 (+)95.2482 (+)95.0000 (+)95.1667 (+)95.0833 (+)96.0833
Std1.45541.10551.2615000.44881.3969
GLIOMAMean91.0000 (+)92.6667 (≈)91.6561 (+)88.3333 (+)95.0000 (≈)90.3333 (+)94.3333
Std5.97225.12085.64914.53385.00006.04614.9554
9_TumorsMean66.9444 (≈)65.8333 (+)45.5556 (+)64.7222 (+)66.3889 (+)66.9444 (≈)69.4444
Std5.88605.42209.06083.52464.55595.03844.4790
LeukemiaMean98.5714 (+)98.5714 (+)97.8571 (+)100 (≈)99.5238 (≈)100 (≈)100
Std2.85712.85713.273301.781700
Nci9Mean60.5556 (≈)60.2778 (≈)60.2778 (≈)58.8889 (+)61.6667 (≈)59.7222 (≈)60.8333
Std4.77915.12954.12952.07874.61483.10563.8188
ArceneMean89.7500 (+)90.0833 (≈)90.0833 (≈)87.9167 (+)90.2500 (≈)89.6667 (≈)90.9167
Std2.26842.45662.45661.45542.68872.71831.8801
orlraws10PMean97.8333 (+)98.8333 (≈)98.0000 (+)98.5000 (+)98.8333 (≈)98.0000 (+)99.6667
Std2.79382.11482.76892.29132.11482.44951.2472
Prostate_TumorMean92.3333 (+)92.3333 (≈)92.0000 (+)89.5000 (+)93.6667 (≈)91.8333 (+)94.1667
Std3.59013.35003.55902.98613.63623.02313.4359
11_TumorsMean89.9020 (+)88.8235 (+)89.8039 (+)88.5294 (+)89.6078 (+)91.0784 (≈)92.0588
Std2.48222.88182.24421.16422.48802.21192.1680
Lung_CancerMean98.0833 (+)97.8333 (+)98.0000 (+)97.6667 (+)98.0833 (+)98.5000 (+)99.1667
Std1.05741.06721.00000.62361.23881.22471.1785
GLI_85Mean98.8235 (+)98.8235 (+)85.0980 (+)98.6275 (+)99.8039 (≈)99.2157 (≈)100
Std2.35292.35298.42002.48801.05591.99960
1st/2nd 0/20/30/12/14/82/717/3
+/−/≈ 16/0/412/0/814/0/617/0/37/0/138/0/12
Table 4. The average number of selected features of seven methods on 20 datasets.
Table 4. The average number of selected features of seven methods on 20 datasets.
DatasetFull WOARDWOASBWOAQWOAWOASAWOACMMSWOA
Sonar6021.266716.900016.233327.400020.533318.266712.2333
Semion265120.966798.300088.1667126.1333109.3333109.133388.3000
Lung_discrete32532.966726.600039.333368.400028.100051.466720.8667
Isolet617346.1333283.9333222.3000396.4000351.1000295.5000263.2667
PDC75662.300034.033347.9000165.966739.4667152.966777.7000
Yale1024436.2333253.2667241.6000633.1000381.4333402.4333201.9667
Colon200050.500015.100030.0000231.266742.000062.266714.9667
SRBCT2308229.700090.8333138.4000449.4667157.666755.800026.8667
warpPIE10P2420818.9333447.6667500.43331260.2333652.36671041.6000457.7000
Lung3312264.5000112.0333217.6667790.0667165.4667349.100077.3000
GLIOMA443466.033339.566795.3667313.533354.8000291.133336.4667
9_Tumors5726709.2333785.8667802.93332078.8333931.00001160.5333668.6333
Leukemia7070161.6667251.8000102.36671525.5333426.96671357.200096.0667
Nci99712879.8667359.4667880.76672972.5000495.93331616.2667324.4667
Arcene10,0001378.0333787.9667985.43334292.70001402.13332573.16671347.0333
orlraws10P10,30438.033336.100042.7333129.666742.2667192.833324.6667
Prostate_Tumor10,509244.4333194.2333195.73331467.5333396.1667476.900080.1000
11_Tumors12,5334783.43332928.90002834.10005626.13334392.56673760.36672437.4000
Lung_Cancer12,6001267.43331067.03331142.43332808.70001096.9333888.5667229.1000
GLI_8522,2832513.8333951.03331337.06676408.56671926.46675108.0000802.4333
Ave 6101721.2750439.0317498.04831588.6067655.6350998.1750364.3767
Rank 5237461
Table 5. The seven algorithms’ average fitness values on 20 datasets.
Table 5. The seven algorithms’ average fitness values on 20 datasets.
DatasetCriteriaWOARDWOASBWOAQWOAWOASAWOACMMSWOA
SonarMean0.0816 (+)0.0841 (+)0.0735 (+)0.0987 (+)0.0670 (≈)0.0634 (≈)0.0576
Std0.02080.02920.02540.01280.02670.02160.0188
SemionMean0.0246 (+)0.0245 (+)0.0217 (+)0.0280 (+)0.0241 (+)0.0209 (+)0.0039
Std0.00410.00490.00410.00240.02000.01810.0246
Lung_discreteMean0.1283 (+)0.1069 (+)0.0979 (+)0.1129 (+)0.0763 (≈)0.0841 (+)0.0572
Std0.03830.03970.04980.03530.04810.04150.0424
IsoletMean0.0963 (+)0.0959 (+)0.0852 (≈)0.1011 (+)0.0922 (+)0.0855 (≈)0.0821
Std0.00890.01140.00790.00690.00730.00770.0081
PDCMean0.1628 (+)0.1543 (≈)0.1597 (+)0.1768 (+)0.1395 (≈)0.1651 (+)0.1457
Std0.02230.02030.02510.02480.02000.03640.0206
YaleMean0.3133 (+)0.3265 (+)0.3064 (≈)0.3282 (+)0.3077 (+)0.3109 (+)0.2940
Std0.03000.03340.02570.01310.02350.02450.0272
ColonMean0.0168 (+)0.0138 (+)0.0084 (+)0.0039 (+)0.0002 (≈)0.0031 (≈)0.0001
Std0.03290.03070.02470.01470.00080.01480.0002
SRBCTMean0.0031 (+)0.0045 (+)0.0027 (+)0.0019 (+)0.0027 (+)0.0002 (+)0.0001
Std0.01180.01560.01110.00200.01130.00020.0001
warpPIE10PMean0.0403 (+)0.0427 (+)0.0382 (≈)0.0508 (+)0.0341 (≈)0.0412 (+)0.0317
Std0.01920.01920.01620.01020.01570.01480.0127
LungMean0.0544 (+)0.0482 (+)0.0477 (+)0.0519 (+)0.0483 (+)0.0497 (+)0.0390
Std0.01430.01100.00760.00210.00900.00480.0139
GLIOMAMean0.0892 (+)0.0727 (≈)0.0827 (+)0.1162 (+)0.0496 (≈)0.0964 (+)0.0562
Std0.05920.05070.06310.04490.04960.05990.0491
9_TumorsMean0.3285 (+)0.3396 (+)0.3342 (+)0.3529 (+)0.3344 (+)0.3293 (+)0.3009
Std0.05830.05370.05010.03530.04530.04990.0397
LeukemiaMean0.0144 (+)0.0145 (+)0.0214 (+)0.0022 (+)0.0053 (+)0.0019 (+)0.0001
Std0.02820.02810.03230.00290.01750.00160.0002
Nci9Mean0.3914 (≈)0.3936 (≈)0.3942 (+)0.4101 (+)0.3800 (≈)0.4004 (+)0.3881
Std0.04760.05070.04120.02080.04590.03090.0377
ArceneMean0.1029 (≈)0.0990 (≈)0.0992 (≈)0.1239 (+)0.0979 (≈)0.1049 (+)0.0913
Std0.02340.02450.02430.01500.02740.02810.0192
orlraws10PMean0.0215 (+)0.0116 (+)0.0198 (+)0.0150 (+)0.0116 (+)0.0200 (+)0.0017
Std0.02770.02090.02740.02270.02100.02440.0089
Prostate_TumorMean0.0761 (+)0.0761 (+)0.0794 (+)0.1053 (+)0.0631 (≈)0.0813 (+)0.0578
Std0.03580.03320.03530.03020.03660.03030.0340
11_TumorsMean0.1038 (+)0.1130 (+)0.1032 (+)0.1180 (+)0.1064 (+)0.0913 (+)0.0806
Std0.02520.02860.02230.01170.02510.02180.0214
Lung_CancerMean0.0200 (+)0.0223 (+)0.0207 (+)0.0253 (+)0.0198 (+)0.0156 (+)0.0084
Std0.01090.01080.01030.00690.01260.01260.0117
GLI_85Mean0.0128 (+)0.0121 (≈)0.0025 (≈)0.0165 (+)0.0028 (≈)0.0101 (+)0.0004
Std0.02300.02320.01040.02460.01050.01980.0004
1st/2nd 0/10/10/40/03/50/617/3
+/−/≈ 18/0/215/0/515/0/520/0/010/0/1017/0/3
Table 6. The average training time of seven algorithms on 20 datasets (unit: s).
Table 6. The average training time of seven algorithms on 20 datasets (unit: s).
DatasetWOARDWOASBWOAQWOAWOASAWOACMMSWOA
Sonar27.076724.377927.594027.405956.084228.166834.1727
Semion170.6904124.2816127.9725218.3407316.6633168.6442167.3333
Lung_discrete28.916225.066729.545329.376161.957831.594038.9689
Isolet156.7579112.1603105.6040169.6362295.2245140.4256153.0561
PDC60.763844.045757.4741133.1237117.2157107.257791.1299
Yale29.933726.277830.169332.584063.135131.901936.0841
Colon25.174623.500725.983627.615456.611728.973133.7010
SRBCT27.509525.735028.767630.223558.370329.730335.1542
warpPIE10P60.898144.659151.363482.9210114.685071.075169.6742
Lung31.171729.077031.949946.042863.275138.877939.8671
GLIOMA25.690528.572028.234629.003155.668129.497734.5528
9_Tumors41.254646.890944.921152.071683.929545.178751.3264
Leukemia43.170956.448949.695059.795993.752058.422759.9864
Nci953.104871.814064.511875.3016113.175369.099574.5359
Arcene81.580595.988593.7707177.8997155.1782125.6167103.2747
orlraws10P26.725748.830337.140841.025758.547138.024837.9534
Prostate_Tumor28.944650.474838.337351.300363.767940.216241.2868
11_Tumors78.081585.413272.7280110.5995157.134480.064078.6899
Lung_Cancer90.6418117.6249105.0547234.5773181.9730140.6153110.4799
GLI_8563.3169160.145297.7187123.5128122.643596.868284.6735
Ave 54.953559.386154.890083.8468109.512167.025765.8194
Rank 1326754
Table 7. Each optimization algorithm’s parameter settings.
Table 7. Each optimization algorithm’s parameter settings.
AlgorithmParameter Settings
MPSOC1 and C2 is 2, r1 and r2 ~ U (0,1), wmin is 0.4, wmax is 0.9
HGWOPThe maximum and minimum values of a are 2 and 0, Crmax is 1, Crmin is 0
HGSOK = α = β = 1, l1, l2 and l3 is 5 × 10−3, 1 × 102 and 1 × 10−2, M1 is 0.1, M2 is 0.2,
DSSAThe maximum number of iterations of LSA is 10, c2 and c3 are chosen randomly in [0, 1]
BGOAThe maximum and minimum values of c are 1 and 0.00004, r linearly decreases from 0.9 to 0
HBBOGa0 = 2, I = 1, E = 1
MSWOAThe value of a ranges from 2 to 0, b is 1, γ = 0.8
Table 8. The average accuracy of seven algorithms on 20 datasets.
Table 8. The average accuracy of seven algorithms on 20 datasets.
DatasetCriteriaMPSOHGWOPHGSODSSABGOAHBBOGMSWOA
SonarMean95.9350 (≈)96.7480 (≈)94.3089 (+)96.0163 (≈)94.7967 (+)96.1789 (≈)96.6667
Std2.02442.12011.70541.60351.37011.85571.4747
SemionMean99.0461 (≈)99.2348 (≈)98.6478 (+)99.0776 (≈)98.7107 (+)99.0566 (≈)99.1300
Std0.22210.26540.21710.31380.24830.21480.3781
Lung_discreteMean93.0952 (+)92.1429 (+)93.5714 (+)92.8571 (+)92.8571 (+)92.6190 (+)96.6667
Std2.24622.14292.14291.84432.60822.24623.5635
IsoletMean89.6368 (+)90.6090 (≈)88.6004 (+)89.8825 (+)89.2735 (+)89.3697 (+)90.3526
Std0.67910.91090.45840.80950.66070.83610.6842
PDCMean77.3951 (+)75.6291 (+)82.2958 (+)75.7174 (+)81.5453 (+)76.5563 (+)83.9073
Std3.11442.00431.32381.35002.06421.99122.1436
YaleMean65.4545 (+)65.8586 (+)64.9495 (+)65.4545 (+)65.7576 (+)65.5556 (+)68.4848
Std2.01012.81382.43472.42423.59832.14043.7196
ColonMean95.2778 (+)91.1111 (+)98.0556 (≈)91.6667 (+)96.3889 (+)91.6667 (+)99.1667
Std4.12952.99183.52460.00004.12950.00002.5000
SRBCTMean99.7917 (≈)98.5417 (+)100 (≈)97.7083 (+)99.3750 (+)97.2917 (+)100
Std1.12192.643503.01181.87503.09710
warpPIE10PMean93.3333 (+)93.3333 (+)93.4127 (≈)93.7302 (+)93.1746 (+)93.3333 (+)94.127
Std1.42860.95241.17981.14741.18781.13361.3375
LungMean97.5000 (+)97.5000 (+)97.5833 (≈)97.5000 (+)97.7500 (≈)97.5000 (+)97.6667
Std000.448800.750000.6236
GLIOMAMean83.3333 (+)80.0000 (+)87.0000 (+)80.0000 (+)86.0000 (+)80.0000 (+)91.3333
Std4.71400.00005.25990.00005.53770.00004.2687
9_TumorsMean66.3889 (+)65.0000 (+)65.5556 (+)63.6111 (+)65.5556 (+)65.0000 (+)70.0000
Std5.03845.85314.15744.01585.15203.96754.0825
LeukemiaMean93.3333 (+)91.4286 (+)98.0952 (+)92.1429 (+)96.4286 (+)92.1429 (+)100
Std4.09634.28573.15872.14294.01952.14290
Nci9Mean61.3889 (+)57.7778 (+)66.1111 (+)58.0556 (+)66.6667 (+)57.7778 (+)74.7222
Std6.26703.68514.77912.62067.45363.68516.6260
ArceneMean95.4167 (+)95.6667 (+)96.1667 (≈)95.9167 (+)95.7500 (≈)95.8333 (≈)96.5833
Std1.45541.10551.40441.20471.60081.34371.3668
orlraws10PMean97.5000 (+)93.5000 (+)99.6667 (≈)93.1667 (+)98.0000 (+)93.5000 (+)100
Std2.81372.29131.24722.40952.76892.29130
Prostate_TumorMean96.8333 (+)97.1667 (+)97.5000 (+)96.3333 (+)96.0000 (+)95.1667 (+)98.6667
Std3.28722.79382.50002.56043.00003.53162.2111
11_TumorsMean91.3725 (+)92.3529 (+)91.7647 (+)91.9608 (+)91.2745 (+)92.1569 (+)93.4314
Std1.30061.44091.40031.99961.41731.38651.4575
Lung_CancerMean97.0000 (≈)97.0000 (≈)97.4167 (≈)96.6667 (≈)97.2500 (≈)97.2500 (≈)97.3333
Std1.00001.00000.44881.34370.98950.75000.8975
GLI_85Mean93.7255 (+)92.7451 (+)99.8039 (≈)92.5490 (+)100 (≈)92.5490 (+)100
Std3.99922.91491.05593.373503.37350
1st/2nd 0/23/22/90/12/10/015/5
+/−/≈ 16/0/416/0/412/0/817/0/316/0/416/0/4
Table 9. The average number of selected features of seven methods on 20 datasets.
Table 9. The average number of selected features of seven methods on 20 datasets.
DatasetFullMPSOHGWOPHGSODSSABGOAHBBOGMSWOA
Sonar6032.400026.466720.466726.566741.400026.466718.6667
Semion265148.6333118.6000115.9333123.2000156.2667124.366796.2667
Lung_discrete32595.3000132.000026.7667135.8667109.3667133.866726.7333
Isolet617394.9667303.5000264.6667304.7000415.4000303.9333248.6333
PDC756337.1333363.300028.5667373.800025.8667365.600080.1667
Yale1024522.8000491.7333217.8667503.8000411.1667494.7333197.1000
Colon2000247.7667895.633320.3000928.9333138.4000879.733317.3667
SRBCT2308500.66671068.333351.50001093.0667190.83331015.633331.7667
warpPIE10P24201229.73331178.2000380.83331185.4333838.90001181.3333426.9000
Lung3312458.76671500.866798.40001550.4000547.66671439.733353.7333
GLIOMA4434188.70001955.533345.96672042.800031.23331806.166753.0667
9_Tumors57262242.30002763.1000542.63332804.83331186.16672721.3667781.9667
Leukemia70701820.93333416.366782.26673464.1000173.10003392.600027.5000
Nci997122244.66674716.4333143.76674795.6333498.76674739.6000135.5000
Arcene10,0002874.40004783.60001113.50004914.10001526.06674774.9333600.7333
orlraws10P10,304983.86674942.866779.56675037.4333113.73334846.633336.0000
Prostate_Tumor10,5093066.53335148.5667593.10005209.8333855.46675109.8000725.6667
11_Tumors12,5335077.16676126.40003234.70006209.13334939.63336056.00002443.2333
Lung_Cancer12,6004629.50006030.86671170.13336182.43334933.30005834.46671091.5333
GLI_8522,2835374.133310,916.9667205.566711,047.0000444.333310,751.2333168.7667
Ave61011620.34042847.5277420.95982900.6781876.90852802.1370362.2713
Rank 4627351
Table 10. The falls and ADLs of the SFall dataset.
Table 10. The falls and ADLs of the SFall dataset.
No.Cat.DescriptionAcquisition Duration (Second)
1ADLWalking slowly10
2ADLJogging10
3ADLSlowly sit down and stand up again10
4ADLBending slowly from standing and stand straight again10
5ADLSlowly squat and then stand up10
6ADLLie down slowly and then stand up10
7ADLWalking upstairs and downstairs slowly10
8ADLMix of daily activities10
9FallFall forward while walking
10FallFall backward while walking
11FallFall left while walking
12FallFall right while walking
13FallVertical fall while walking
14FallFall while jogging
Table 11. Structure of the confusion matrix.
Table 11. Structure of the confusion matrix.
Confusion MatrixPrediction
Falls (P)ADLs (N)
Reference Falls (P)TPFN
ADLs (N)FPTN
Table 12. Recognition results of the four algorithms on the SFall dataset.
Table 12. Recognition results of the four algorithms on the SFall dataset.
AlgorithmsFS (%)TPFNFPFN+FPMean Accuracy
MSWOA89.656712.21.61.22.881.33%
SASS88.119811.10.52.42.979.29%
COLSHADE80.61839.40.14.24.368.61%
sCMAgES76.431211.54.92.27.161.83%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, H.; Yan, B.; Dong, L.; Yuan, X. Multipopulation Whale Optimization-Based Feature Selection Algorithm and Its Application in Human Fall Detection Using Inertial Measurement Unit Sensors. Sensors 2024, 24, 7879. https://doi.org/10.3390/s24247879

AMA Style

Cao H, Yan B, Dong L, Yuan X. Multipopulation Whale Optimization-Based Feature Selection Algorithm and Its Application in Human Fall Detection Using Inertial Measurement Unit Sensors. Sensors. 2024; 24(24):7879. https://doi.org/10.3390/s24247879

Chicago/Turabian Style

Cao, Haolin, Bingshuo Yan, Lin Dong, and Xianfeng Yuan. 2024. "Multipopulation Whale Optimization-Based Feature Selection Algorithm and Its Application in Human Fall Detection Using Inertial Measurement Unit Sensors" Sensors 24, no. 24: 7879. https://doi.org/10.3390/s24247879

APA Style

Cao, H., Yan, B., Dong, L., & Yuan, X. (2024). Multipopulation Whale Optimization-Based Feature Selection Algorithm and Its Application in Human Fall Detection Using Inertial Measurement Unit Sensors. Sensors, 24(24), 7879. https://doi.org/10.3390/s24247879

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop