Next Article in Journal
A Comprehensive Review of Computational and Experimental Studies on Skin Mechanics and Meshing: Discrepancies, Challenges, and Optimization Strategies
Previous Article in Journal
Optimization of Actuator Stiffness and Actuation Timing of a Passive Ankle Exoskeleton: A Case Study Using a Musculoskeletal Modeling Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IHBOFS: A Biomimetics-Inspired Hybrid Breeding Optimization Algorithm for High-Dimensional Feature Selection

1
School of Computer Science, Hubei University of Technology, No. 28 Nanli Road, Hongshan District, Wuhan 430068, China
2
Hubei Provincial Key Laboratory of Green Intelligent Computing Power Network, No. 28 Nanli Road, Hongshan District, Wuhan 430068, China
*
Author to whom correspondence should be addressed.
Biomimetics 2026, 11(1), 3; https://doi.org/10.3390/biomimetics11010003
Submission received: 12 November 2025 / Revised: 27 November 2025 / Accepted: 16 December 2025 / Published: 22 December 2025
(This article belongs to the Section Biological Optimisation and Management)

Abstract

With the explosive growth of data across various fields, effective data preprocessing has become increasingly critical. Evolutionary and swarm intelligence algorithms have shown considerable potential in feature selection. However, their performance often deteriorates in large-scale problems, due to premature convergence and limited exploration ability. To address these limitations, this paper proposes an algorithm named IHBOFS, a biomimetics-inspired optimization framework that integrates multiple adaptive strategies to enhance performance and stability. The introduction of the Good Point Set and Elite Opposition-Based Learning mechanisms provides the population with a well-distributed and diverse initialization. Furthermore, adaptive exploitation–exploration balancing strategies are designed for each subpopulation, effectively mitigating premature convergence. Extensive ablation studies on the CEC2022 benchmark functions verify the effectiveness of these strategies. Considering the discrete nature of feature selection, IHBOFS is further extended with continuous-to-discrete mapping functions and applied to six real-world datasets. Comparative experiments against nine metaheuristic-based methods, including Harris Hawk Optimization (HHO) and Ant Colony Optimization (ACO), demonstrate that IHBOFS achieves an average classification accuracy of 92.57%, confirming its superiority and robustness in high-dimensional feature selection tasks.

1. Introduction

With the rapid expansion of dataset dimensions, high-dimensional data pose numerous challenges. In such spaces, data become sparse, making it difficult for models to learn effective patterns, which reduces the generalization capability [1]. When models contain excessive or redundant features, they tend to capture irrelevant patterns from the training samples, which in turn degrades their generalization ability on new data [2]. Moreover, as the dimensionality increases, the number of potential feature combinations grows exponentially, causing traditional search-based selection procedures to become computationally prohibitive [3]. Therefore, recent research has increasingly turned toward adaptive and scalable feature reduction techniques designed for complex high-dimensional datasets [4].
Feature selection (FS) serves as a fundamental procedure in the fields of machine learning and data analytics. Its objective is to extract the most informative subset of attributes from the full feature set, thereby enhancing model efficiency, reducing the computational burden, and improving the interpretability of learning outcomes [5]. Broadly, FS methods can be divided into three major categories: filter-based, wrapper-based, and embedded approaches. The filter-based category assesses feature importance without relying on any particular learning algorithm, often through statistical relevance measures such as correlation analysis, chi-square evaluation, or information gain [6]. Their low computational overhead makes them an effective choice for preprocessing extensive datasets [7]. In wrapper approaches, subsets of features are examined using a designated learning algorithm, training and assessing models to select the optimal subset, though incurring high computational cost [8], with popular approaches like recursive feature elimination (RFE) [9]. Embedded methods integrate FS within the model training procedure by tuning model parameters, allowing for an efficient and effective identification of relevant features [10], and are generally more computationally efficient, often yielding better model performance compared to wrapper methods [11].
Hybrid approaches integrate the advantages of both filter and wrapper techniques. Typically, a preliminary filter step is applied to limit the dimensionality, followed by a wrapper method to refine the optimal feature subset [12]. For example, a recent hybrid feature selection method combines ReliefF and fuzzy entropy with an enhanced equilibrium optimizer and achieves superior performance on medical datasets [13].
Despite the effectiveness of traditional and hybrid feature selection strategies, they often encounter limitations when applied to large-scale or highly complex datasets, particularly owing to the combinatorial complexity of feature subset selection. As the dimensionality of the data increases, exhaustive search becomes computationally infeasible, and heuristic-based methods may fall into local optima. To cope with this challenge, increasing attention has been devoted to evolutionary computation methods, which demonstrate remarkable global exploration ability and adaptability when tackling high-dimensional and nonlinear optimization tasks.
Evolutionary algorithms such as Grey Wolf Optimizer (GWO) [14], Genetic Algorithm (GA) [15], and Particle Swarm Optimization (PSO) [16] simulate natural evolution or swarm intelligence to explore high-quality solutions within challenging optimization landscapes. However, standard metaheuristics tend to converge slowly, with low efficiency in high-dimensional settings. As a result, researchers have considered various approaches to improve the algorithm in order to enhance the feature selection performance [17]. An improved Crayfish Optimization Algorithm (MCOA) was proposed to lead the exploration toward optimal results using an environmental update mechanism [18]. A self-adaptive weighted differential evolution algorithm (SaWDE) was proposed to tackle large-scale feature selection tasks and showed remarkable effectiveness across diverse high-dimensional datasets [19]. The Feature Weighting Particle Swarm Optimization method (FWPSO) effectively identifies biomarker genes within complex microarray data by integrating feature relevance assessment with optimized feature selection [20]. A hybrid Sine Cosine–Firehawk Algorithm (HSCFHA) was designed to perform feature selection by optimizing dataset variance [21]. In the past four years, other methods have been proposed, including VGS-MOEA [22], MTPSO [23], PSO-EMT [24], and DENCA [25].
Despite the effectiveness demonstrated by existing metaheuristic-based feature selection methods, their inherent random strategies, while ensuring global search capability, struggle to completely avoid local optima. Therefore, developing a computationally efficient high-dimensional feature selection method that exhibits good consistency and robustness remains a significant direction in current research. Ye et al. inspired by the theory of hybrid vigor and hybrid breeding mechanisms, proposed a hybrid breeding optimization algorithm [26]. HBO exhibits outstanding performance in exploration, search efficiency, and adaptability and has been employed to address various optimization challenges, including the 0–1 knapsack problem [27]. Despite its novel structure, standard HBO suffers from limited adaptability and may converge prematurely. To overcome these issues, several improved variants have been proposed. For instance, a Cooperative Hybrid Breeding Swarm Intelligence (CHBSI) algorithm was proposed to address the challenges of FS in high-dimensional spaces [28]. Furthermore, a Double-Stage Multimodal HBO (DSMHBO), integrating dynamic niching, neighborhood search, and elite mutation, was proposed to locate multiple optima effectively [29]. In the domain of network security, Ye et al. developed a Cooperative Co-evolution Improved HBO (CCIHBO) framework [30], effectively enhancing intrusion detection accuracy. These improvements significantly broaden the applicability and capability of the original HBO to handle intricate optimization problems involving high-dimensional search spaces.
The performance of HBO has already been validated on function optimization and engineering optimization problems. However, HBO and its improved variants face challenges such as low convergence accuracy and insufficient robustness in high-dimensional FS scenarios. Therefore, this paper focuses on HBO research, with the objective of thoroughly analyzing its potential and limitations in high-dimensional feature selection problems. We propose IHBOFS, which enhances the original HBO framework to better handle high-dimensional feature selection tasks.
The main findings highlighted in this work are as follows:
  • An integrated multi-strategy improved hybrid breeding optimization algorithm is proposed, which effectively balances the exploration and exploitation capabilities.
  • Ablation experiments are conducted on various types of benchmark functions to verify the effectiveness of the different enhancement strategies incorporated into the improved algorithm.
  • The optimal combination of classifier and transfer function is identified based on experimental evaluations of the performance of different options.
  • The performance of IHBOFS is evaluated against various metaheuristic-based FS methods on high-dimensional datasets to validate its effectiveness.
The structure of this paper is as follows: Section 2 provides a review of recent related studies, presenting the improvement strategies and their corresponding mathematical models. Section 3 introduces the proposed IHBOFS. Experimental findings and corresponding discussions are detailed in Section 4. Finally, Section 5 concludes the paper and outlines potential directions for future research.

2. Related Works

In recent years, numerous evolutionary algorithms have been proposed to address the challenges of high-dimensional feature selection, and attempts have been made to improve algorithm performance from various perspectives [31]. Table 1 summarizes the performance and limitations of several nascent hybrid algorithms tailored for feature selection tasks in various research fields. Among them, the HBO has shown strong potential due to its biological inspiration and excellent global search ability. As this work builds upon HBO as a baseline, this section first introduces the fundamental principles of the HBO and then presents recent improvements and related methods relevant to high-dimensional feature selection.

2.1. Hybrid Breeding Optimization Algorithm

To address the inefficiency of FS, Ye et al. [26] proposed the HBO, which exhibits strong exploratory capability and algorithm efficiency. During the initial stage, all individuals are sorted according to their fitness values, forming the population: X = { X 1 , X 2 , , X n } , where n represents the population size. The maintainer line, which consists of individuals with the highest fitness values, is defined as M = { X 1 , X 2 , X 3 , , X m } ,   m = n / 3 . The sterile line is composed of individuals exhibiting the poorest fitness X s = { X 2 m + 1 , X 2 m + 2 , , X n } , while the remaining individuals make up the restorer line X r = { X m , X m + 1 , , X 2 m } .
Hybridization: The sterile individuals are updated during this step. The process of generating new individuals through hybridization is given by Equation (1),
X s ( i ) ( t + 1 ) = k 1 · X s ( j ) ( t ) + ( 1 k 1 ) · X m ( t ) ,
where X s ( i ) ( t + 1 ) represents a new sterile-line individual generated in iteration t, and X m ( t ) and X s ( j ) ( t ) represent individuals randomly drawn from the maintainer and sterile lines, respectively. While k 1 denotes a uniformly distributed random variable within [ 0 , 1 ] , the subsequent parameters k 2 and k 3 share the same distribution and are utilized in later stages of the algorithm.
Selfing: In this step, the individuals in the restorer line integrate genetic information from other individuals to facilitate the evolution of the subpopulation towards an optimal solution. This process is formulated as Equation (2),
X r ( i ) ( t + 1 ) = k 2 · ( X best ( t ) X r ( j ) ( t ) ) + X r ( i ) ( t ) .
In this equation, X r ( i ) ( t + 1 ) refers to a newly generated individual obtained through self-fertilization between two restorer individuals i and j. The element X best ( t ) represents the current best individual in iteration t, whereas X r ( j ) ( t ) represents another restorer individual randomly chosen from the population.
Renewal: If a restorer-line individual fails to update for a predefined number of consecutive iterations ( SC max ), the algorithm resets it by randomly selecting new values from the search space. This mechanism is described as Equation (3),
X r ( i ) ( t + 1 ) = k 3 · ( V max V min ) + X r ( i ) ( t ) + V min ,
where X r ( i ) ( t ) represents the restorer individual that has not been updated, and V max and V min are the maximum and minimum values of the search space.

2.2. Good Point Set

The Good Point Set (GPS) method, proposed by the Chinese mathematician Luogeng Hua, aims to generate a set of uniformly distributed points within the search space, which helps establish a balanced coverage of the search domain during initialization [41]. Its principle is as follows:
Suppose G D is the unit cube in the D-dimensional Euclidean space; then, P n ( k ) = ( { r 1 n · k } , { r 2 n · k } , , { r D n · k } ) is a Good Point Set, where 1 k n and r G D . The discrepancy is φ ( n ) = C ( r , ε ) n 1 + ε , and C ( r , ε ) is a constant that only depends on r and ε ( ε is an arbitrarily small positive number), r is the good point, { r s n · k } represents taking the fractional part, and n represents the number of points. In this paper, we take r = 2 cos ( 2 k π p ) , where 1 k D , and p is the smallest prime number satisfying p 3 2 D . The initialization mapping is defined in Equation (4),
X i d = ( UB d LB d ) × { r d i × k } + LB d ,
where LB d and UB d represent the lower and upper bounds of the d-th dimension in the search space, respectively. Figure 1 presents the two-dimensional initial population distributions generated by four different methods—logistic chaotic mapping, Good Point Set initialization, Gaussian perturbation, and uniform random initialization. As depicted, for relatively small population sizes, the Good Point Set-based initialization method significantly improves the uniformity of population distribution. It improves the coverage of the search space, reducing the sensitivity to the initial population distribution.

2.3. Elite Opposition-Based Learning (EOBL)

The core idea of EOBL is to exploit the symmetry of the problem space by mirroring each solution in the representation space. The principle is as follows:
Consider an elite individual X i = ( x i , 1 , x i , 2 , , x i , D ) , where i = 1 , 2 , , n , in a D-dimensional search space. Its opposite solution is defined as Equations (5) and (6):
X ˜ i = ( x ˜ i , 1 , x ˜ i , 2 , , x ˜ i , D ) ,
x ˜ i , j = K LB j + UB j x i , j , j = 1 , 2 , , D ,
where K is a dynamic coefficient taking values in [0, 1], and L B j and U B j denote the lower and upper bounds of the j-th dimension, respectively. When the computed opposite solution falls outside the search boundaries, it is adjusted according to a uniform distribution, as described in Equation (7).
x ˜ i , j = Uniform ( LB j , UB j )
The standard HBO mainly utilizes heterosis to update sterile lines and restorer lines, thereby achieving population evolution. However, the maintainer line lacks an effective update strategy, leading to underutilization. In addition, the original operators exhibit performance deficiencies in high-dimensional complex scenarios. With the aim of resolving these limitations, we designed four advanced strategies to optimize and improve the standard HBO.

3. The Proposed Method

A thorough explanation of the proposed method is provided in this section.

3.1. Integrated Multi-Strategy Improved HBO

The subsequent discussion offers a comprehensive introduction to the algorithm’s process and improvement measures.

3.1.1. Optimization of Initial Population Generation

Initial population generation is a crucial step in metaheuristic algorithms, directly affecting the search efficiency and accuracy. In the traditional HBO, random initialization is used. However, an excessively large population increases the computational burden, whereas a too-small population leads to uneven distribution, making it susceptible to becoming trapped in local optima, particularly for complex high-dimensional problems. Therefore, this paper introduces the GPS and EOBL to optimize the initial population generation. The GPS-initialized population is used to construct elite opposition solutions, from which the best individuals are selected as the final initial solution set.

3.1.2. Operations for the Maintainer Line

The sterile line holds significant potential. When guided in evolution by maintainer lines with currently high fitness values, there is a considerable probability of generating superior individuals. However, due to the low stability of the sterile line, strategy adjustments are required at different stages of the iteration process. Therefore, three refined variants of differential operators are introduced in this paper, each designed to operate at different stages of the search process. These operators, denoted as D E 1 , D E 2 , and D E 3 , are mathematically defined in the following Equations (8)–(10).
DE 1 : X i ( t + 1 ) = X i ( t ) + r 1 · ( X m ( i ) ( t ) X m ( j ) ( t ) ) + ( 1 r 1 ) · ( X m ( k ) ( t ) X i ( t ) ) ,
DE 2 : X i ( t + 1 ) = X i ( t ) + F · ( X m ( i ) ( t ) X m ( j ) ( t ) ) + F · ( X best ( t ) X i ( t ) ) ,
DE 3 : X i ( t + 1 ) = X best ( t ) + r 2 · ( X m ( i ) ( t ) X m ( j ) ( t ) ) + ( 1 r 2 ) · ( X m ( k ) ( t ) X m ( l ) ( t ) ) ,
where X best ( t ) indicates the optimal solution obtained at iteration t, and X m ( · ) refers to individuals randomly drawn from the maintainer line, with all indices being distinct. The updated value of the i-th maintainer line individual at iteration t + 1 is represented by X i ( t + 1 ) . The smoothing factor F controls the transition from a global to a local search, and the random coefficients r 1 and r 2 are independently generated within the interval [0, 1].
To allow different differential operators to be applied with varying probabilities during different phases of optimization, the following strategy is implemented: at the initial stage, the global differential operator DE 1 is assigned a greater likelihood of selection, whereas during the the global-to-local search transition, the transition operator DE 2 is preferred. At the later stage, the local differential operator DE 3 is preferred to further refine the optimal solution. An adaptive algorithm is proposed to dynamically determine the selection probabilities corresponding to the three differential operators, while utilizing a roulette wheel selection algorithm for operator selection, as shown in Equations (11)–(17):
s 1 = k · 1 1 + e ( t T / 6 ) / ( 0.04 T ) + s min ,
S 2 ( t ) = k · e ( t T / 2 ) 2 / ( 10 · T ) + S min ,
S 3 ( t ) = k · 1 1 + e ( t 5 T / 6 ) / ( 0.04 T ) + S min ,
k = s max s min ,
S = s 1 + s 2 + s 3 ,
p i = s i S , i = 1 , 2 , 3 ,
DE s = RW _ Select ( p 1 , p 2 , p 3 ) ,
where s min and s max are parameters to constrain the lower and upper bounds before normalization. P i denotes the probability for selecting each differential operator after normalization. Here, t and T denote the current and maximum iteration counts, respectively. D E s corresponds to the differential operator finally selected via the roulette wheel strategy.
As shown in Figure 2, the algorithm adjusts the selection probabilities of each operator as the iteration progresses.

3.1.3. Operations for the Hybridization Phase

To strengthen the hybridization process, a t-distribution mutation-based perturbation method is adopted. The modification of the HBO crossover phase is expressed in Equations (18)–(21):
X s ( i ) ( t + 1 ) = t · X s ( j ) ( t ) + ( 1 t ) · X m ( t ) , i j ,
f ( t | d f , σ ) = Γ d f + 1 2 Γ d f 2 d f π σ 1 + 1 d f t σ 2 d f + 1 2 ,
d f = 2 + t T · 28 ,
σ = ( σ max σ min ) · 1 t T g r + σ min .
The original random number r is replaced by a value selected through a random sampling strategy based on the t-distribution. Γ denotes the Gamma function, while σ denotes the mutation scaling factor. This strategy enables broader exploration of the search space during the early iterations, while favoring more focused local search in the later stages.

3.1.4. Operations for the Selfing Phase

During the selfing phase, restorer individuals are updated. When an individual has reached the maximum selfing limit S C but fails to update the global optimal individual, this suggests that the algorithm has converged to a local optimum, thus requiring a reset.
This paper refines S C by allowing it to vary adaptively throughout the iterative process, as depicted in Equation (22).
S C = SC min + ( SC max SC min ) · 1 t T 2 .
During the algorithm’s initialization and global exploration, most individuals are highly likely to make productive moves within a short iteration cycle. Accordingly, the S C parameter is initialized with a comparatively high value. When an individual meets this criterion, this suggests premature convergence and triggers a reset. As the search advances, the chance of stagnation in local optima grows; so, a smaller S C is used to accelerate escape from such regions.
In addition, this paper also improves the selfing strategy for restorer individuals. The gene update formula for restorer individuals is replaced by Equations (23)–(25),
x r ( i ) ( t + 1 ) = r 3 · X best X r ( j ) + c 1 · ( L B + L ) ,
L = α · L é v y ( β ) · ( U B L B ) ,
c 1 = 2 · exp 4 t T 2 ,
where α serves as the step-length scaling coefficient, and c 1 is an adaptively varying factor that gradually decreases from 2 toward 0 in a nonlinear manner as the iteration proceeds. x r ( j ) denotes a restorer individual randomly chosen from the population ( j i ), while r 3 is a uniformly distributed random variable within [0, 1]. The term Lévy(β) follows a Lévy flight distribution characterized by the parameter β whose heavy-tailed nature allows the algorithm to perform frequent local movements interspersed with occasional long jumps, thereby improving both global exploration and local refinement capabilities [42]. The explicit mathematical form of this distribution is presented in Equation (26).
L é v y ( β ) t μ , 1 < β 3 .
The flowchart of the IHBOFS is shown in Figure 3.

3.2. IHBOFS Method

The proposed feature selection method, based on IHBOFS, follows the workflow outlined below. Initially, the algorithm parameters are set, and the search space is defined according to the feature count in the dataset. Next, cross-validation is applied to partition the data. IHBOFS is then utilized on the training set to identify the optimal solution. Although HBO is designed for continuous space optimization, FS is inherently a discrete problem. Thus, a binary encoding technique is employed to adapt the original algorithm for continuous domains to a form that can address discrete challenges. To evaluate the performance of the selected feature subset, an SVM classifier is used, and a fitness function is calculated to assess the subset’s quality by balancing the classification accuracy and feature minimization.

3.2.1. Binary Encoding

Through experimental comparison among eight transfer functions, we selected the most suitable one for the current scenario, as shown in Equation (27):
S ( x ) = 1 1 + exp ( x / 2 ) .
The continuous values are mapped into the [ 0 , 1 ] interval through the transfer function. Then, the transformed value is converted into a binary solution representation using Equation (28):
X i = 1 , if T ( X i ) > rand 0 , otherwise .
Here, “1” indicates that the corresponding feature is selected for classifier training, while “0” means the feature is not selected. T ( · ) denotes the transfer function, and rand is a random value within the range [ 0 , 1 ] .

3.2.2. Fitness Function

The fitness function in feature selection must balance two conflicting objectives: classification performance and subset dimensionality. Relying solely on accuracy tends to produce oversized feature subsets, whereas minimizing the number of features alone often results in severe information loss and degraded predictive performance. To select as few features as possible while improving the accuracy, the fitness function designed in this paper is defined as follows:
Fitness = λ · ( 1 acc ) + μ · n N ,
where Fitness represents the fitness value, acc is the classification accuracy, and n and N represent the proportion of selected features to the total number of features, λ + μ = 1 . The term ( 1 acc ) ensures that maximizing the accuracy corresponds to minimizing the fitness value, which aligns with the minimization nature of evolutionary optimization. The ratio n N provides a scale-invariant measure of subset size, preventing bias toward datasets with different dimensionalities.

3.2.3. Complexity Analysis

With N as the population size, D as the dimension of the search space, and T as the iteration limit, the computational complexity for each operational stage is presented in Table 2.
Even though the time complexity for initializing the population in IHBOFS is roughly twice that of HBO, its effect on the overall time complexity remains minimal, as the population initialization occurs only once. IHBOFS incurs a moderate increase in time complexity, approximately one-third, due to the addition of a preservation sub-population update. However, this enhancement significantly improves the exploitation of high-quality individuals and enhances the population diversity. Notably, the time complexities of the population sorting and optimal solution updates remain consistent between IHBOFS and HBO. Therefore, the proposed improvement strategy maintains a relatively low additional computational cost while enhancing the algorithm’s search performance and population diversity, making it a relatively excellent improvement strategy.

4. Experiment Results and Discussion

The effectiveness of HBO has been demonstrated in function optimization and engineering problems. As a result, this paper concentrates on feature selection experiments, and three groups of experiments are organized:
  • Ablation Study: this analysis aims to evaluate the role and significance of each component within the proposed IHBOFS, and experiments are conducted using different combinations of strategies on the CEC2022 test functions.
  • Classifier and Transfer Function Selection: using high-dimensional real datasets from the Scikit feature selection repository, six classifiers (KNN, SVM, XGBoost, Decision Tree, Random Forest, Naive Bayes) are evaluated and selected based on performance; in addition, both standard HBO and IHBOFS are combined with eight different transfer functions to perform feature selection, in order to identify the best classifier and transfer function.
  • Comparison with Other Algorithms: based on the selected SVM classifier and the S3 transfer function, IHBOFS is compared with various feature selection algorithms based on metaheuristics across multiple high-dimensional datasets.
All algorithms in the experiments are implemented in Python 3.8.20. Experiments are conducted on a computer equipped with an AMD Ryzen 7 5800H CPU @ 3.2 GHz and 32 GB RAM, using Windows 10 as the operating system.

4.1. Datasets

The CEC2022 benchmark suite includes 1 unimodal function (F1), 4 basic multimodal functions (F2–F5), 3 hybrid functions (F6–F8), and 4 composite functions (F9–F12), with a dimensionality of 20. A detailed description is provided in Appendix A.
The high-dimensional datasets are obtained from the standard Scikit feature selection database, as shown in Table 3. These datasets contain between 1024 and 10,304 features and 100 to 400 samples. All are multi-class classification problems, each presenting unique challenges. Their diversity provides a robust basis for evaluating IHBOFS’s ability to identify optimal feature subsets across various complex scenarios.

4.2. Experimental Settings

Table 4 presents the key parameter settings of HBO and its four improved strategy variants. For equitable evaluation and to confirm the contribution of each enhancement mechanism, each algorithm is independently executed 30 times on each test function at different dimensions. The maximum number of iterations is set to 1000 for all algorithms.
To ensure fairness in the number of fitness evaluations, the population size is set to 40 for IHBOFS and HBO_DE and to 60 for standard HBO and the other three improved algorithms. Five evaluation metrics are employed to comprehensively assess the algorithms’ performance: best fitness value (Best), worst fitness value (Worst), mean fitness value (Mean), standard deviation of fitness values (Std), and average execution time (Time).
To further assess the performance of IHBOFS, it is compared with several metaheuristic-based algorithms, including GA, Flower Pollination Algorithm (FPA) [43], Sparrow Search Algorithm (SSA) [44], Whale Optimization Algorithm (WOA) [45], the Rime-Ice Optimization Algorithm (RIME) [46], JAYA [47], GWO, and Harris Hawk Optimization (HHO) [48]. The parameter settings of these algorithms are listed in Table 5. All algorithms use a population size of 40 and a maximum of 1000 iterations. Each test function is run independently 30 times to minimize the impact of randomness. The parameter settings and evaluation metrics of HBO and IHBOFS are consistent with those used in the ablation study.
In the classifier evaluation and selection experiment, all classifiers are independently run 10 times on each dataset. A five-fold cross-validation and stratified sampling strategy is employed to reduce overfitting.
Properly designed transfer mechanisms can improve the interpretability and controllability of the feature selection process [49]. This study selects eight transfer functions (four S-shaped and four V-shaped), whose details are provided in Appendix B.

4.3. Comparison of Experimental Results and Analysis

Detailed experimental results and analyses are presented in the following subsections.

4.3.1. Ablation Study

Based on the results shown in Table 6, integrating all four improvement strategies into IHBOFS yields clear and consistent advantages over both the original HBO and all individually enhanced variants. IHBOFS achieves superior performance in terms of both the Best and Mean metrics, while further reducing the standard deviation, demonstrating significantly enhanced stability across multiple runs. Overall, the integrated strategy effectively leverages the complementary strengths of each individual enhancement, resulting in an optimized exploration process across the three populations and substantially improving the algorithm’s overall problem-solving performance and robustness.

4.3.2. Classifier Performance Evaluation and Selection

Ten independent runs were conducted for each classifier–dataset combination to ensure the statistical robustness of the validation findings. A five-fold cross-validation combined with stratified sampling was applied to mitigate overfitting risks. Table 7 summarizes the average classification accuracy and training time over 10 runs for several classifiers, including KNN, SVM, XGBoost, DT, RF, and NB.
As shown in Table 7, SVM exhibits strong overall performance with an average accuracy of 83.70%, ranking second only to RF, which attained 83.86%. Notably, SVM achieved exceptional accuracy on the pixraw10P and orlraws10P datasets, reaching 97.00% and 96.00% respectively, significantly outperforming other classifiers. This highlights the superior ability of SVM to handle high-dimensional data. In contrast, NB performs poorly on certain datasets such as pixraw10P, with an accuracy of only 22.00%, indicating that it may not be suitable for specific scenarios. Moreover, NB exhibits limited generalization across datasets, with its accuracy consistently lagging behind that of SVM and RF.
To provide a more intuitive comparison, Figure 4 presents the bar chart of training times for different classifiers. From the training time perspective, KNN is the fastest among all classifiers, significantly outperforming the others. SVM combines high accuracy with favorable training speed, averaging 0.145 s per run. This is 87.52% faster than RF (1.1617 s on average) and also substantially lower than XGBoost (2.155 s). This advantage is particularly important in real-time or resource-constrained environments.
In conclusion, SVM demonstrates dual strengths in classification accuracy and operational efficiency. Although RF offers a slight edge in accuracy, its longer training time may become a constraint in practical applications. Therefore, for high-dimensional classification tasks where both speed and accuracy are critical, KNN and SVM are generally more effective options. This study will further evaluate the classification performance of KNN and SVM under different parameter settings.
This study primarily focuses on the dimensionality reduction capability of evolutionary algorithms and their discriminative power across diverse categories. Given the complexity of parameter combinations across different classifiers, only key parameters significantly impacting classification performance are locally tuned. This approach emphasizes the search performance and efficiency of the evolutionary algorithm rather than the classifier itself.
Figure 5 and Figure 6 illustrate the classification accuracy of KNN and SVM under different parameter settings. Specifically, for KNN, the number of neighbors K is tested at values of 4, 5, and 6. For SVM, the penalty parameter C is evaluated at 3, 3.5, and 4. These parameters strongly influence the model’s performance and fitting behavior. In KNN, K determines the number of neighbors considered during classification, affecting the smoothness and generalization. For SVM, C controls the decision boundary margin and the tolerance to misclassifications, thereby determining the model complexity. As observed from Figure 5, as the value of K increases, KNN’s accuracy tends to decrease across most datasets. This suggests that a smaller K may provide more concentrated information and thus higher accuracy. On the other hand, SVM exhibits improved performance as C increases. Notably, SVM with C = 4 (SVM4) shows superior accuracy on multiple datasets, especially on pixraw10P and orlraws10P, where it achieves 97.00% and 96.00%, respectively.
To prevent potential overfitting and to ensure generalization across datasets, this study adopts cross-validation and stratified sampling. Both strategies help reduce bias in the train–test split and provide a more reliable estimate of model performance. Moreover, since C in SVM determines the model complexity, a larger value may increase the risk of overfitting. However, the results in Table 7 show that using smaller C values does not improve the performance; instead, the accuracy decreases. Objectively, if the model were overfitting at C = 4 , reducing the value of C (e.g., to 3 or 3.5) would simplify the model and alleviate overfitting, thereby improving the accuracy. However, the opposite trend is observed, indicating that the model is not overfitting at C = 4 . Although larger C values may potentially yield higher accuracy, they also tend to enforce stricter fitting to the training samples, which could overshadow the feature-filtering capability of the metaheuristic algorithm. Therefore, excessively large C values are not further explored in this study.
With regard to the training time, the table shows that KNN and SVM demonstrate stable training times across different parameter settings. Overall, SVM4 exhibits a favorable trade-off, especially on complex and high-dimensional datasets. Its high accuracy and stability highlight strong modeling capability. Moreover, SVM shows better generalization than KNN across datasets. A reasonable training time further ensures its practicality. Based on these findings, SVM with C = 4 is selected as the final classifier configuration for the subsequent experiments.

4.3.3. Transfer Function Experiments

To identify the most suitable transfer function for the algorithm, this subsection evaluates the performance of the multi-strategy improved IHBOFS in combination with the eight transfer functions introduced in Section 4.2. Feature selection experiments are conducted using the optimal classifier (SVM) selected in the previous subsection. Table 8 presents the classification accuracy of IHBOFS when combined with different transfer functions. In the table, Best denotes the highest accuracy among 10 independent runs, Worst denotes the lowest accuracy, while Mean and Std represent the average accuracy and standard deviation, respectively.
According to the results, the feature subsets obtained using S-shaped transfer functions significantly outperform those obtained using V-shaped functions. Among all transfer functions, S3 yields the best average classification performance when combined with IHBOFS, achieving the highest mean accuracy on five out of the tested datasets.
The combinations IHBOFS+S1, IHBOFS+S2, and IHBOFS+S3 all achieve the highest Best score on the last three high-dimensional datasets. Notably, S3 further enhances the lower-bound performance of IHBOFS, as evidenced by its superior Worst scores compared to other transfer functions. As a result, IHBOFS+S3 outperforms all other combinations in terms of overall average accuracy.
For the first three datasets, IHBOFS+S3 performs best across almost all accuracy metrics. Although IHBOFS+S3 does not achieve the highest average accuracy on the warpAR10P dataset (87.69% for IHBOFS+S4), the difference between the two is less than 1.79%. Overall, S3 produces feature subsets with better classification performance.
Table 9 records the number of features selected by IHBOFS with different transfer functions. Although V-shaped functions sometimes select fewer features on certain datasets, their classification accuracy is significantly inferior to that of S-shaped functions. Therefore, when classification performance is prioritized, V-shaped functions are not suitable as the final choice.
Moreover, the IHBOFS+S3 combination achieves the lowest standard deviation on three datasets, indicating enhanced stability. It also demonstrates the ability to select smaller feature subsets with higher classification accuracy, especially on the latter three high-dimensional datasets, outperforming other S-shaped functions.

4.3.4. Comparison with Other Algorithms and Result Analysis

In the previous sections, we compared the performance of different classifiers and transfer functions and selected SVM and the S3 transfer function as the final settings for the classification model. Based on this configuration, this section compares IHBOFS with various metaheuristic-based feature selection algorithms, including HHO, the Ant Colony Optimization Algorithm (ACO) [50], Slime Mould Algorithm (SMA) [51], Artificial Bee Colony Algorithm (ABC) [52], Tree Growth Algorithm (TGA) [53], Henry Gas Solubility Optimization Algorithm (HGSO) [54], Emperor Penguin Optimizer Algorithm (EPO) [55], Manta Ray Foraging Optimization Algorithm (MRFO) [56], and the standard HBO.
As shown in Table 10, IHBOFS achieves higher feature subset accuracy compared to all other algorithms. The only exception is the pixraw10P dataset, where IHBOFS ranks second to MRFO in terms of the mean classification accuracy. In all other datasets, IHBOFS outperforms the other algorithms across Best, Worst, and Mean metrics.
In particular, HGSO produces relatively weak feature subsets. On the Yale dataset, IHBOFS surpasses HGSO by 14.65% in mean classification accuracy and by 13.19% and 8.70% on warpAR10P and warpPIE10P, respectively. It is worth emphasizing that, while the original HBO performs better than most algorithms on the first four datasets, it fails to effectively search for optimal feature subsets on the two higher-dimensional datasets—an issue addressed by the improved IHBOFS.
In terms of the number of selected features and runtime, the EPO achieves the best performance. However, the EPO tends to prioritize selecting fewer features while compromising the classification performance. For instance, on the Yale dataset, the EPO’s Best and Mean scores are 73.33% and 72.06%, respectively—10.75% and 11.19% lower than IHBOFS.
Figure 7 presents the convergence curves of the average fitness values for all comparison algorithms. As shown, IHBOFS only ranks second to MRFO on the pixraw10P dataset. For all other datasets, IHBOFS achieves the smallest fitness values, significantly outperforming other algorithms. Moreover, IHBOFS exhibits better initial fitness values, thanks to the use of the Good Point Set and Elite Opposition-Based Learning strategies for population initialization, which helps generate higher-quality candidate solutions and accelerates convergence toward the global optimum.
Figure 8 visually compares the runtime of different algorithms. Although the EPO has the shortest runtime, IHBOFS—despite integrating multiple advanced strategies—does not exhibit performance degradation compared to HBO. Instead, it shows improvement, as IHBOFS selects fewer and higher-quality feature subsets, thus accelerating the search process. On higher-dimensional datasets, IHBOFS also demonstrates higher efficiency compared to other algorithms, confirming its suitability for high-dimensional feature selection tasks.
Table 11 shows the algorithm ranking obtained using the Friedman non-parametric statistical test based on mean accuracy. A lower rank indicates better performance. As shown, IHBOFS achieves the best average rank of 1.1667, followed by MRFO. TGA and HGSO rank lowest, indicating their inability to identify optimal feature subsets in this context. The computed p-value is significantly lower than the standard significance threshold (0.05), supporting the conclusion that performance differences among the algorithms are statistically significant.

5. Conclusions

In this paper, an adaptive improvement scheme for the Hybrid Breeding Optimization algorithm is proposed to address the problem of high-dimensional feature selection. The main contributions of this study are summarized as follows: First, the performance bottlenecks of HBO in high-dimensional and complex optimization tasks are analyzed, and IHBOFS is developed. IHBOFS incorporates several innovative strategies in a targeted manner, and these improvements comprehensively enhance the search ability and robustness of the algorithm at different evolutionary stages. Ablation experiments are conducted on 12 benchmark functions of different types and dimensions, and IHBOFS achieves the optimal average performance on 10 of these functions with significant reductions in the objective function values. In addition, IHBOFS exhibits zero or near-zero variance on several functions, reflecting extremely strong stability.
For high-dimensional feature selection tasks, a wrapper framework based on IHBOFS is designed in this paper, which maps continuous search spaces to discrete ones by means of multiple transfer functions. Combined with classifier selection and local fine-tuning strategies, this framework further improves the quality of the selected feature subsets and simultaneously enhances the classification performance. Extensive experimental validations show that IHBOFS achieves the highest mean classification accuracy on all six datasets, with the average accuracy improvement ranging from 2.54% to 3.49% compared with the original HBO, and its performance on each dataset outperforms other state-of-the-art metaheuristic algorithms. In terms of solution robustness, IHBOFS also has lower or comparable standard deviations to the comparison algorithms, indicating more stable search behavior. Furthermore, the improvements in IHBOFS lead to a maximum reduction of 98.8% and 94.0% in the size of the selected feature subsets while still maintaining extremely high accuracy. IHBOFS also demonstrates competitiveness in terms of the computational time.
Despite the promising results achieved above, several directions are worthy of further exploration. First, future research can incorporate the stability of feature subsets into the design of the fitness function to realize the collaborative optimization of classification accuracy, subset size, and stability. Second, considering the challenge of label scarcity in practical applications, future work can explore semi-supervised or unsupervised feature selection methods to better utilize unlabeled or partially labeled data, thereby improving the adaptability and practical application value of the algorithm.

Author Contributions

Conceptualization, C.X.; methodology, J.Z. and C.X.; software, J.Z.; validation, J.Z. and W.Z.; formal analysis, C.X. and W.Z.; investigation, J.Z.; resources, C.X.; data curation, W.Z. and J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, C.X. and W.Z.; visualization, J.Z.; supervision, C.X.; project administration, W.Z.; funding acquisition, C.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (Grant Nos. 62376089, 62302153, 62302154, U23A20318), the Key Research and Development Program of Hubei Province, China (Grant No. 2023BEB024), and the Young and Middle-aged Scientific and Technological Innovation Team Plan in Higher Education Institutions in Hubei Province, China (Grant No. T2023007).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets employed in this research were obtained from the open-access UCI Machine Learning Repository (https://archive.ics.uci.edu/ (accessed on 15 March 2025)).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Description of the CEC2022 Benchmark Functions

The CEC2022 test suite is one of the latest benchmark suites widely used for evaluating the performance of optimization algorithms. It consists of 12 single-objective test functions with boundary constraints, covering four types: unimodal functions, multimodal functions, hybrid functions, and composite functions. The design objectives of these four categories of functions are described as follows:
  • Unimodal Functions
    Unimodal functions have only one global optimal solution and no other local optimal solutions in the search space. This category of functions is mainly used to test the convergence speed and accuracy of optimization algorithms, which is a core indicator to measure the basic optimization capability of algorithms.
  • Multimodal Functions
    Multimodal functions possess multiple local optimal solutions and one global optimal solution in the search space. They are designed to evaluate the global search capability of optimization algorithms and their ability to avoid falling into local optima. To achieve better performance on multimodal functions, algorithms need to balance exploration and exploitation in the search process, where exploration focuses on discovering new search regions, and exploitation refines the current promising regions.
  • Hybrid Functions
    Hybrid functions are usually composed of multiple basic test functions, integrating the characteristics of both unimodal and multimodal functions to create more complex and challenging optimization problems. Such functions may contain both easily optimizable regions and difficult-to-optimize regions simultaneously, requiring algorithms to have the ability to switch search strategies adaptively and perform effective searches in different types of search spaces.
  • Composition Functions
    Composition functions further extend the concept of hybrid functions by combining multiple hybrid functions or basic functions to construct extremely complex test environments. These functions typically have higher dimensions and more intricate search spaces. The core objective of composite functions is to provide a broader range of challenges to assess the adaptability and robustness of algorithms when facing various types of optimization problems, while also requiring algorithms to have stronger global optimization capabilities and higher computational efficiency to handle the increased complexity.
All benchmark functions share the same D-dimensional global search range:
x i [ 100 , 100 ] , i = 1 , , D .
Table A1 summarizes the 12 benchmark functions used in this study, including their IDs, types, names, and global optimum values.
Table A1. CEC2022 benchmark functions used in this study.
Table A1. CEC2022 benchmark functions used in this study.
TypeIDFunction NameOptimal Value
Unimodal FunctionF1Shifted and Full Rotated Zakharov Function300
Multimodal FunctionsF2Shifted and Full Rotated Rosenbrock’s Function400
F3Shifted and Full Rotated Expanded Schaffer’s f6 Function600
F4Shifted and Full Rotated Non-Continuous Rastrigin’s Function800
F5Shifted and Full Rotated Levy Function900
Hybrid FunctionsF6Hybrid Function 1 (N = 3)1800
F7Hybrid Function 2 (N = 6)2000
F8Hybrid Function 3 (N = 5)2200
Composition FunctionsF9Composition Function 1 (N = 3)2300
F10Composition Function 2 (N = 4)2400
F11Composition Function 3 (N = 5)2600
F12Composition Function 4 (N = 6)2700
Global search range: [ 100 , 100 ] D .
Unimodal Function (F1): tests global convergence capability without complex local minima.
Multimodal Functions (F2–F5): evaluate the ability to escape local minima while maintaining exploration.
Hybrid Functions (F6–F8): divide the decision vector into segments, each associated with different basic optimization functions.
Composition Functions (F9–F12): represent the most challenging landscapes by combining several base functions through weighted adaptive composition strategies.

Appendix B. Transfer Functions in Feature Selection

The feature selection problem is often regarded as a complex combinatorial optimization challenge. However, many metaheuristic algorithms, originally designed for continuous space problems, cannot be directly applied to discrete optimization problems such as feature selection. Therefore, how to effectively convert the search in the continuous space into decisions in the discrete space is a key challenge. Transfer functions act as a bridge in this conversion process: through transfer functions, algorithms can search freely in the continuous domain, while their solutions can be interpreted as the selection or exclusion of features.
Transfer functions allow the search capabilities and flexibility of metaheuristic algorithms to be preserved, enabling the algorithms to perform excellently on global optimization problems. They can be applied to various types of feature selection tasks, whether for text, image, or bioinformatics data. By rationally designing the conversion mechanism, the interpretability and controllability of the feature selection process can be improved. In this paper, eight transfer functions (four sigmoid-type and four V-type) are selected; detailed information on these transfer functions is shown in Table A2. Transfer functions ensure the consistency of the search agents’ positions in the continuous and discrete spaces. Specifically, sigmoid-type transfer functions tend to map large positive values in the continuous space to 1 with a high probability; in contrast, V-type transfer functions tend to map values with large absolute values in the continuous space to 1 with a high probability.
Table A2. Details of the S-shaped and V-shaped transfer functions.
Table A2. Details of the S-shaped and V-shaped transfer functions.
S-Shaped FunctionsV-Shaped Functions
Function SymbolTransfer FunctionFunction SymbolTransfer Function
S 1 S ( x ) = 1 1 + e x p ( 2 x ) V 1 V ( x ) = 2 π arctan ( π x 2 )
S 2 S ( x ) = 1 1 + e x p ( x ) V 2 V ( x ) = x 1 + x 2
S 3 S ( x ) = 1 1 + e x p ( x / 2 ) V 3 V ( x ) = tanh ( x )
S 4 S ( x ) = 1 1 + e x p ( x / 3 ) V 4 V ( x ) = 2 π 0 π / 2 x e x p ( s 2 ) d s

References

  1. Yang, G.; He, J.; Lan, X.; Li, T.; Fang, W. A fast dual-module hybrid high-dimensional feature selection algorithm. Inf. Sci. 2024, 681, 121185. [Google Scholar] [CrossRef]
  2. Asghar, N.; Khalil, U.; Ahmad, B.; Alshanbari, H.M.; Hamraz, M.; Ahmad, B.; Khan, D.M. Improved nonparametric survival prediction using CoxPH, Random Survival Forest & DeepHit Neural Network. BMC Med. Inform. Decis. Mak. 2024, 24, 120. [Google Scholar]
  3. Saraf, T.O.Q.; Fuad, N.; Taujuddin, N.S.A.M. Framework of meta-heuristic variable length searching for feature selection in high-dimensional data. Computers 2022, 12, 7. [Google Scholar] [CrossRef]
  4. Osama, S.; Shaban, H.; Ali, A.A. Gene reduction and machine learning algorithms for cancer classification based on microarray gene expression data: A comprehensive review. Expert Syst. Appl. 2023, 213, 118946. [Google Scholar] [CrossRef]
  5. Hancer, E.; Xue, B.; Zhang, M. A survey on feature selection approaches for clustering. Artif. Intell. Rev. 2020, 53, 4519–4545. [Google Scholar] [CrossRef]
  6. Zandvakili, A.; Mansouri, N.; Javidi, M.M. A new feature selection algorithm based on fuzzy-pathfinder optimization. Neural Comput. Appl. 2024, 36, 17585–17614. [Google Scholar] [CrossRef]
  7. Alsaeedi, A.H.; Al-Mahmood, H.H.R.; Alnaseri, Z.F.; Aziz, M.R.; Al-Shammary, D.; Ibaida, A.; Ahmed, K. Fractal feature selection model for enhancing high-dimensional biological problems. BMC Bioinform. 2024, 25, 12. [Google Scholar] [CrossRef] [PubMed]
  8. Nssibi, M.; Manita, G.; Korbaa, O. Advances in nature-inspired metaheuristic optimization for feature selection problem: A comprehensive survey. Comput. Sci. Rev. 2023, 49, 100559. [Google Scholar] [CrossRef]
  9. Mahawar, K.; Rattan, P. Empowering education: Harnessing ensemble machine learning approach and ACO-DT classifier for early student academic performance prediction. Educ. Inf. Technol. 2025, 30, 4639–4667. [Google Scholar] [CrossRef]
  10. Alsahaf, A.; Petkov, N.; Shenoy, V.; Azzopardi, G. A framework for feature selection through boosting. Expert Syst. Appl. 2022, 187, 115895. [Google Scholar] [CrossRef]
  11. Van Zyl, C.; Ye, X.; Naidoo, R. Harnessing eXplainable artificial intelligence for feature selection in time series energy forecasting: A comparative analysis of Grad-CAM and SHAP. Appl. Energy 2024, 353, 122079. [Google Scholar] [CrossRef]
  12. Ganjei, M.A.; Boostani, R. A hybrid feature selection scheme for high-dimensional data. Eng. Appl. Artif. Intell. 2022, 113, 104894. [Google Scholar] [CrossRef]
  13. Vommi, A.M.; Battula, T.K. A hybrid filter-wrapper feature selection using Fuzzy KNN based on Bonferroni mean for medical datasets classification: A COVID-19 case study. Expert Syst. Appl. 2023, 218, 119612. [Google Scholar] [CrossRef]
  14. Chang, D.; Rao, C.; Xiao, X.; Hu, F.; Goh, M. Multiple strategies based Grey Wolf Optimizer for feature selection in performance evaluation of open-ended funds. Swarm Evol. Comput. 2024, 86, 101518. [Google Scholar] [CrossRef]
  15. Fang, Y.; Yao, Y.; Lin, X.; Wang, J.; Zhai, H. A feature selection based on genetic algorithm for intrusion detection of industrial control systems. Comput. Secur. 2024, 139, 103675. [Google Scholar] [CrossRef]
  16. Gad, A.G. Particle swarm optimization algorithm and its applications: A systematic review. Arch. Comput. Methods Eng. 2022, 29, 2531–2561. [Google Scholar] [CrossRef]
  17. Mostafa, R.R.; Hussien, A.G.; Gaheen, M.A.; Ewees, A.A.; Hashim, F.A. AEOWOA: Hybridizing whale optimization algorithm with artificial ecosystem-based optimization for optimal feature selection and global optimization. Evol. Syst. 2024, 15, 1753–1785. [Google Scholar] [CrossRef]
  18. Jia, H.; Zhou, X.; Zhang, J.; Abualigah, L.; Yildiz, A.R.; Hussien, A.G. Modified crayfish optimization algorithm for solving multiple engineering application problems. Artif. Intell. Rev. 2024, 57, 127. [Google Scholar] [CrossRef]
  19. Wang, X.; Wang, Y.; Wong, K.C.; Li, X. A self-adaptive weighted differential evolution approach for large-scale feature selection. Knowl.-Based Syst. 2022, 235, 107633. [Google Scholar] [CrossRef]
  20. Wang, X.; Jia, W. A feature weighting particle swarm optimization method to identify biomarker genes. In Proceedings of the 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Las Vegas, NV, USA, 6–8 December 2022; pp. 830–834. [Google Scholar]
  21. Moosavi, S.K.R.; Saadat, A.; Abaid, Z.; Ni, W.; Li, K.; Guizani, M. Feature selection based on dataset variance optimization using hybrid sine cosine–firehawk algorithm (hscfha). Future Gener. Comput. Syst. 2024, 155, 272–286. [Google Scholar] [CrossRef]
  22. Cheng, F.; Cui, J.; Wang, Q.; Zhang, L. A variable granularity search-based multiobjective feature selection algorithm for high-dimensional data classification. IEEE Trans. Evol. Comput. 2022, 27, 266–280. [Google Scholar] [CrossRef]
  23. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. Evolutionary multitasking for feature selection in high-dimensional classification via particle swarm optimization. IEEE Trans. Evol. Comput. 2021, 26, 446–460. [Google Scholar] [CrossRef]
  24. Chen, K.; Xue, B.; Zhang, M.; Zhou, F. An evolutionary multitasking-based feature selection method for high-dimensional classification. IEEE Trans. Cybern. 2020, 52, 7172–7186. [Google Scholar] [CrossRef] [PubMed]
  25. Hancer, E.; Xue, B.; Zhang, M. An evolutionary filter approach to feature selection in classification for both single-and multi-objective scenarios. Knowl.-Based Syst. 2023, 280, 111008. [Google Scholar] [CrossRef]
  26. Ye, Z.; Ma, L.; Chen, H. A hybrid rice optimization algorithm. In Proceedings of the 2016 11th International Conference on Computer Science & Education (ICCSE), Nagoya, Japan, 23–25 August 2016; pp. 169–174. [Google Scholar]
  27. Shu, Z.; Ye, Z.; Zong, X.; Liu, S.; Zhang, D.; Wang, C.; Wang, M. A modified hybrid rice optimization algorithm for solving 0–1 knapsack problem. Appl. Intell. 2022, 52, 5751–5769. [Google Scholar] [CrossRef]
  28. Mei, M.; Zhang, S.; Ye, Z.; Wang, M.; Zhou, W.; Yang, J.; Zhang, J.; Yan, L.; Shen, J. A cooperative hybrid breeding swarm intelligence algorithm for feature selection. Pattern Recognit. 2025, 169, 111901. [Google Scholar] [CrossRef]
  29. Cai, T.; Ma, F.; Ye, Z.; Zhou, W.; Wang, M.; He, Q.; Pan, H.; Shen, J. Dynamic niche technology based hybrid breeding optimization algorithm for multimodal feature selection. Sci. Rep. 2025, 15, 8001. [Google Scholar] [CrossRef] [PubMed]
  30. Ye, Z.; Luo, J.; Zhou, W.; Wang, M.; He, Q. An ensemble framework with improved hybrid breeding optimization-based feature selection for intrusion detection. Future Gener. Comput. Syst. 2024, 151, 124–136. [Google Scholar] [CrossRef]
  31. Song, X.; Zhang, Y.; Zhang, W.; He, C.; Hu, Y.; Wang, J.; Gong, D. Evolutionary computation for feature selection in classification: A comprehensive survey of solutions, applications and challenges. Swarm Evol. Comput. 2024, 90, 101661. [Google Scholar] [CrossRef]
  32. Gao, J.; Wang, Z.; Jin, T.; Cheng, J.; Lei, Z.; Gao, S. Information gain ratio-based subfeature grouping empowers particle swarm optimization for feature selection. Knowl.-Based Syst. 2024, 286, 111380. [Google Scholar] [CrossRef]
  33. Zaman, E.A.K.; Ahmad, A.; Mohamed, A. Adaptive threshold optimisation for online feature selection using dynamic particle swarm optimisation in determining feature relevancy and redundancy. Appl. Soft Comput. 2024, 156, 111477. [Google Scholar] [CrossRef]
  34. Wang, T.; Zhao, Q.; Wang, S.; Jing, C. Research on Stochastic Feature Selection Optimization Algorithm Fusing Symmetric Uncertainty. In Proceedings of the 2024 IEEE 9th International Conference on Data Science in Cyberspace (DSC), Jinan, China, 23–26 August 2024; pp. 549–554. [Google Scholar]
  35. Abbasi, M.S.; Al-Sahaf, H.; Mansoori, M.; Welch, I. Behavior-based ransomware classification: A particle swarm optimization wrapper-based approach for feature selection. Appl. Soft Comput. 2022, 121, 108744. [Google Scholar] [CrossRef]
  36. Mahesh, T.; Santhakumar, D.; Balajee, A.; Shreenidhi, H.; Kumar, V.V.; Annand, J.R. Hybrid ant lion mutated ant colony optimizer technique with particle swarm optimization for leukemia prediction using microarray gene data. IEEE Access 2024, 12, 10910–10919. [Google Scholar] [CrossRef]
  37. Zhang, K.; Liu, Y.; Wang, X.; Mei, F.; Sun, G.; Zhang, J. Enhancing IoT (Internet of Things) feature selection: A two-stage approach via an improved whale optimization algorithm. Expert Syst. Appl. 2024, 256, 124936. [Google Scholar] [CrossRef]
  38. Wang, Y.C.; Song, H.M.; Wang, J.S.; Song, Y.W.; Qi, Y.L.; Ma, X.R. GOG-MBSHO: Multi-strategy fusion binary sea-horse optimizer with Gaussian transfer function for feature selection of cancer gene expression data. Artif. Intell. Rev. 2024, 57, 347. [Google Scholar] [CrossRef]
  39. Houssein, E.H.; Abdalkarim, N.; Samee, N.A.; Alabdulhafith, M.; Mohamed, E. Improved Kepler Optimization Algorithm for enhanced feature selection in liver disease classification. Knowl.-Based Syst. 2024, 297, 111960. [Google Scholar] [CrossRef]
  40. Sharma, N.; Arora, B. HFCCW: A novel hybrid filter-clustering-coevolutionary wrapper feature selection approach for network anomaly detection. Int. J. Mach. Learn. Cybern. 2024, 15, 4887–4922. [Google Scholar] [CrossRef]
  41. Zhu, F.; Li, G.; Tang, H.; Li, Y.; Lv, X.; Wang, X. Dung beetle optimization algorithm based on quantum computing and multi-strategy fusion for solving engineering problems. Expert Syst. Appl. 2024, 236, 121219. [Google Scholar] [CrossRef]
  42. Balakrishnan, K.; Dhanalakshmi, R.; Akila, M.; Sinha, B.B. Improved equilibrium optimization based on Levy flight approach for feature selection. Evol. Syst. 2023, 14, 735–746. [Google Scholar] [CrossRef]
  43. Zhang, H.; Gao, J.; Kang, L.; Zhang, Y.; Wang, L.; Wang, K. State of health estimation of lithium-ion batteries based on modified flower pollination algorithm-temporal convolutional network. Energy 2023, 283, 128742. [Google Scholar] [CrossRef]
  44. Gharehchopogh, F.S.; Namazi, M.; Ebrahimi, L.; Abdollahzadeh, B. Advances in sparrow search algorithm: A comprehensive survey. Arch. Comput. Methods Eng. 2023, 30, 427–455. [Google Scholar] [CrossRef] [PubMed]
  45. Nadimi-Shahraki, M.H.; Zamani, H.; Asghari Varzaneh, Z.; Mirjalili, S. A systematic review of the whale optimization algorithm: Theoretical foundation, improvements, and hybridizations. Arch. Comput. Methods Eng. 2023, 30, 4113–4159. [Google Scholar] [CrossRef]
  46. Su, H.; Zhao, D.; Heidari, A.A.; Liu, L.; Zhang, X.; Mafarja, M.; Chen, H. RIME: A physics-based optimization. Neurocomputing 2023, 532, 183–214. [Google Scholar] [CrossRef]
  47. Rao, R. Jaya: A simple and new optimization algorithm for solving constrained and unconstrained optimization problems. Int. J. Ind. Eng. Comput. 2016, 7, 19–34. [Google Scholar] [CrossRef]
  48. Shehab, M.; Mashal, I.; Momani, Z.; Shambour, M.K.Y.; AL-Badareen, A.; Al-Dabet, S.; Bataina, N.; Alsoud, A.R.; Abualigah, L. Harris hawks optimization algorithm: Variants and applications. Arch. Comput. Methods Eng. 2022, 29, 5579–5603. [Google Scholar] [CrossRef]
  49. Umarani, S.; Balaji, N.A.; Balakrishnan, K.; Guptha, N. Binary northern goshawk optimization for feature selection on micro array cancer datasets. Evol. Syst. 2024, 15, 1551–1565. [Google Scholar] [CrossRef]
  50. Kolamroudi, M.K.; Ilkan, M.; Egelioglu, F.; Safaei, B. Feature selection by ant colony optimization and experimental assessment analysis of PV panel by reflection of mirrors perpendicularly. Renew. Energy 2023, 218, 119238. [Google Scholar] [CrossRef]
  51. Li, S.; Chen, H.; Wang, M.; Heidari, A.A.; Mirjalili, S. Slime mould algorithm: A new method for stochastic optimization. Future Gener. Comput. Syst. 2020, 111, 300–323. [Google Scholar] [CrossRef]
  52. Kaya, E.; Gorkemli, B.; Akay, B.; Karaboga, D. A review on the studies employing artificial bee colony algorithm to solve combinatorial optimization problems. Eng. Appl. Artif. Intell. 2022, 115, 105311. [Google Scholar] [CrossRef]
  53. Cheraghalipour, A.; Hajiaghaei-Keshteli, M.; Paydar, M.M. Tree Growth Algorithm (TGA): A novel approach for solving optimization problems. Eng. Appl. Artif. Intell. 2018, 72, 393–414. [Google Scholar] [CrossRef]
  54. Hashim, F.A.; Houssein, E.H.; Mabrouk, M.S.; Al-Atabany, W.; Mirjalili, S. Henry gas solubility optimization: A novel physics-based algorithm. Future Gener. Comput. Syst. 2019, 101, 646–667. [Google Scholar] [CrossRef]
  55. Xing, Z.; He, Y. A two-step image segmentation based on clone selection multi-object emperor penguin optimizer for fault diagnosis of power transformer. Expert Syst. Appl. 2024, 244, 122940. [Google Scholar] [CrossRef]
  56. Zhu, D.; Wang, S.; Zhou, C.; Yan, S. Manta ray foraging optimization based on mechanics game and progressive learning for multiple optimization problems. Appl. Soft Comput. 2023, 145, 110561. [Google Scholar] [CrossRef]
Figure 1. Comparison of population distributions obtained from different initialization approaches.
Figure 1. Comparison of population distributions obtained from different initialization approaches.
Biomimetics 11 00003 g001
Figure 2. The evolving selection chances of the three differential operators over successive iterations.
Figure 2. The evolving selection chances of the three differential operators over successive iterations.
Biomimetics 11 00003 g002
Figure 3. The flowchart of the IHBOFS algorithm.
Figure 3. The flowchart of the IHBOFS algorithm.
Biomimetics 11 00003 g003
Figure 4. Average Training Time of Different Classifiers on High-Dimensional Datasets.
Figure 4. Average Training Time of Different Classifiers on High-Dimensional Datasets.
Biomimetics 11 00003 g004
Figure 5. KNN classification accuracy under different K values.
Figure 5. KNN classification accuracy under different K values.
Biomimetics 11 00003 g005
Figure 6. SVM classification accuracy under different C values.
Figure 6. SVM classification accuracy under different C values.
Biomimetics 11 00003 g006
Figure 7. Average convergence curves of comparison algorithms on high-dimensional datasets. (a) Yale. (b) ORL. (c) warpAR10P. (d) warpPIE10P. (e) pixraw10P. (f) orlraws10P.
Figure 7. Average convergence curves of comparison algorithms on high-dimensional datasets. (a) Yale. (b) ORL. (c) warpAR10P. (d) warpPIE10P. (e) pixraw10P. (f) orlraws10P.
Biomimetics 11 00003 g007
Figure 8. Average running time of comparison algorithms on high-dimensional datasets.
Figure 8. Average running time of comparison algorithms on high-dimensional datasets.
Biomimetics 11 00003 g008
Table 1. Summary of nascent hybrid algorithms for feature selection.
Table 1. Summary of nascent hybrid algorithms for feature selection.
AlgorithmDomainPerformance MetricsLimitations
ISPSO [32]High-dimensional feature selection, classification tasksAccuracy reaches 93.5%, with redundant features reduced by approximately 12%.Sensitive to initial parameter selection; requires high adaptability to specific high-dimensional scenarios; susceptible to feature dimensionality.
Dynamic PSO [33]Feature selection for dynamic data streamsAccuracy ranges from 88% to 92%; capable of handling high-dimensional dynamic data.Suitable for dynamic environments but struggles with real-time high-frequency changes; high computational resources.
SU-GA [34]Feature selection for bioinformatics dataAverage accuracy attains 91%.Poor performance on small-scale datasets.
PSO-GWO [35]Various domains in UCI repositories, hybrid feature selectionAverage accuracy of 90% with reduced computation time.Further research needed on alternative fitness and transfer functions.
Hybrid Ant Lion PSO [36]Feature identification for gene expression dataPrediction accuracy of 87.88%.Room for improvement in convergence rate for large-scale data processing.
EWOA [37]Internet of Things (IoT)Average accuracy of 95.5% on Aalto IoT dataset and 98.8% on RT-IoT 2022 dataset; average feature reduction rate of 82.5% (Aalto IoT) and 62.3% (RT-IoT 2022).Prone to the curse of dimensionality due to initial parameters; excellent performance on static tasks but weak adaptability to dynamic data.
GOG-MBSHO [38]Cancer gene feature selection, high-dimensional classificationOutperforms all comparative algorithms in average fitness value across selected datasets; average number of selected features is 63.Risk of performance degradation when processing extremely large datasets.
I-KOA [39]Liver disease classification, medical data feature selectionOverall classification accuracy of 93.46%; feature selection size of 0.1042.High computational cost and insufficient algorithm adaptability.
HFCCW [40]Network anomaly detectionAccuracy of 92%; feature selection rate ranges from 10% to 20%.High computational time consumption.
Table 2. Time complexity comparison analysis.
Table 2. Time complexity comparison analysis.
StageHBOIHBOFS
Population Initialization O ( N ) O ( 2 N )
Preservation Sub-population Update- O N · D 3
Crossover O N · D 3 O N · D 3
Self-crossover O N · D 3 O N · D 3
Population Sorting O ( N ) O ( N )
Optimal Solution Update O ( 1 ) O ( 1 )
Total O T · ( N + 2 N · D / 3 + 1 ) + N O T · ( N + N · D + 1 ) + 2 N
Table 3. Dataset description.
Table 3. Dataset description.
IDDatasetNumber of SamplesNumber of FeaturesNumber of Classes
D1Yale165102415
D2ORL400102410
D3warpAR10P130240040
D4warpPIE10P210240010
D5pixraw10P10010,00010
D6orlraws10P10010,30410
Table 4. Parameter settings for different improved strategies.
Table 4. Parameter settings for different improved strategies.
AlgorithmsParameter Setting
HBO S C = 8
HBO_INIT S C = 8 , s [ 0.01 , 0.92 ]
HBO_DE F = 0.5 , c r = 0.9 , S C = 8
HBO_TD σ [ 0.1 , 1.0 ] , g w = 2 , S C = 8
HBO_SC S C [ 4 , 10 ]
Table 5. Key parameter settings for comparison algorithms.
Table 5. Key parameter settings for comparison algorithms.
AlgorithmsParameter Settings
IHBOFS F = 0.5 , c r = 0.9 , g r = 2 , a = 0.1 ,
s [ 0.01 , 0.92 ] , σ [ 0.10 , 1.0 ] ,
SC max = 10 , SC min = 4
GA C R = 0.8 , M R = 0.01
FPA p = 0.8 , β = 1.5
SSA c 1 , c 2 [ 0 , 1 ]
WOAb = 1
RIMEW = 1.5
JAYA
GWO α [ 0 , 2 ]
HHO β = 1.5
Table 6. Ablation study results of the improvement strategies.
Table 6. Ablation study results of the improvement strategies.
FunctionHBOHBO_INITHBO_DEHBO_TDHBO_SCIHBOFS
Mean ± StdMean ± StdMean ± StdMean ± StdMean ± StdMean ± Std
F1310.55 ± 17.63317.80 ± 47.96316.91 ± 31.64307.11 ± 13.784787.90 ± 1189.40300.00 ± 0.00
F2465.03 ± 9.89461.10 ± 10.63459.11 ± 14.66485.41 ± 19.14588.89 ± 30.81448.94 ± 0.75
F3600.00 ± 0.00600.00 ± 0.00600.00 ± 0.01600.00 ± 0.00600.00 ± 0.00600.00 ± 0.00
F4930.62 ± 34.88927.49 ± 26.82914.28 ± 46.021048.30 ± 20.70982.61 ± 31.13878.03 ± 28.94
F5901.13 ± 0.61900.23 ± 0.26900.49 ± 0.35902.36 ± 0.70901.45 ± 0.64900.24 ± 0.20
F6707,110 ± 565,1601,445,900 ± 2,195,10034,424 ± 12,92923,070,000 ± 15,035,000631,090 ± 991,95045,885 ± 13,119
F71951.40 ± 83.461980.70 ± 79.731955.40 ± 81.692192.60 ± 247.342067.40 ± 51.131987.40 ± 76.82
F82887.40 ± 510.222984.00 ± 583.252303.90 ± 166.676223.30 ± 2264.103241.90 ± 352.972154.00 ± 101.03
F92642.60 ± 7.062628.80 ± 61.092624.70 ± 60.272639.30 ± 2.232741.20 ± 123.172635.70 ± 0.05
F102358.00 ± 887.072144.30 ± 1011.402480.90 ± 678.152577.20 ± 512.512847.50 ± 481.642527.60 ± 370.76
F112613.90 ± 11.512612.50 ± 11.262615.60 ± 18.512623.30 ± 3.042609.50 ± 11.712600.00 ± 0.01
F122939.40 ± 18.222942.70 ± 20.432927.80 ± 9.852929.40 ± 7.683068.50 ± 55.482900.00 ± 0.00
Table 7. Classification accuracy (%) and training time of different classifiers.
Table 7. Classification accuracy (%) and training time of different classifiers.
ClassifiersYaleORLwarpAR10PwarpPIE10Ppixraw10Porlraws10PAverage
ACCTIMEACCTIMEACCTIMEACCTIMEACCTIMEACCTIMEACCTIME
KNN64.850.0487.250.0649.230.0593.330.0695.000.0890.000.0879.940.0617
SVM66.060.0584.750.2670.770.0787.620.1597.000.1696.000.1883.700.1450
Xgboost60.610.9275.003.8468.461.3484.761.5180.001.9372.003.3973.472.1550
DT49.520.3123.400.9769.540.3269.290.6589.600.6174.101.1462.580.6667
RF70.121.1675.901.1678.541.1686.481.1695.801.1696.301.1783.861.1617
NB63.640.0588.250.0959.230.0687.140.0722.000.1693.000.1668.880.0983
KNN464.850.0487.250.0649.230.0593.330.0695.000.0890.000.0879.940.0617
KNN562.420.0483.500.0643.080.0590.480.0696.000.0889.000.0877.410.0617
KNN658.180.0479.750.0746.150.0586.190.0794.000.0887.000.0875.210.0650
SVM363.640.0582.250.2764.620.0781.900.1596.000.1693.000.1880.240.1467
SVM3.564.240.0582.750.2666.920.0786.190.1596.000.1796.000.1882.020.1467
SVM466.060.0584.750.2670.770.0787.620.1597.000.1696.000.1883.700.1450
Table 8. Classification accuracy (%) of IHBOFS with different transfer functions.
Table 8. Classification accuracy (%) of IHBOFS with different transfer functions.
DatasetMeasures (%)S1S2S3S4V1V2V3V4
YaleBest80.0080.0081.2180.0079.3980.0079.3978.79
Worst78.1878.1878.7978.1875.7675.7676.3675.15
Mean79.3379.2780.1279.4577.4577.2777.5877.03
Std0.570.450.590.571.111.091.051.26
ORLBest93.0092.0093.0093.0091.5091.0092.0091.25
Worst90.7590.5091.7591.2589.7589.5589.7589.75
Mean91.3391.2592.3592.2390.7390.2590.5090.30
Std0.630.400.440.610.570.500.720.46
warpAR10PBest85.3886.9286.1587.6984.6284.6286.9283.85
Worst81.5482.3183.8582.3180.7779.2380.7780.77
Mean84.1584.2385.1584.3882.1582.3883.0882.38
Std1.101.340.601.621.131.521.651.11
warpPIE10PBest100.00100.00100.0099.5299.0599.5299.5298.10
Worst97.1497.1497.6297.1495.7195.2496.6794.76
Mean98.5298.3898.8198.5297.0597.4897.7196.52
Std0.860.860.860.691.221.510.921.04
pixraw10PBest100.00100.00100.00100.00100.00100.00100.00100.00
Worst99.0099.0099.0099.0099.0099.0099.0099.00
Mean99.4099.6099.5099.6099.5099.6099.4099.60
Std0.490.490.500.490.500.490.490.49
orlraws10PBest100.00100.00100.00100.00100.0099.00100.00100.00
Worst98.0099.0099.0098.0099.0098.0098.0098.00
Mean99.5099.5099.5099.2099.3098.8099.0099.00
Std0.810.500.500.600.460.400.450.63
Table 9. Number of selected features by IHBOFS with different transfer functions.
Table 9. Number of selected features by IHBOFS with different transfer functions.
Transfer
Functions
YaleORLwarpAR10PwarpPIE10Ppixraw10Porlraws10P
AvgStdAvgStdAvgStdAvgStdAvgStdAvgStd
S1409.1031.49410.6050.08454.10176.20430.20159.8724.009.86138.90188.01
S2397.6048.38451.0042.01474.90233.89430.00205.9227.9013.4991.0038.04
S3361.7077.88420.7025.98472.00131.48253.60117.9120.005.39110.4068.14
S4408.4051.10364.5070.05419.30202.34351.20155.3621.306.07154.90119.42
V1262.6059.37373.1068.78400.40208.35430.10264.0339.3015.32123.7055.84
V2287.3066.58408.7069.11458.80204.12332.40261.4945.1028.07132.7053.86
V3290.2080.84401.4050.58386.50137.02218.9093.7640.3031.87151.1092.95
V4313.1071.76405.9050.04417.60141.61418.30232.1041.8026.16155.40105.44
Table 10. Experimental results of comparison algorithms on high-dimensional datasets.
Table 10. Experimental results of comparison algorithms on high-dimensional datasets.
DatasetsAlgorithmsBestWorstMeanStdNumTime
D1HBO78.7976.3677.580.61351.694.47
IHBOFS81.2178.7980.120.59361.7088.69
HHO75.7671.5273.391.34238.30194.68
ACO75.7672.1273.581.09464.80100.05
SMA71.5269.0970.180.7689.6096.89
ABC75.1571.5274.001.03418.70201.66
TGA72.1269.7070.850.69504.80110.49
HGSO70.3069.0969.880.39493.20103.12
EPO73.3370.9172.060.8863.6058.53
MRFO77.5872.1276.061.52246.60192.29
D2HBO91.5089.7590.720.55384.20472.91
IHBOFS93.0091.7592.350.44420.70447.11
HHO90.5087.5088.631.13281.901040.11
ACO90.2587.5088.920.74479.00521.61
SMA87.5085.0085.670.68113.10440.60
ABC89.7589.2589.470.21457.901046.83
TGA87.0085.7586.280.36503.40567.56
HGSO86.7585.2585.830.45506.40514.93
EPO87.7585.2586.780.79112.90293.91
MRFO91.5088.0089.150.97339.20990.85
D3HBO83.8580.0081.311.09751.60119.10
IHBOFS86.1583.8585.150.60472.00108.14
HHO83.0878.4680.921.75280.10210.61
ACO79.2376.9278.080.711093.70126.27
SMA80.0075.3876.851.2141.90108.36
ABC79.2377.6978.380.411059.10253.99
TGA76.9274.6276.080.641186.90141.09
HGSO76.9273.8575.230.831190.60127.22
EPO83.0876.1580.152.2534.0074.33
MRFO83.0879.2380.921.23432.90237.55
D4HBO98.1094.2995.811.04765.30230.18
IHBOFS100.0097.6298.810.86253.60190.31
HHO97.6295.7196.430.65115.10359.06
ACO94.2991.9093.190.741109.20250.45
SMA94.7690.9592.951.1245.40171.51
ABC94.7692.8693.480.711078.30491.03
TGA92.3890.9591.620.491203.90277.40
HGSO91.4390.4890.090.261203.20236.71
EPO95.7193.3394.430.8040.60107.09
MRFO99.0592.8696.671.92313.90350.61
D5HBO99.0099.0099.000.001716.90308.34
IHBOFS100.0099.0099.500.5020.00212.40
HHO100.0099.0099.100.3023.00350.00
ACO99.0099.0099.000.004523.60322.90
SMA100.0099.0099.500.5016.90266.08
ABC99.0099.0099.000.004480.80661.02
TGA99.0099.0099.000.004917.50365.06
HGSO99.0098.0098.900.304958.10298.39
EPO100.0099.0099.300.4613.60134.40
MRFO100.0099.0099.600.4931.50372.00
D6HBO98.0097.0097.800.401830.80330.51
IHBOFS100.0099.0099.500.50110.40256.00
HHO100.0098.0098.900.7047.30383.50
ACO98.0097.0097.700.464730.30349.58
SMA100.0098.0098.600.6645.20270.70
ABC98.0098.0098.000.004670.80716.17
TGA98.0097.0097.400.495080.50397.87
HGSO97.0097.0097.000.005107.60338.67
EPO100.0098.0098.700.9028.40140.43
MRFO100.0098.0099.000.77110.20445.54
Table 11. Friedman test for the average accuracy of comparison algorithms.
Table 11. Friedman test for the average accuracy of comparison algorithms.
AlgorithmAverage RankingFinal Rankp-Value
IHBOFS1.16666718.182420 × 10−10
MRFO2.6666672
HBO3.8333333
HHO4.3333334
EPO5.3333335
ABC5.5000006
ACO6.5000007
SMA7.1666678
TGA8.6666679
HGSO9.83333310
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiang, C.; Zhou, J.; Zhou, W. IHBOFS: A Biomimetics-Inspired Hybrid Breeding Optimization Algorithm for High-Dimensional Feature Selection. Biomimetics 2026, 11, 3. https://doi.org/10.3390/biomimetics11010003

AMA Style

Xiang C, Zhou J, Zhou W. IHBOFS: A Biomimetics-Inspired Hybrid Breeding Optimization Algorithm for High-Dimensional Feature Selection. Biomimetics. 2026; 11(1):3. https://doi.org/10.3390/biomimetics11010003

Chicago/Turabian Style

Xiang, Chunli, Jing Zhou, and Wen Zhou. 2026. "IHBOFS: A Biomimetics-Inspired Hybrid Breeding Optimization Algorithm for High-Dimensional Feature Selection" Biomimetics 11, no. 1: 3. https://doi.org/10.3390/biomimetics11010003

APA Style

Xiang, C., Zhou, J., & Zhou, W. (2026). IHBOFS: A Biomimetics-Inspired Hybrid Breeding Optimization Algorithm for High-Dimensional Feature Selection. Biomimetics, 11(1), 3. https://doi.org/10.3390/biomimetics11010003

Article Metrics

Back to TopTop