Next Article in Journal
Effect of Surfactant on Bubble Formation on Superhydrophobic Surface in Quasi-Static Regime
Next Article in Special Issue
An Enhanced Snow Geese Optimizer Integrating Multiple Strategies for Numerical Optimization
Previous Article in Journal
Enhancing Legged Robot Locomotion Through Smooth Transitions Using Spiking Central Pattern Generators
Previous Article in Special Issue
An Improved Northern Goshawk Optimization Algorithm for Mural Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Three Strategies Enhance the Bionic Coati Optimization Algorithm for Global Optimization and Feature Selection Problems

1
School of Mechanical Engineering, Hunan Institute of Science and Technology, Yueyang 414006, China
2
School of Data Science and Ecommerce, Henan University of Economics and Law, Zhengzhou 450016, China
3
School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Biomimetics 2025, 10(6), 380; https://doi.org/10.3390/biomimetics10060380
Submission received: 4 May 2025 / Revised: 4 June 2025 / Accepted: 5 June 2025 / Published: 7 June 2025

Abstract

With the advancement of industrial digitization, utilizing large datasets for model training to boost performance is a pivotal technical approach for industry progress. However, raw training datasets often contain abundant redundant features, which increase model training’s computational cost and impair generalization ability. To tackle this, this study proposes the bionic ABCCOA algorithm, an enhanced version of the bionic Coati Optimization Algorithm (COA), to improve redundant feature elimination in datasets. To address the bionic COA’s inadequate global search performance in feature selection (FS) problems, leading to lower classification accuracy, an adaptive search strategy is introduced. This strategy combines individual learning capability and the learnability of disparities, enhancing global exploration. For the imbalance between the exploration and exploitation phases in the bionic COA algorithm when solving FS problems, which often traps it in suboptimal feature subsets, a balancing factor is proposed. By integrating phase control and dynamic adjustability, a good balance between the two phases is achieved, reducing the likelihood of getting stuck in suboptimal subsets. Additionally, to counter the bionic COA’s insufficient local exploitation performance in FS problems, increasing classification error rates, a centroid guidance strategy is presented. By combining population centroid guidance and fractional-order historical memory, the algorithm lowers the classification error rate of feature subsets and speeds up convergence. The bionic ABCCOA algorithm was tested on the CEC2020 test functions and engineering problem, achieving an over 90% optimization success rate and faster convergence, confirming its efficiency. Applied to 27 FS problems, it outperformed comparative algorithms in best, average, and worst fitness function values, classification accuracy, feature subset size, and running time, proving it an efficient and robust FS algorithm.

1. Introduction

Advancements in the field of Artificial Intelligence (AI) have spurred developments across various domains, including education, healthcare, and manufacturing [1,2,3,4,5,6,7,8,9,10]. Currently, large-scale model technologies within AI have emerged as a pivotal focal point [11,12,13,14,15]. Leveraging data for model training to enhance predictive and inferential performance has become a crucial technical approach, driving progress in these industries [16,17,18,19,20,21,22,23,24,25]. Unfortunately, however, the raw datasets employed for model training often contain a substantial number of redundant features, which diminishes the interpretability of the data and incurs computational costs while compromising the model’s generalization ability [26,27,28,29,30]. To enhance the interpretability and generalization capacity of models, it is typically necessary to eliminate redundant features within datasets, a process known as feature dimensionality reduction [31,32,33,34,35]. This not only improves model interpretability but also liberates additional computational resources. Assuming the original dataset comprises M feature information units, the task of feature dimensionality reduction involves searching through 2M possible feature subset combinations, which constitutes an NP-hard problem [36,37,38]. To reduce the computational cost associated with the feature subset search process, employing bionic-inspired feature selection methods to address this NP-hard problem has been demonstrated to be more appropriate [39,40,41,42,43,44]. Consequently, the objective of this paper is to propose a robust bionic-inspired feature selection (FS) method with efficient search capabilities. The aim is to leverage the computational simplicity of bionic-inspired algorithms to cut down on the computational expenses during the data denoising process, effectively eliminate redundant features, and enhance the interpretability and generalization ability of model training.
Currently, there are two primary approaches to FS: the filter method [45] and the wrapper method [46]. The filter method evaluates the significance of each feature independently by assessing statistical measures such as correlation [47], information gain [48], variance [49], and the chi-square test [50], among others. This enables the selection of an appropriate feature subset combination to represent the original dataset. However, this approach is prone to overlooking interactions among features [51], potentially leading to the exclusion of important features and a subsequent decline in classification accuracy. In contrast, the wrapper method is model performance-centric. It iteratively selects feature subsets, which are then incorporated into the model training process. The current feature subset is evaluated using algorithms such as K-Nearest Neighbors (KNN) [52], neural networks [53], and Support Vector Machines (SVM) [54]. Features are either removed or added to obtain the optimal feature subset that best represents the entire dataset, thereby enhancing classification accuracy. The primary advantage of this method lies in its consideration of interactions among features, significantly improving the model’s interpretability and substantially increasing the classification accuracy of the feature subset [41]. Compared to the filter method, the wrapper method better accounts for interactions among features and achieves higher classification accuracy for feature subsets. However, the wrapper method’s computational cost is elevated due to the need to exhaustively search through a vast number of feature subset combinations. Meta-heuristic algorithms derived from bionic simulations of natural organisms can significantly reduce the computational cost of traversing feature subsets while effectively enhancing classification accuracy, owing to their structural flexibility and simplicity [55].
Given the advantage of bionic algorithms in terms of low computational costs, numerous bionic-based FS methods have been proposed by current scholars to address FS problems, aiming to effectively eliminate redundant features while enhancing classification accuracy. For instance, Askr et al. [56] introduced a bionic algorithm named Binary Enhanced Golden Jackal Optimization (BEGJO) to improve classification accuracy in feature selection for high-dimensional datasets. They enhanced the original bionic Golden Jackal Optimization (GJO) algorithm by incorporating Copula entropy and four enhanced search strategies. Experimental results on 15 FS problems demonstrated bionic BEGJO’s superior performance in terms of classification accuracy and feature dimensionality, proving its effectiveness in high-dimensional feature selection. However, its drawback lies in its runtime ranking fourth, indicating potential for improvement. Abdel-Salam et al. [57] proposed an adaptive chaotic version of the bionic RIME algorithm, termed bionic ACRIME, which integrates four improvement strategies to enhance the algorithm’s FS performance. Subsequently, bionic ACRIME was applied to solve FS problems across 14 different datasets. Experimental results indicated that the proposed bionic algorithm effectively located the optimal feature subsets, improved classification accuracy, and eliminated redundant features. Nevertheless, this work still exhibits certain performance limitations when addressing high-dimensional FS problems and does not comprehensively consider runtime. Tijjani et al. [58] introduced a novel bionic Particle Swarm Optimization (PSO)-based FS method called bionic EBPSO, which enhances feature selection performance by incorporating a dimensionality reduction mechanism and a position update mechanism. Experimental results demonstrated its ability to effectively reduce feature dimensionality and runtime while improving classification accuracy. However, the proposed bionic method still has some shortcomings when dealing with high-dimensional FS problems, such as an imbalance between local and global searches, which increases the risk of becoming trapped in locally optimal feature subsets. Fang et al. [59] proposed a bionic genetic algorithm-based feature selection method to reduce redundant features in ICS data. The experimental results showed that, through a feature ranking fusion mechanism and a growth tree clustering idea, the method achieved effective feature selection for ICS data, leading to higher ICS processing performance. Despite its good performance, the issue of runtime has not been addressed. Ye et al. [60] proposed an improved bionic Hybrid Breeding Optimization (HBO) algorithm, termed bionic CCI-HBO, by introducing Lévy flights, elite opposition-based learning strategies, and cooperative co-evolution to enhance the algorithm’s FS performance. Experimental results indicated that bionic CCI-HBO outperformed over 90% of the FS problems, but it did not show a significant advantage in classification accuracy for high-dimensional FS problems. Wang et al. [61] proposed a role-oriented bionic Binary Grey Wolf Optimizer (BFLGWO) for feature selection, which enhances FS performance by incorporating a V-shaped linear transfer function and a role-oriented strategy. Experimental results demonstrated that the method excels in balancing exploitation and exploration and is competitive in feature selection tasks. Although bionic BFLGWO outperforms other bionic algorithms in terms of the number of selected features and classification accuracy, it has limitations in terms of time consumption. Houssein et al. [62] proposed an improved bionic mSTOA algorithm that optimizes FS performance by balancing exploration and exploitation, adaptive parameter control, and population reduction strategies. Experimental results on nine FS problems showed that it achieved the highest classification accuracy on over 89.5% of the problems. However, its limitation is that the tested FS problems are all in low-dimensional spaces, failing to represent high-dimensional FS performance. It may still exhibit performance deficiencies when addressing high-dimensional FS problems, and runtime has not been fully considered.
The aforementioned related works have demonstrated that metaheuristic-based bionic FS methods possess powerful capabilities for dataset dimensionality reduction, effectively eliminating redundant features and enhancing classification accuracy. However, as the dimensionality and scale of FS problems increase, existing bionic FS methods encounter issues such as low classification accuracy, an inability to eliminate redundant features, suboptimal runtime, and insufficient stability. The primary cause of these problems lies in the exponential growth of feature subset combinations that algorithms must search through as the dimensionality of FS problems rises. Due to limitations in their global exploration and local exploitation capabilities, bionic algorithms are unable to thoroughly traverse the feature space, thereby constraining classification accuracy. Additionally, most existing bionic works fail to achieve a good balance across multiple metrics, such as classification accuracy, feature subset size, runtime, stability, and convergence speed, when solving high-dimensional FS problems. Although they may excel in a specific metric, performance limitations in other metrics can compromise the bionic algorithm’s reliability. Consequently, these facts motivate us to propose a robust bionic FS algorithm that comprehensively considers multiple performance metrics to alleviate the problems of reduced classification accuracy and limited runtime caused by becoming trapped in locally suboptimal feature subsets in high-dimensional FS problems. Fortunately, the bionic Coati Optimization Algorithm (COA) has been validated as a robust tool with efficient optimization performance [63], and it has been applied in various fields, demonstrating strong application scalability. To date, however, no relevant researchers have applied bionic COA to solve FS problems. Therefore, this paper applies bionic COA to address FS problems to improve their solution performance while also filling a research gap in the bionic COA algorithm. Furthermore, considering the gradual increase in the dimensionality of current FS problems, the original bionic COA may suffer from insufficient global search and local exploitation capabilities during the solution process. To further enhance the algorithm’s FS performance, this paper proposes an enhanced bionic COA (ABCCOA) by integrating an adaptive search strategy, a balancing factor, and a centroid guidance strategy into the original bionic COA. Tests conducted on 27 FS problems confirm that the proposed bionic ABCCOA-based FS method exhibits significant advantages in terms of classification accuracy, stability, and runtime and can be considered a promising bionic FS approach. The main contributions of this paper are as follows:
  • An adaptive search strategy is proposed, which integrates individual learning capabilities and the learnability of disparities, effectively enhancing the algorithm’s global exploration capabilities.
  • A balancing factor is introduced, incorporating phase-based and dynamically adjustable characteristics, to achieve a well-balanced interplay between the exploration and exploitation phases.
  • A centroid guidance strategy is devised, which combines the concept of centroid guidance with the utilization of fractional-order historical memory, thereby improving the algorithm’s exploitation capabilities.
  • By integrating the aforementioned three strategies, the bionic ABCCOA algorithm is formulated. Experimental results on 27 FS problems demonstrate its superiority in terms of classification accuracy, stability, and runtime. This confirms that bionic ABCCOA is a promising bionic FS method.
The subsequent work plan for this paper is outlined as follows: In Section 2, the concept of the bionic COA is introduced. Section 3 proposes the bionic ABCCOA algorithm by integrating three learning strategies into the bionic COA framework. Section 4 validates the robust optimization performance of bionic ABCCOA through experiments conducted on the CEC2020 benchmark functions and engineering problems. Section 5 applies bionic ABCCOA to solve 27 FS problems, confirming its promise as an effective FS method. Finally, Section 6 provides a summary of the work presented in this paper and outlines the future research directions.

2. Mathematical Model of Coati Optimization Algorithm

In this section, the primary focus is on introducing the natural inspiration and mathematical model of the COA. The COA is a swarm intelligence optimization algorithm primarily formulated by simulating the attacking and escaping behaviors of coatis [63]. Specifically, the attacking behavior in the algorithm mainly simulates the global exploration phase of the COA, while the escaping behavior predominantly models the local exploitation phase. Additionally, when solving optimization problems using the COA, it is necessary to initialize a set of candidate solutions to form the initial population for subsequent iterative refinement. This process is referred to as the population initialization phase. Following this, the initial population undergoes updates in both the global search phase and the local exploitation phase to enhance the quality of the candidate solutions. In summary, the COA mainly consists of three phases: population initialization, global exploration, and local exploitation. A detailed description of each phase will be provided in the following subsections.

2.1. Population Initialization Phase

In this section, the focus is on introducing the population initialization phase of the COA. When utilizing the COA to solve optimization problems, the first step involves initializing a population comprising N individuals for subsequent algorithmic iterations. Each individual within this population represents a potential solution to the optimization problem and is characterized as a vector of size 1 × D, where D denotes the number of variables in the problem to be optimized. Under the aforementioned definitions, the initial population is computed using Equation (1).
X = X 1 X i X N N × 1 = x 1 , 1 x 1 , j x 1 , D x i , 1 x i , j x i , D x N , 1 x N , j x N , D N × D
where X represents the initialized population, X i denotes the ith individual within the population, and x i , j signifies the value of the jth dimension for the ith individual. Specifically, x i , j is calculated using Equation (2).
x i , j = l b j + r a n d · ( u b j l b j ) ,   i = 1 , 2 , ,   N ,   j = 1 , 2 , , D .
where rand represents a random number within the interval [0, 1], while u b j and l b j denote the jth dimensional values of the upper and lower bounds, respectively, for the optimization problem to be solved. Subsequently, each individual is evaluated using its fitness function value to characterize its problem-solving capability. For minimization optimization problems, a smaller fitness function value indicates that the corresponding individual possesses a higher quality of problem-solving. The fitness function value corresponding to each individual is represented using Equation (3).
F = F 1 F i F N N × 1 = F ( X 1 ) F ( X i ) F ( X N ) N × 1
where F denotes the vector of fitness function values for all individuals in the population, F i represents the fitness function value corresponding to the ith individual, and F ( · ) signifies the computational function for the fitness function value. After the population initialization is completed, it is necessary to optimize the quality of individuals within the population through the global exploration phase and the local exploitation phase. In the following subsections, the global exploration phase and the local exploitation phase of the COA algorithm will be introduced in detail.

2.2. Global Exploration Phase

In this section, the global exploration phase of the COA will be introduced in detail. This phase is primarily inspired by the attacking behavior of coatis towards iguanas. In this behavior, coatis climb trees to intimidate iguanas, causing them to fall to the ground. Subsequently, the coatis on the ground attempt to attack the iguanas, enabling them to move to various positions within the search space. As illustrated in Figure 1, this attacking behavior effectively mirrors the global search characteristic of the algorithm, allowing it to explore a wider range of potential optimal solution regions when solving optimization problems. In this context, we assume that the position of the iguanas represents the optimal individual within the population. Half of the coatis climb trees to launch attacks on the iguanas, while the other half remain on the ground to target the fallen iguanas. Based on this assumption, the behavior of coatis climbing trees to attack the iguanas is mathematically modeled using Equation (4).
x i , j P 1 = x i , j + r a n d · I g u a n a j I · x i , j ,   i = 1 , 2 , , N 2 ,   j = 1 , 2 , , D .
where x i , j P 1 denotes the new state of the jth dimension of the ith individual after simulating the attacking behavior. I g u a n a j represents the value of the jth dimension of the best individual, and I is a randomly selected constant from the set {1,2}. After the iguanas fall to random positions on the ground, the coatis on the ground launch attacks on them. This behavior is mathematically simulated by Equation (5).
x i , j P 1 = x i , j + r a n d · I g u a n a j G I · x i , j , i f   F I g u a n a G < F i x i , j + r a n d · x i , j I g u a n a j G , e l s e , i = N 2 + 1 , N 2 + 2 , , N ,   j = 1 , 2 , , D .
where F i represents the fitness function value of the ith individual, and   F I g u a n a G denotes the fitness function value of the iguanas that have fallen to the ground. I g u a n a j G signifies the value of the jth dimension of the position of the iguanas that have fallen to the ground; this is calculated using Equation (6).
I g u a n a j G = l b j + r a n d · ( u b j l b j ) ,   j = 1 , 2 , , D .
Subsequently, the new state generated by an individual during the attacking behavior is retained using Equation (7).
X i = X i P 1 , i f   F i P 1 < F i X i , e l s e
where X i P 1 denotes the new individual state generated by the ith individual through the simulation of the attacking behavior, and F i P 1 represents the fitness function value corresponding to the individual X i . By simulating the attacking behavior of coatis, the global exploration capability of the COA algorithm is ensured. In the following subsection, the local exploitation phase of the COA algorithm will be introduced in detail.

2.3. Local Exploitation Phase

In this section, a detailed description is provided regarding the local exploitation phase of the COA algorithm. During this phase, the focus is on simulating the escape behavior of coatis when confronted with predators. When coatis are subjected to attacks from predators, they promptly flee within a designated safe zone to ensure their safety. As illustrated in Figure 2, by simulating this behavior, the algorithm’s local exploitation performance is effectively guaranteed, enabling it to further explore and exploit the identified regions of potential optimal solutions. Specifically, Equation (8) is utilized to simulate the escape behavior of coatis.
x i , j P 2 = x i , j + ( 1 2 r ) · l b j l o c a l + r · ( u b j l o c a l l b j l o c a l ) , i = 1 , 2 , , N ,   j = 1 , 2 , , D .
where x i , j P 2 represents the new state of the jth dimension for the ith individual after simulating the escape behavior, r denotes a random number generated within the interval [0, 1], and both l b j l o c a l and u b j l o c a l are calculated using Equation (9).
l b j l o c a l = l b j t ,   u b j l o c a l = u b j t ,   t = 1 , 2 , , T .
where t represents the current iteration number of the algorithm, and T denotes the maximum number of iterations for the algorithm. Subsequently, the new state generated by an individual during the escape behavior is retained using Equation (10).
X i = X i P 2 , i f   F i P 2 < F i X i , e l s e
where X i P 2 represents the new individual state generated by the ith individual after simulating the escape behavior, and F i P 2 denotes the fitness function value corresponding to the individual X i P 2 . By simulating the escape behavior of coatis, the algorithm’s local exploitation capability is ensured, thereby guaranteeing the optimization precision. After introducing the population initialization phase, the global exploration phase, and the local exploitation phase involved in the COA algorithm, the next subsection will present the execution logic of the COA algorithm.

2.4. Execution Framework of COA Algorithm

In this section, the execution logic of the COA algorithm when solving optimization problems will be introduced. When the COA algorithm is applied to solve an optimization problem, it first initializes the population to generate a set of high-quality candidate solutions. Subsequently, through the synergistic effects of the global exploration phase and the local exploitation phase, the algorithm enhances the quality of individuals on a population-wide basis, thereby effectively addressing the optimization problem. The pseudo code for the COA algorithm is presented in Algorithm 1. Additionally, to facilitate an intuitive understanding of the algorithm’s execution logic, Figure 3a illustrates the flowchart of the algorithm’s execution.
Algorithm 1: Pseudo code of COA algorithm
Input: Population size (N), Dimension (D), Upper bound (ub) and lower bound (lb), Maximum of iterations (T).
Output: Global best solution (Xbest).
1.  Initialize population based on Equation (1) and calculate the individual fitness function values of the population.
2.  for t = 1:T
3.    Calculate positions of iguanas using the globally best individual.
4.    Phase 1: The attack behavior of coatis (Global Exploration Phase)
5.    for i = 1 : N / 2
6.     for j = 1:D
7.       Calculate the jth dimensional new state of the ith individual based on Equation (4).
8.     end for
9.    end for
10.  for i = 1 + N / 2 : N
11.    for j = 1:D
12.      Calculate the jth dimensional new state of the ith individual based on Equation (5).
13.    end for
14.  end for
15.  Use Equation (7) to preserve the new state of individual X i .
16.  Phase 2: The escape behavior of coatis (Local Exploitation Phase)
17.  for i = 1:N
18.    for j = 1:D
19.      Calculate the jth dimensional new state of the ith individual based on Equation (8).
20.    end for
21.  end for
22.  Use Equation (10) to preserve the new state of individual X i .
23.  Save the global best solution Xbest.
24. end for
25. Output the global best solution Xbest obtained by solving the optimization problem using the COA algorithm.

3. The Proposal of ABCCOA Algorithm

The original COA algorithm suffers from insufficient global exploration capability, inadequate local exploitation capability, and an imbalance between the exploration and exploitation phases when solving FS problems. These issues lead to the algorithm easily becoming trapped in suboptimal feature subset combinations and experiencing local stagnation, particularly when addressing high-dimensional FS problems, thereby compromising the classification accuracy of feature subsets. To enhance the FS performance of the COA algorithm, this paper proposes an enhanced COA algorithm, termed ABCCOA, by integrating three learning strategies into the COA framework, thereby improving its FS performance from various perspectives. In the ABCCOA algorithm, firstly, an adaptive search strategy is introduced to address the deficiency in global search performance of the COA algorithm when solving FS problems due to the exponential increase in the number of feature subset combinations, which, in turn, reduces classification accuracy. This strategy facilitates learning from the disparities between different types of individuals, taking into account both the individual’s own learning capacity and the learnability of each disparity group. It effectively enhances the population diversity during the iterative process, thereby strengthening the algorithm’s global exploration capability and improving the classification accuracy of feature subsets. Secondly, to tackle the issue of the COA algorithm’s tendency to become trapped in suboptimal feature subset combinations due to an imbalance between the exploration and exploitation phases when solving FS problems, a balancing factor is proposed to better control the global exploration and local exploitation phases. In the proposed balancing factor, the phase is controlled through sine and cosine functions, while arcsine and arccosine functions are also incorporated for dynamic adjustment. This ensures that the algorithm maintains higher global exploration capability while also guaranteeing a certain level of exploitation capability during the early iterations, contributing to improved convergence speed and accuracy. Meanwhile, during the later iterations, the algorithm exhibits higher local exploitation capability while still ensuring a certain level of global search capability, reducing the probability of the algorithm becoming trapped in suboptimal FS subsets. Finally, to address the issue of increased classification error rates resulting from insufficient local exploitation capability of the COA algorithm when solving FS problems, a centroid guidance strategy is proposed to enhance the algorithm’s local exploitation capability. In this strategy, guidance is provided by leveraging the centroids of individuals from the previous four generations of the population, combined with the fractional-order historical memory effect. This approach not only reduces the classification error rate of FS subsets but also accelerates the convergence speed of the algorithm. In the following sections, we will elaborate on the mathematical concepts of the adaptive learning strategy, the balancing factor, and the centroid guidance strategy, as well as the execution logic of the ABCCOA algorithm.

3.1. Adaptive Search Strategy

The original COA algorithm exhibits insufficient global exploration capability when addressing complex FS problems, primarily due to the increase in the number of features within the original dataset. This deficiency leads to reduced population diversity, hindering the algorithm’s ability to locate regions containing potentially optimal FS subset combinations and subsequently decreasing the classification accuracy of the subsets. To mitigate this issue, there is an urgent need to propose a search strategy with efficient exploratory capabilities. Zhang et al. [64] have indicated that learning by individuals from the gaps between different types of individuals within the population can enhance the algorithm’s global exploration capability. Inspired by this, this section proposes an adaptive search strategy to augment the algorithm’s global exploratory performance. In the adaptive search strategy, individuals primarily learn from four sets of individual gaps: the gap between the globally best individual and better individuals ( G a p 1 ), the gap between the globally best individual and worse individuals ( G a p 2 ), the gap between better and worse individuals ( G a p 3 ), and the gap between two randomly selected individuals ( G a p 4 ). These four sets of gaps are calculated using Equation (11).
G a p 1 = X b e s t X b e t t e r G a p 2 = X b e s t X w o r s e G a p 3 = X b e t t e r X w o r s e G a p 4 = X r a n d 1 X r a n d 2
where X b e s t denotes the globally best individual, and X b e t t e r represents a better individual, which is defined as a randomly selected individual from the set of the top five individuals with the smallest fitness function values within the population. X w o r s e denotes a worse individual, defined as a randomly selected individual from the set of the top five individuals with the largest fitness function values within the population. X r a n d 1 and X r a n d 2 represent two distinct random individuals selected from the population. Moreover, the learnability of each gap of individual disparities also varies. We employed Equation (12) to calculate the learnability of each gap of disparities.
L B k = G a p k k = 1 4 G a p k ,   ( k = 1 , 2 , 3 , 4 )
where L B k denotes the learnability of the kth gap of disparities. As can be observed from the equation, the greater the disparity between individuals, the higher the learnability of this gap of disparities. This is because a larger disparity implies a greater amount of information embedded within this gap. Furthermore, individuals with different characteristics exhibit varying learning capacities. We utilized Equation (13) to calculate the learning capacity of each individual.
L C i = F i F m a x
where L C i represents the learning capacity of the ith individual, F i denotes the fitness function value of the ith individual, and F m a x signifies the maximum fitness function value within the population. As can be inferred from the equation, individuals with smaller fitness function values possess lower learning capacities. This is primarily because a smaller fitness function value implies that the individual is of higher quality. Consequently, it is necessary to reduce the global exploration phase to preserve the individual’s quality and enhance the optimization precision, and vice versa. Based on the aforementioned definitions, Equation (14) is employed to calculate the learning process of the ith individual from the kth gap of disparities.
K A k = L C i · L B k · G a p k ,   ( k = 1 , 2 , 3 , 4 )
where K A k represents the information acquired by the ith individual after learning from the kth gap of disparities. Subsequently, Equation (15) is utilized to compute the new state generated by the ith individual after learning from all four gaps of disparities.
X i P 1 = X i + K A 1 + K A 2 + K A 3 + K A 4
Subsequently, Equation (7) is employed to retain the new state of the individual. To provide a more intuitive representation of the adaptive search strategy, Figure 4 visualizes the entire search process.

3.2. Balancing Factor

When the original COA algorithm is applied to solve feature selection (FS) problems, as the number of feature elements in the original dataset increases dramatically, the algorithm is required to search through feature subset combinations that grow exponentially in magnitude. This results in the COA algorithm lacking a proper balance between global exploration and local exploitation, making it prone to becoming trapped in locally suboptimal feature subset combinations and exacerbating the deterioration of classification error rates. To alleviate this issue, this section proposes a novel balancing factor to reasonably balance the global exploration phase and the local exploitation phase. This factor is primarily composed of trigonometric and inverse trigonometric functions. Specifically, the sine and cosine functions mainly control the phase of the balancing factor, while the arcsine and arccosine functions primarily provide dynamic adjustments to the balancing factor. This configuration enables the algorithm to possess higher global exploration capabilities while ensuring a certain level of exploitation capability during the early iterations, thereby enhancing the algorithm’s convergence speed and precision. Simultaneously, it ensures higher local exploitation performance during the later iterations while maintaining a certain degree of global search capability, reducing the probability of the algorithm becoming stuck in suboptimal FS subsets. Specifically, the phase adjustment term formed by the combination of the sine and cosine functions is expressed as Equation (16).
p a t = ( s i n ( π · t / 2 T ) ) 2 + 0.5 · ( 1 c o s ( π · t / T ) )
where pat denotes the phase adjustment term. The dynamic adjustment term, which is formed by combining the arcsine and arccosine functions, is expressed as Equation (17).
d a t = 2 π · ( a r c s i n ( t / T ) + a r c c o s ( t / T ) )
where dat represents the dynamic adjustment term, a r c s i n ( · ) denotes the arcsine operation, and a r c c o s ( · ) signifies the arccosine operation. Subsequently, the phase adjustment term pat and the dynamic adjustment term dat are combined to form the balancing factor proposed in this section, which is expressed as Equation (18).
B F = ( t T ) d a t · p a t i f   ( t T ) < 0.5 1 ( t T ) d a t · p a t e l s e
where BF denotes the balancing factor, and its value decreases non-linearly as the number of iterations t increases. To better illustrate this decreasing process, Figure 5 visualizes the changes in BF with the growth of the number of iterations. As can be observed from the figure, during the early stages of iteration, the global exploration phase of the algorithm dominates. However, as iterations progress, the local exploitation phase takes precedence. It is noteworthy that, even in the later stages of iteration, the algorithm still retains a certain degree of global exploration capability. This enables the algorithm to enhance classification accuracy while simultaneously reducing the probability of becoming trapped in locally suboptimal FS subset combinations. In summary, the balancing factor proposed in this section facilitates a better equilibrium between the global exploration and local exploitation phases of the algorithm when addressing FS problems. This contributes to identifying superior FS subset combinations and mitigating the classification error rate.

3.3. Centroid Guidance Strategy

When the original COA algorithm is employed to solve FS problems after locating a potentially optimal FS subset combination region during the global exploration phase, the algorithm’s classification accuracy and convergence speed are compromised due to its inadequate local exploitation capabilities. To address this issue, this section proposes a centroid guidance strategy to enhance the algorithm’s local exploitation performance. In the proposed centroid guidance strategy, individuals are primarily guided by the centroid individuals from the current population and the centroids of the preceding three generations. Simultaneously, the fractional-order theory is incorporated to weigh the centroid individuals of populations from different generations, thereby endowing the centroid guidance strategy with adaptability and local exploitation properties. Specifically, the centroid individual within a population is computed using Equation (19).
X c e n t = i = 1 N F m a x t F i t + ε j = 1 N ( F m a x t F j t + ε ) · X i t
where X c e n t denotes the centroid individual of the population at the tth iteration, F m a x t represents the maximum fitness function value within the population at the tth iteration, F i t signifies the fitness function value of the ith individual at the tth iteration, ε is a very small constant introduced to ensure that the denominator of the equation does not equal zero, and X i t indicates the information of the ith individual at the tth iteration. Subsequently, the centroid individuals of the population at the tth, ( t 1 ) t h , ( t 2 ) t h , and ( t 3 ) t h iterations are calculated using Equation (19), denoted as X c e n t , X c e n t 1 , X c e n t 2 , and X c e n t 3 , respectively. Then, fractional-order weights are employed as the learning weights for these four centroid individuals to guide the ith individual, thereby forming the centroid guidance strategy, which is expressed as Equation (20).
X i P 2 = X i t + 1 1 ! · q · ( X c e n t X i t ) + 1 2 ! · q · ( 1 q ) · ( X c e n t 1 X i t ) + 1 3 ! · q · ( 1 q ) · ( 2 q ) · ( X c e n t 2 X i t ) + 1 4 ! · q · ( 1 q ) · ( 2 q ) · ( 3 q ) · ( X c e n t 3 X i t )
where q represents the adaptive factor, which is utilized to control the probability of the guidance process occurring, and it is calculated using Equation (21).
q = 1 1 + e t T · cos 2 · π · t T
To provide a more intuitive illustration of the implementation process of the centroid guidance strategy, we have visualized it as shown in Figure 6. In the centroid guidance strategy, firstly, the concept of centroid individuals is introduced. These individuals, being the most representative within the population, are employed to guide the population’s evolution, offering significant advantages in local exploitation. Additionally, by integrating fractional-order theory, comprehensive consideration is given to the centroid individuals of four consecutive generations. This further enhances the algorithm’s local exploitation capabilities, enabling it to achieve faster convergence speeds and higher classification accuracy when solving complex FS problems.

3.4. Execution Framework of ABCCOA Algorithm

To address the issues of inadequate global exploration capability, insufficient local exploitation capability, and imbalance between the global exploration and local exploitation phases in the COA when solving FS problems, this paper enhances its performance from various perspectives by integrating an adaptive search strategy, a balancing factor, and a centroid guidance strategy. These enhancements improve the algorithm’s global search and local exploitation capabilities while maintaining a balance between the global exploration and local exploitation phases, thereby further enhancing its classification accuracy. Consequently, this section primarily introduces the execution logic of the enhanced COA algorithm that incorporates these three learning strategies, referred to as the ABCCOA algorithm. Algorithm 2 presents the pseudo code for the execution of ABCCOA. Additionally, to visually demonstrate the execution logic of the ABCCOA algorithm, Figure 3b provides a flowchart of the ABCCOA algorithm.
Algorithm 2: Pseudo code of ABCCOA algorithm
Input: Population size ( N ), Dimension ( D ), Upper bound ( u b ) and lower bound ( l b ), Maximum of iterations ( T ).
Output: Global best solution ( X b e s t ).
1.  Initialize population based on Equation (1) and calculate the individual fitness function values of the population.
2.  for  t = 1 : T
3.    Calculate positions of iguanas using the globally best individual.
4.    Calculate Balancing Factor based on Equation (18).
5.    if  r a n d < B F
6.     Phase 1: The attack behavior of coatis (Global Exploration Phase)
7.     if  r a n d < 0.5
8.      for i = 1 : N / 2
9.       for j = 1 : D
10.        Calculate the j t h dimensional new state of the i t h individual based on Equation (4).
11.     end for
12.   end for
13.   for i = 1 + N / 2 : N
14.    for j = 1 : D
15.     Calculate the j t h dimensional new state of the i t h individual based on Equation (5).
16.    end for
17.   end for
18.  else
19.   for i = 1 : N
20.    Calculate the new state of the i t h individual based on Equation (15).
21.   end for
22.  end if
23.  Use Equation (7) to preserve the new state of individual X i .
24.  else
25.  Phase 2: The escape behavior of coatis (Local Exploitation Phase)
26.   if  r a n d < 0.5
27.   for i = 1 : N
28.    for j = 1 : D
29.     Calculate the j t h dimensional new state of the i t h individual based on Equation (8).
30.    end for
31.   end for
32.  else
33.   for i = 1 : N
34.    Calculate the new state of the i t h individual based on Equation (20).
35.   end for
36.  end if
37.  end if
38.  Use Equation (10) to preserve the new state of individual X i .
39.  Save the global best solution X b e s t .
40. end for
41. Output the global best solution X b e s t obtained by solving the optimization problem using the COA algorithm.

4. Experimental Results and Discussion on CEC2020 and Real Problems

In this section, the primary focus is on evaluating the global optimization performance of the proposed ABCCOA algorithm. Specifically, tests are conducted using the CEC2020 benchmark function suite and five real engineering optimization problems, the details of which are outlined in Table 1. To ensure experimental fairness, the population size is set to 40 and the maximum number of function evaluations is capped at 40,000. Each experiment is independently replicated 30 times without repetition. Subsequently, the experimental results are subjected to comprehensive analyses, including population diversity analysis, exploration/exploitation balance analysis, fitness function value analysis, solution stability analysis, nonparametric statistical testing, and convergence analysis. These evaluations collectively assess the optimization performance of the ABCCOA algorithm. Additionally, to visually demonstrate ABCCOA’s optimization capabilities, we compare its performance against seven state-of-the-art optimization algorithms. The parameter configurations for these comparative algorithms are provided in Table 2. To ensure the reproducibility of the experiments, the hardware configuration employed was an Intel(R) Core (TM) Ultra 5 processor with 32 GB of RAM. The software environment consisted of MATLAB R2021b running on the Windows 11 operating system.

4.1. Strategies and Parameter Analysis

In this section, we primarily analyze the sensitivity of the algorithm with respect to population size and learning strategies. Specifically, experiments were conducted on the CEC2020 test functions. To select an appropriate population size that enhances the algorithm’s performance, we set the population sizes to 10, 30, 40, 50, and 100 for separate experimental runs. The experimental results are depicted in Figure 7. It can be observed that when the population size is 40, the algorithm achieves an average ranking of 1.10, securing the top position. This is primarily because a larger population tends to weaken the algorithm’s convergence tendency, while a smaller population makes the algorithm more prone to getting trapped in local optima, thereby compromising convergence accuracy. Consequently, this study opts for a population size of 40 in subsequent experimental settings to better leverage the algorithm’s performance.
Meanwhile, to validate the effectiveness of the proposed learning strategies, we define several modified algorithms based on the CO. Specifically, we introduce the adaptive search strategy into COA to form ACOA, incorporate the balancing factor to create BCOA, adopt the centroid guidance strategy to develop CCOA, and integrate all three learning strategies to establish ABCCOA. Subsequently, we conduct experiments on these defined algorithms using the CEC2020 benchmark suite. The results are illustrated in Figure 8, which clearly shows that introducing each individual learning strategy into COA enhances the algorithm’s performance, confirming the effectiveness of each strategy in improving algorithmic performance. Furthermore, it is noteworthy that when all three learning strategies are integrated simultaneously, the algorithm exhibits superior optimization performance compared to when only a single strategy is employed. The aforementioned discussion substantiates the effectiveness of each learning strategy proposed in this paper and demonstrates that integrating all three strategies can further enhance the algorithm’s performance.

4.2. Population Diversity Analysis

In this section, we primarily analyze the population diversity of the ABCCOA algorithm during the execution of optimization problems. In the context of solving optimization problems, a higher population diversity indicates that the algorithm possesses a stronger global search capability, which helps it to escape local optima traps and thereby enhances the optimization precision of the algorithm. The experimental results are illustrated in Figure 9, where the x-axis represents the number of iterations, and the y-axis denotes the population diversity during the algorithm’s execution for the optimization problem. The blue line depicts the changes in population diversity for the ABCCOA algorithm, while the red line represents the same for the COA algorithm.
As depicted in the figure, during the early iterations of solving both unimodal and complex multimodal optimization problems, the ABCCOA algorithm exhibits higher population diversity compared to the COA algorithm. This is primarily attributed to the adaptive search strategy proposed in this study, which leverages its gap-learning mechanism to enhance the algorithm’s global search capabilities. This enables the ABCCOA algorithm to effectively locate potential optimal solution regions in the early iterations, thereby improving optimization precision. Subsequently, as shown in the figure, the ABCCOA algorithm maintains higher population diversity than the COA algorithm in the later iterations. This is advantageous for enhancing the algorithm’s ability to escape local optima traps. The introduction of the balancing factor in this study coordinates the exploration and exploitation phases, ensuring that the ABCCOA algorithm retains a certain degree of global search capability in the later iterations. This strengthens its ability to escape local optima and mitigates premature convergence. In summary, these observations confirm the effectiveness of the proposed adaptive search strategy and balancing factor in improving global exploration capabilities and enhancing the algorithm’s ability to escape local optima. They also validate that, compared to the COA algorithm, the ABCCOA algorithm achieves higher population diversity, leading to superior optimization performance.

4.3. Exploration/Exploitation Balance Analysis

In this section, we primarily analyze the balance between the exploration and exploitation phases of the ABCCOA algorithm during the execution of optimization problems. An effective optimization algorithm should typically maintain a well-balanced trade-off between exploration and exploitation. Specifically, in the early iterations, the algorithm should focus on global exploration to identify superior regions containing potential optimal solutions. In the later iterations, it should prioritize local exploitation to refine the regions identified during the exploration phase, thereby enhancing optimization precision and convergence speed. Simultaneously, the algorithm must retain a certain level of global search capability to ensure its ability to escape local optima traps. The experimental results are illustrated in Figure 10, where the x-axis represents the number of iterations, and the y-axis denotes the exploration-to-exploitation ratio. The blue line indicates the global exploration rate, while the red line represents the local exploitation rate.
As depicted in Figure 10, when solving unimodal and complex multimodal optimization problems, the ABCCOA algorithm exhibits a dominant global exploration phase in the early iterations. During this phase, the algorithm effectively identifies promising regions containing optimal solutions, primarily due to the enhanced global search capabilities conferred by the adaptive search strategy proposed in this study. As iterations progress, the global exploration rate gradually decreases while the local exploitation rate increases, with the local exploitation phase becoming dominant. This facilitates the algorithm’s refinement of the potential optimal solution regions, thereby improving optimization precision and convergence speed. This improvement is largely attributed to the centroid guidance strategy, which strengthens the algorithm’s local exploitation capabilities. Notably, even in the later iterations when local exploitation dominates, the global exploration rate remains approximately 30%. This is primarily due to the introduction of the balancing factor, which ensures a well-coordinated trade-off between the exploration and exploitation phases throughout the algorithm’s execution. Consequently, the algorithm retains a certain level of global exploration capability in the later iterations, enhancing its ability to escape local optima traps. In summary, these observations confirm that the integration of the adaptive search strategy, balancing factor, and centroid guidance strategy in the ABCCOA algorithm enables a superior balance between the exploration and exploitation phases. This effectively improves the algorithm’s optimization precision while accelerating convergence.

4.4. Fitness Function Values Analysis on CEC2020

In this section, we primarily analyze the fitness function values of the ABCCOA algorithm when solving the CEC2020 test suite to comprehensively evaluate its optimization performance. The evaluation metrics primarily include the mean and standard deviation derived from 30 independent experimental runs. The experimental results are presented in Table A1, where “Mean” denotes the average fitness function value obtained across the 30 independent experiments, and “Std” represents the standard deviation of these 30 runs. Additionally, “Mean Rank” indicates the average ranking of the algorithm based on the mean fitness function values, while “Final Rank” reflects the algorithm’s overall performance ranking determined from the “Mean Rank.”
As shown in the table, for the unimodal function F1, the ABCCOA algorithm achieves the optimal value compared to the benchmark algorithms, indicating that ABCCOA possesses stronger local exploitation capabilities. This is primarily attributed to the centroid guidance strategy proposed in this study, which effectively enhances the algorithm’s local exploitation performance, thereby improving optimization precision. Additionally, when solving complex multimodal functions F2 to F10, the ABCCOA algorithm ranks first in terms of the average fitness value across eight test functions compared to the benchmark algorithms, achieving a win rate of 88.8%. This success is largely due to the adaptive search strategy and balancing factor introduced in this study, which collectively improve the algorithm’s global search capabilities and its ability to escape local optima traps. This enables the algorithm to effectively escape modal traps and better locate regions containing potential optimal solutions. Meanwhile, Figure 11 visualizes the box plots of the algorithm’s results over 30 independent runs. As illustrated, in most cases, ABCCOA exhibits a lower box height and fewer outliers compared to the benchmark algorithms, confirming that the proposed ABCCOA algorithm demonstrates higher solution stability and concentration, ensuring its applicability in real-world optimization scenarios. Furthermore, to provide a more intuitive comparison of algorithm performance, Figure 12 visualizes the mean ranking plot. As shown, ABCCOA achieves a mean rank of 1.70, securing the first overall rank with the lowest bar height, outperforming the second-ranked IMODE by approximately 39.2%. In summary, these observations confirm that the adaptive search strategy, balancing factor, and centroid guidance strategy proposed in this study significantly enhance the algorithm’s performance. They also validate that the proposed ABCCOA is a robust optimization algorithm with efficient optimization capabilities, offering higher solution stability and broader applicability.

4.5. Nonparametric Test Analysis on CEC2020

The optimization performance of the ABCCOA algorithm was primarily evaluated using numerical results in the preceding analysis. However, since outliers can influence numerical outcomes to some extent, this section employs nonparametric statistical tests to mitigate their impact on the algorithm’s numerical results. Specifically, we conduct Wilcoxon and Friedman nonparametric tests with a significance level of 0.05 on the results of 30 independent experimental runs to further assess the algorithm’s optimization performance. The results of the Wilcoxon nonparametric test are presented in Table A2, where “+” indicates that the performance of the benchmark algorithm is significantly superior to that of the proposed ABCCOA algorithm, “−” denotes that the benchmark algorithm’s performance is significantly inferior to ABCCOA, and “=” signifies no statistically significant difference in performance between the benchmark algorithm and ABCCOA. The outcomes of the Friedman nonparametric test are illustrated in Figure 13.
As shown in the table, the COA, GGO, and QAGO algorithms exhibit significantly inferior performance compared to the ABCCOA algorithm on nine test functions, representing a 90% ratio. The IPOA and PRO algorithms underperform ABCCOA on all ten test functions, achieving a 100% ratio. The QHDBO, HEOA, and STOA algorithms demonstrate weaker performance than ABCCOA on eight test functions, corresponding to an 80% ratio. In summary, the ABCCOA algorithm achieves superior optimization performance over the benchmark algorithms due to its advantageous learning strategies, positioning it as an efficient algorithm. However, it is noteworthy that the COA, GGO, QAGO, QHDBO, and HEOA algorithms outperform ABCCOA on the F8 test function, indicating that ABCCOA’s performance could be further enhanced for certain specific optimization problems. From a comprehensive perspective, across the 90 conducted tests, ABCCOA achieves victory in 78 cases, yielding an 86.6% success rate. This suggests that, despite the occasional shortcomings in specific optimization scenarios, ABCCOA is a robust algorithm with highly efficient optimization capabilities. Additionally, as illustrated by the Friedman nonparametric test results in Figure 13, ABCCOA attains a Friedman mean rank of 1.95, securing the top position. It outperforms the second-ranked IMODE algorithm by 29.8%, demonstrating superior optimization performance. The conclusions drawn from the nonparametric tests collectively validate that the adaptive search strategy, balancing factor, and centroid guidance strategy proposed in this study significantly enhance the algorithm’s performance. Furthermore, the ABCCOA algorithm can be regarded as an efficient optimization algorithm with robust performance.

4.6. Convergence Analysis on CEC2020

The preceding sections primarily analyzed the ABCCOA algorithm’s population diversity, exploration/exploitation balance, fitness function values, solution stability, and nonparametric test results, collectively validating the algorithm’s efficient optimization capabilities. However, beyond these metrics, the convergence speed of the algorithm is also critical, as it reflects the algorithm’s practical applicability in real-world optimization scenarios. Therefore, this section focuses on analyzing the convergence behavior of the algorithm when solving the CEC2020 test function suite. The experimental results are depicted in Figure 14, where the x-axis represents the number of function evaluations, and the y-axis denotes the common logarithm (base 10) of the fitness function values.
As illustrated in the figure, all algorithms exhibit regional convergence behavior and effectively solve the optimization problems. However, it is noteworthy that after approximately 15,000 function evaluations, the ABCCOA algorithm demonstrates a distinct advantage, outperforming the benchmark algorithms in terms of convergence precision. This superiority is primarily attributed to the strategies proposed in this study, which enable the algorithm to rapidly locate regions containing potential optimal solutions, thereby enhancing convergence precision. Moreover, as iterations progress, the ABCCOA algorithm’s leading margin gradually widens, ultimately achieving stable and regionally consistent convergence behavior with a faster convergence rate. This is mainly due to the enhancement of the algorithm’s exploitation capabilities by the proposed strategies, which accelerates convergence. In summary, these observations confirm that the adaptive search strategy, balancing factor, and centroid guidance strategy proposed in this study significantly improve the algorithm’s convergence properties. Consequently, the ABCCOA algorithm can be regarded as an algorithm with favorable convergence characteristics and practical applicability.
The analyses conducted in the preceding sections—encompassing population diversity, exploration/exploitation balance, fitness function values, solution stability, nonparametric testing, and convergence behavior-validate the enhancing effects of the three proposed learning strategies on the ABCCOA algorithm. Furthermore, these analyses confirm that, compared to the benchmark algorithms, ABCCOA is a robust optimization tool characterized by efficient optimization performance, high solution stability, and rapid convergence properties.

4.7. Fitness Function Values Analysis on Real Problems

The optimization performance of the ABCCOA algorithm has already been validated on synthetic functions, as mentioned above. Additionally, to further assess the algorithm’s performance, this section evaluates the ABCCOA algorithm on five real-world engineering optimization problems, including the 10-bar truss design problem (RP1), Himmel Blau’s function (RP2), tension/compression spring design case 1 (RP3), tension/compression spring design case 2 (RP4), and the welded beam design problem (RP5). The experimental results are presented in Table A3. As can be seen from the table, the ABCCOA algorithm achieves optimal fitness function values across all five engineering optimization problems considered. Compared to the comparative algorithms, it demonstrates the highest win rate, further corroborating that the ABCCOA algorithm is an efficient optimizer with high-performance capabilities.

5. Experimental Results and Discussion on FS Problems

The preceding sections evaluated the optimization performance of the ABCCOA algorithm on the CEC2020 test function suite, confirming its efficacy as a high-performance optimization algorithm. In this section, the proposed ABCCOA algorithm is applied to solve FS problems. The primary objective is to minimize redundancy in the original dataset by identifying a compact subset of highly informative features, thereby enhancing classification accuracy while reducing computational costs. Specifically, the experiments were conducted on 27 FS datasets, and the performance of ABCCOA was compared against seven state-of-the-art algorithms. The details of the 27 FS datasets are provided in Table 3, and the benchmark algorithms are listed in Table 4. To ensure experimental fairness, the population size was set to 40, the maximum number of iterations was set to 100, and each experiment was independently repeated 30 times to aggregate the results. The FS performance of the ABCCOA algorithm was objectively evaluated by comprehensively considering the fitness function values, classification accuracy, feature subset size, and runtime metrics.

5.1. FS Problems Model

In this section, we focus on formulating the model for feature selection (FS) problems. FS plays a pivotal role in machine learning and data mining, as its primary objective is to reduce redundancy in the original dataset, thereby decreasing dataset complexity while enhancing classification accuracy. This process mitigates computational complexity in subsequent machine learning and data mining tasks and improves the generalization capability of model training. Specifically, the FS process involves searching through a vast space of feature subset combinations to identify a compact yet effective set of features that adequately represent the entire dataset, ultimately boosting model classification accuracy. To achieve this, we employ Equation (22) as the objective function for the FS problem.
min f ( Y i ) = λ 1 · e r r o r + λ 2 · R / n
where Y i denotes the ith FS subset combination, error represents the classification error rate when using Y i , R indicates the size of the FS subset combination, n denotes the number of features in the original dataset, λ 1 is a constant within the interval [0, 1], and λ 2 = 1 λ 1 . In this paper, λ 1 is set to 0.9.
The FS problem is a classic combinatorial optimization task that involves selecting a subset of features from the original dataset and subsequently evaluating the quality of this subset by computing an objective function value using the selected features. In this section, we apply the proposed ABCCOA algorithm to solve the FS problem. Consequently, it is necessary to convert the real-valued solutions generated by the ABCCOA algorithm during iterations into binary (0–1) values, which indicate whether a feature is selected or not. Below, we detail the step-by-step procedure for selecting feature subsets and computing the fitness function values when solving the FS problem using the ABCCOA algorithm. The flowchart is illustrated as shown in Figure 15.
  • Step 1: Using Equation (23), we convert real-valued individual X i = ( x i , 1 , x i , 2 , , x i , j , , x i , D i m ) in the ABCCOA algorithm into a binary individual Y i = ( y i , 1 , y i , 2 , , y i , j , , y i , D i m ) .
    y i , j = 1 i f   x i , j < 0 . 5   0 i f   x i , j 0 . 5   ,   i = 1 , 2 , , N ,   j = 1 , 2 , , D i m .
  • Step 2: The binary individual Y i = ( y i , 1 , y i , 2 , , y i , j , , y i , D i m ) is utilized to select a feature subset combination from the original dataset. Here, y i , j = 1 indicates that the jth feature in the original dataset is selected in the ith feature subset combination, whereas y i , j = 0 indicates that the jth feature is not selected.
  • Step 3: The classification accuracy of the selected feature subset combinations is computed using a K-Nearest Neighbors (KNN) classifier, where K is set to 5.
  • Step 4: The objective function value for the feature subset combination is computed using Equation (22) with the information output by the KNN classifier.

5.2. Fitness Function Values Analysis of FS Problems

In this section, the focus is on analyzing the fitness function values of the ABCCOA algorithm when solving FS problems. Specifically, the algorithm was tested on 27 FS datasets encompassing low-dimensional, medium-dimensional, and high-dimensional cases while being compared with seven other algorithms known for their high efficiency. The experimental results are presented in Table A4. Here, “MIN,” “AVG,” and “MAX” denote the minimum, average, and maximum fitness function values obtained from 30 independent experimental runs, respectively. “Mean Rank” represents the average ranking of the algorithm across the 27 FS problem instances, and “Final Rank” indicates the algorithm’s ultimate ranking based on the “Mean Rank” metric.
As can be observed from the table, when solving low-dimensional FS problems, the ABCCOA algorithm ranks first on seven FS problems in terms of the optimal fitness function value, achieving a win rate of 87.5%. This is primarily attributed to the introduction of the centroid guidance strategy in this paper, which enhances the algorithm’s local exploitation capability, enabling it to explore more suitable combinations of feature subsets. Meanwhile, in terms of the average fitness function value, the ABCCOA algorithm ranks first on eight FS problems, with a ratio of 100%, indicating its high solution stability when solving FS problems. Additionally, in terms of the worst fitness function value, the ABCCOA algorithm ranks first on six FS problems, with a ratio of 75%, demonstrating its higher fault tolerance compared to the comparative algorithms when solving FS problems. In summary, it can be concluded that, due to the incorporation of the three learning strategies proposed in this paper, the ABCCOA algorithm is capable of obtaining excellent feature subset combinations for low-dimensional FS problems while showcasing its robust solution stability and fault tolerance.
When solving medium-dimensional FS problems, algorithms often require better global search capabilities to effectively explore superior potential optimal regions. As evident from the table, in terms of the optimal fitness function value for medium-dimensional FS problems, the ABCCOA algorithm ranks first on eight FS problems, accounting for 100%, demonstrating its stronger global optimization capability. Furthermore, in terms of the average fitness function value, the ABCCOA algorithm ranks first on seven FS problems, with a ratio of 87.5%, exhibiting greater solution stability that facilitates its implementation in practical application scenarios. In terms of the worst fitness function value, the ABCCOA algorithm ranks first on six FS problems, with a ratio of 75%, showcasing its stronger solution fault tolerance and robustness.
When addressing high-dimensional FS problems where the number of features exceeds 100, algorithms demand a higher balance between global exploration and local exploitation. Firstly, in terms of the optimal fitness function value, the ABCCOA algorithm ranks first on nine FS problems, with a ratio of 81.8%, demonstrating its stronger global optimization capability. This is largely attributed to the balance factor proposed in this paper, which achieves a good equilibrium between the exploration and local exploitation phases, thereby enhancing its search accuracy. Additionally, in terms of the average fitness function value, the ABCCOA algorithm ranks first on nine FS problems, with a ratio of 81.8%, achieving better overall FS performance compared to the comparative algorithms and exhibiting stronger solution stability. In terms of the worst fitness function value, the ABCCOA algorithm ranks first on eight FS problems, with a ratio of 72.7%, demonstrating its stronger solution fault tolerance and robustness compared to the comparative algorithms.
Meanwhile, Figure 16 visualizes the average rankings of the algorithms across different metrics. As can be seen from the figure, the ABCCOA algorithm achieves significant advantages in terms of the optimal, average, and worst fitness function values with the lowest bar heights. In conclusion, it can be concluded that the ABCCOA algorithm proposed in this paper is a robust and stable algorithm with efficient feature search performance, exhibiting higher solution stability and fault tolerance.

5.3. Classification Accuracy and Feature Subset Size Analysis of FS Problems

The analysis of the fitness function values of the ABCCOA algorithm when solving FS problems in the previous subsection has confirmed the algorithm’s high efficiency. The FS problem aims to extract the most representative feature subset to characterize the entire dataset, thereby improving classification accuracy. Therefore, in this section, we primarily evaluate the two most crucial metrics in FS problems: classification accuracy and the size of the feature subset. The experimental results for classification accuracy are presented in Table A5, while those for the size of the feature subset are shown in Table A6. Here, “Mean Rank” denotes the average ranking of the algorithm across 27 FS problems, and “Final Rank” represents the algorithm’s ultimate ranking based on the “Mean Rank” metric.
As can be observed from Table A5, when solving low-dimensional FS problems, the ABCCOA algorithm achieved the optimal classification accuracy on seven FS problems, accounting for 87.5%. Compared to the comparative algorithms, it obtained a more superior combination of feature subsets, thereby achieving higher classification accuracy with fewer features. This is primarily attributed to the strategies proposed in this paper, which effectively enhance the algorithm’s local exploitation capability. Meanwhile, as revealed in Table A6, all algorithms were capable of effectively reducing the size of the feature subsets. However, the ABCCOA algorithm did not have a distinct advantage in terms of feature reduction. This is mainly because other algorithms filtered out highly informative features, which, despite their advantage in feature subset size, resulted in a significant loss of classification accuracy. This also highlights the insufficient global trade-off capability of the comparative algorithms, leading to the deletion of useful feature information. The ABCCOA algorithm, due to its higher global search capability, was able to strike a reasonable balance between classification accuracy and feature subset size, thereby preserving effective features. When addressing medium-dimensional FS problems, the ABCCOA algorithm ranked first on four FS problems, accounting for 50%, demonstrating a relatively high advantage compared to the comparative algorithms. Nevertheless, it must be acknowledged that the ABCCOA algorithm has certain limitations on specific datasets, and its performance still needs improvement. However, from a comprehensive perspective, it exhibits higher overall performance compared to the comparative algorithms. This is largely attributed to the learning strategies proposed in this paper, which effectively enhance the algorithm’s global search performance. Furthermore, when solving high-dimensional FS problems, the ABCCOA algorithm ranked first in terms of accuracy on eight FS problems, accounting for 72.7%, achieving higher performance compared to the comparative algorithms and effectively preserving key features. Meanwhile, Figure 17 and Figure 18, respectively, illustrate the average rankings of the algorithms in terms of classification accuracy and feature subset size. It can be observed that the ABCCOA algorithm holds a significant advantage in terms of classification accuracy. This is primarily due to the strategies proposed in this paper, which improve the algorithm’s global optimization capability and, consequently, enhance its FS performance. Additionally, in terms of the feature subset size metric, the ABCCOA algorithm ranks first, effectively reducing the dimensionality of features. In summary, the ABCCOA algorithm is capable of effectively reducing the dimensionality of the features in datasets when solving FS problems, thereby improving the algorithm’s classification accuracy. It can be considered a promising FS method.

5.4. Runtime Analysis of FS Problems

In this section, we primarily analyze the actual runtime of the ABCCOA algorithm when solving FS problems, which mainly reflects the algorithm’s real-time performance in practical applications. The experimental results are presented in Table A7, where “Mean Rank” denotes the average ranking of the algorithm in terms of runtime, and “Final Rank” represents the algorithm’s ultimate ranking based on the “Mean Rank” metric. As can be observed from the table, the ABCCOA algorithm achieved the shortest runtime on 25 FS problems, with a win rate of 92.5%. To more intuitively demonstrate the algorithm’s advantage in terms of runtime, Figure 19 displays a bar chart of the algorithms’ average rankings. From the figure, it is visually evident that the ABCCOA algorithm has a shorter bar length, indicating the shortest runtime. This confirms that the ABCCOA algorithm proposed in this paper is a real-time FS method.

5.5. Comprehensive Analysis of FS Problems

In this section, a comprehensive analysis of the ABCCOA algorithm’s performance is conducted. By synthesizing the average rankings of the involved metrics, including the optimal fitness function value, average fitness function value, worst fitness function value, classification accuracy, feature subset size, and runtime, a stacked bar chart, as depicted in Figure 20, is formed. As can be observed from the figure, although the ABCCOA algorithm performs slightly worse than MCOA and MSAACO in terms of the feature subset size, its overall stacked bar exhibits the lowest height, indicating the best overall performance. This can be regarded as an FS algorithm with comprehensive and efficient performance.

6. Conclusions and Future Works

This paper addresses the limitations of the original COA in solving global optimization and FS problems, which stem from insufficient global exploration, inadequate local exploitation, and an imbalance between the exploration and exploitation phases. These issues often lead the algorithm to converge to suboptimal feature subset combinations, resulting in low classification accuracy and suboptimal optimization precision. To overcome these challenges, an enhanced COA algorithm, termed ABCCOA, is proposed by integrating three learning strategies. Firstly, to tackle the problem of inadequate global search performance in the COA algorithm when solving FS problems, which arises from the exponential increase in the number of feature subset combinations and consequently reduces classification accuracy, an adaptive search strategy is introduced. This strategy combines individual learning capabilities with the learnability of disparities, effectively enhancing the algorithm’s global exploration capabilities in solving FS problems and thereby improving the classification accuracy of feature subsets. Secondly, to address the issue of the COA algorithm’s tendency to converge to suboptimal feature subset combinations due to an imbalance between the exploration and exploitation phases when solving FS problems, a balancing factor is proposed for control. This factor’s phase is modulated using sine and cosine functions, while its dynamic adjustment is achieved through the integration of arcsine and arccosine functions. This approach ensures a well-balanced transition between the two phases, reducing the likelihood of the algorithm converging to suboptimal FS subsets. Lastly, to counter the problem of increased classification error rates resulting from insufficient local exploitation in the COA algorithm when solving FS problems, a centroid guidance strategy is introduced to enhance the algorithm’s local exploitation capabilities. By leveraging the centroids of individuals from the previous four generations and utilizing fractional-order historical memory, this strategy not only reduces the classification error rate of FS subsets but also accelerates convergence speed. Subsequently, the ABCCOA algorithm is applied to the CEC2020 test functions and five engineering optimization problems, demonstrating an optimization win rate exceeding 90% and a faster convergence speed compared to comparative algorithms, thereby confirming its efficient optimization performance. Furthermore, the ABCCOA algorithm, with its proven efficient optimization capabilities, was employed to solve 27 FS problems to evaluate its FS performance. The experimental results indicate that, compared to comparative algorithms, the ABCCOA algorithm exhibits significant advantages in terms of the optimal fitness function value, average fitness function value, worst fitness function value, classification accuracy, feature subset size, and runtime, achieving remarkable FS performance. Consequently, ABCCOA can be regarded as an efficient and robust FS algorithm.
However, despite ABCCOA achieving the best overall performance in solving FS problems, there is still room for improvement in its performance on certain specific FS datasets. Meanwhile, although the algorithm proposed in this paper may still offer advantages in combinatorial optimization processes within forward-looking fields, no further extended testing has been conducted in this regard. Therefore, the future work outlined in this paper primarily focuses on the following three aspects: 1. Developing strategies with more efficient search capabilities for specific FS problems to enhance their classification accuracy. 2. Extending the application of the ABCCOA algorithm to a broader range of combinatorial optimization problems, thereby expanding the application scenarios of the algorithm proposed in this paper. Examples include airline scheduling and photovoltaic parameter identification. 3. This paper has only developed a single-objective optimization model for the algorithm. In subsequent work, we will also extend it to a multi-objective model to further enhance the algorithm’s performance.

Author Contributions

Conceptualization, Q.C. and S.Y.; methodology, Q.C.; software, Q.C.; validation, S.Y. and Y.F.; formal analysis, Q.C. and Y.F.; investigation, S.Y. and Y.F.; resources, Q.C.; data curation, S.Y. and Y.F.; writing—original draft preparation, Q.C.; writing—review and editing, Q.C., S.Y. and Y.F.; visualization, Q.C. and S.Y.; supervision, Y.F.; project administration, Y.F.; funding acquisition, Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All experimental data can be obtained by contacting the corresponding author.

Acknowledgments

We greatly appreciate the efforts made by the reviewers and editorial team for our article.

Conflicts of Interest

The author declares that there are no conflicts of interest in this article.

Appendix A

Table A1. The fitness function value on the CEC2020 test function.
Table A1. The fitness function value on the CEC2020 test function.
ProblemsMetricsCOAGGOIPOAPROQAGOQHDBOHEOASTOAIMODEABCCOA
CEC2020_F1Mean3.1909 × 1032.8022 × 1052.3920 × 1034.0006 × 1033.7112 × 1031.4334 × 1039.2230 × 1051.7426 × 1031.0000 × 1021.0000 × 102
Std3.3719 × 1031.2304 × 1062.4564 × 1033.0737 × 1033.5303 × 1031.5126 × 1036.8025 × 1052.1007 × 1030.0000 × 1001.0556 × 10−14
CEC2020_F2Mean1.9082 × 1031.8364 × 1031.4853 × 1031.4333 × 1031.9506 × 1031.6920 × 1031.2850 × 1031.8119 × 1031.3453 × 1031.2784 × 103
Std3.3197 × 1023.3372 × 1022.1359 × 1021.8320 × 1022.6218 × 1021.6673 × 1021.0412 × 1022.6013 × 1022.1524 × 1021.1347 × 102
CEC2020_F3Mean7.6108 × 1027.3357 × 1027.2055 × 1027.2118 × 1027.4366 × 1027.2806 × 1027.2263 × 1027.2482 × 1027.1801 × 1027.1718 × 102
Std1.5046 × 1011.2631 × 1016.1067 × 1004.3600 × 1001.8048 × 1016.4219 × 1004.5734 × 1006.8581 × 1003.8629 × 1003.5531 × 100
CEC2020_F4Mean1.9045 × 1031.9033 × 1031.9011 × 1031.9017 × 1031.9017 × 1031.9011 × 1031.9022 × 1031.9016 × 1031.9009 × 1031.9007 × 103
Std2.9976 × 1001.5294 × 1003.1868 × 10−18.0634 × 10−16.9471 × 10−15.5801 × 10−15.4287 × 10−18.4758 × 10−13.4417 × 10−12.1239 × 10−1
CEC2020_F5Mean1.0452 × 1041.3219 × 1044.7648 × 1031.9134 × 1052.2577 × 1033.6506 × 1033.8663 × 1043.1948 × 1031.8119 × 1031.7500 × 103
Std1.0876 × 1041.5027 × 1042.6024 × 1032.7411 × 1052.5618 × 1029.3224 × 1023.6543 × 1046.6865 × 1026.4939 × 1015.4569 × 101
CEC2020_F6Mean1.7598 × 1031.8005 × 1031.6973 × 1031.7170 × 1031.7293 × 1031.6193 × 1031.6773 × 1031.6311 × 1031.6249 × 1031.6113 × 103
Std1.0650 × 1021.1254 × 1026.9794 × 1017.4794 × 1011.0833 × 1023.1798 × 1017.9426 × 1014.9920 × 1014.7920 × 1012.9837 × 101
CEC2020_F7Mean9.1304 × 1035.0144 × 1032.4814 × 1031.4209 × 1042.4609 × 1032.8087 × 1031.1044 × 1042.4544 × 1032.1410 × 1032.1062 × 103
Std7.1463 × 1037.4174 × 1031.7018 × 1022.5252 × 1042.3036 × 1022.8361 × 1021.0598 × 1041.1367 × 1028.0007 × 1012.1880 × 101
CEC2020_F8Mean2.3080 × 1032.3057 × 1032.3100 × 1032.3100 × 1032.3072 × 1032.3077 × 1032.3030 × 1032.3068 × 1032.3069 × 1032.3100 × 103
Std1.0866 × 1011.8825 × 1011.0581 × 10−122.3739 × 10−31.5463 × 1011.2572 × 1012.3518 × 1011.7244 × 1011.7146 × 1014.5475 × 10−13
CEC2020_F9Mean2.7744 × 1032.7070 × 1032.7354 × 1032.7440 × 1032.7310 × 1032.6821 × 1032.7374 × 1032.7055 × 1032.7344 × 1032.6550 × 103
Std1.9217 × 1011.0097 × 1024.4992 × 1014.7359 × 1017.9166 × 1019.2976 × 1015.8952 × 1017.8665 × 1014.4548 × 1011.1648 × 102
CEC2020_F10Mean2.9306 × 1032.9390 × 1032.9337 × 1032.9304 × 1032.9309 × 1032.9197 × 1032.9220 × 1032.9315 × 1032.9212 × 1032.9026 × 103
Std4.7628 × 1012.6650 × 1012.1468 × 1012.4738 × 1012.1773 × 1012.2733 × 1012.0299 × 1012.1700 × 1012.3416 × 1011.3847 × 101
Mean Rank 8.207.505.707.106.404.206.104.702.801.70
Final Rank 10958736421
Table A2. Wilcoxon nonparametric test results on CEC2020 test function.
Table A2. Wilcoxon nonparametric test results on CEC2020 test function.
ProblemsCOAGGOIPOAPROQAGOQHDBOHEOASTOAIMODE
CEC2020_F1=
CEC2020_F2=
CEC2020_F3
CEC2020_F4
CEC2020_F5
CEC2020_F6=
CEC2020_F7
CEC2020_F8++++++=
CEC2020_F9==
CEC2020_F10
+/−/=1/9/01/9/00/10/00/10/01/9/01/8/11/8/11/8/10/7/3
Table A3. The fitness function value in real-world optimization problems.
Table A3. The fitness function value in real-world optimization problems.
ProblemsMetricsCOAGGOIPOAPROQAGOQHDBOHEOASTOAIMODEABCCOA
RP1Mean5.261 × 1025.347 × 1025.257 × 1025.599 × 1025.262 × 1025.247 × 1025.249 × 1025.360 × 1025.246 × 1025.245 × 102
Std1.981 × 1008.537 × 1001.596 × 1001.022 × 1012.812 × 1001.118 × 1001.179 × 1006.981 × 1006.270 × 10−021.919 × 10−2
RP2Mean−3.067 × 104−3.067 × 104−3.067 × 104−3.005 × 104−3.067 × 104−3.067 × 104−3.067 × 104−3.067 × 104−3.067 × 104−3.067 × 104
Std7.768 × 10−73.223 × 10−43.374 × 10−112.249 × 1021.100 × 10−111.110 × 10−111.110 × 10−113.598 × 10−21.042 × 10−61.392 × 10−6
RP3Mean1.280 × 10−21.274 × 10−21.267 × 10−22.226 × 10−21.268 × 10−21.267 × 10−21.267 × 10−21.314 × 10−21.267 × 10−21.267 × 10−2
Std1.402 × 10−41.021 × 10−45.570 × 10−68.296 × 10−32.380 × 10−55.254 × 10−65.682 × 10−84.130 × 10−41.170 × 10−74.378 × 10−13
RP4Mean2.691 × 1002.916 × 1002.664 × 1003.429 × 1002.660 × 1002.673 × 1002.668 × 1002.735 × 1002.660 × 1002.659 × 100
Std5.083 × 10−23.480 × 10−11.415 × 10−24.178 × 10−17.474 × 10−35.196 × 10−21.761 × 10−29.737 × 10−27.470 × 10−31.712 × 10−6
RP5Mean1.670 × 1001.685 × 1001.670 × 1001.974 × 1001.670 × 1001.670 × 1001.670 × 1001.737 × 1001.670 × 1001.670 × 100
Std2.497 × 10−46.173 × 10−21.187 × 10−81.204 × 10−15.800 × 10−102.258 × 10−162.182 × 10−164.694 × 10−21.950 × 10−71.934 × 10−16
Mean Rank 5.806.603.2010.003.403.002.407.201.601.00
Final Rank 78510643921
Table A4. The fitness function value on FS problem.
Table A4. The fitness function value on FS problem.
DatasetsMetricsCOAFATAHOAMCOAMSAACOSBOAPLOABCCOA
AggregationMIN0.1000.1000.1000.1000.1000.1060.1060.100
AVG0.1000.1000.1000.1000.1000.1060.1060.100
MAX0.1000.1000.1000.1000.1000.1060.1060.100
BananaMIN0.2110.1960.1920.1990.2120.1950.1970.188
AVG0.2110.1960.1920.1990.2120.1950.1970.188
MAX0.2110.1960.1920.1990.2120.1950.1970.188
IrisMIN0.0250.0800.0250.0500.0500.0750.1150.025
AVG0.0250.0880.0250.0500.0510.0750.1150.025
MAX0.0250.1100.0250.0500.0550.0750.1150.025
BupaMIN0.2850.2850.3330.3330.3070.3500.3110.285
AVG0.2960.2920.3340.3330.3130.3500.3110.285
MAX0.3440.3110.3370.3330.3370.3500.3110.285
GlassMIN0.3120.2900.2690.2590.2590.3120.2480.226
AVG0.3140.3080.2690.2650.2600.3230.2550.250
MAX0.3330.3560.2690.2800.2700.3440.3120.270
BreastcancerMIN0.0590.0590.0530.0530.0550.0530.0530.048
AVG0.0600.0600.0630.0530.0550.0530.0530.050
MAX0.0660.0660.0680.0530.0550.0530.0550.059
LipidMIN0.2450.2470.2530.2290.2450.2370.2580.222
AVG0.2470.2630.2530.2400.2490.2520.2620.222
MAX0.2510.2680.2570.2610.2650.2630.2680.222
HeartEWMIN0.0900.1400.1900.1470.1380.1130.1220.105
AVG0.1100.1650.2480.1640.1420.1230.1290.107
MAX0.1470.1900.2810.1880.1470.1540.1560.122
ZooMIN0.0890.0700.0380.0700.0760.0380.0440.031
AVG0.1210.0950.0410.0730.1140.0710.0650.035
MAX0.1340.1090.0500.0760.1340.0830.1150.044
VoteMIN0.0290.0290.0310.0350.0460.0480.0500.027
AVG0.0420.0490.0350.0450.0500.0610.0540.027
MAX0.0480.0600.0500.0700.0580.0980.0620.027
CongressMIN0.0170.0390.0370.0270.0600.0480.0420.017
AVG0.0170.0520.0380.0270.0650.0480.0480.025
MAX0.0170.0620.0420.0270.0680.0480.0560.035
LymphographyMIN0.0640.1100.0590.0840.1070.0900.0840.059
AVG0.0860.1320.0760.0900.1390.1130.1030.067
MAX0.1100.1630.1070.0950.1720.1320.1260.084
VehicleMIN0.2570.2620.2410.2670.2570.2360.2470.236
AVG0.2750.2850.2660.2790.2730.2510.2570.251
MAX0.2880.3000.2890.2940.2950.2630.2670.263
WDBCMIN0.0610.0370.0690.0660.0460.0680.0560.037
AVG0.0660.0530.0860.0670.0490.0890.0700.046
MAX0.0750.0640.1020.0700.0580.1030.0810.055
BreastEWMIN0.0390.0360.0370.0350.0390.0410.0390.021
AVG0.0580.0450.0430.0390.0470.0500.0540.038
MAX0.0670.0510.0550.0560.0570.0610.0640.059
SonarEWMIN0.0540.1250.0570.0490.0340.0350.0440.015
AVG0.0730.1520.0740.0770.0530.0500.0730.038
MAX0.0930.1890.0890.0910.0810.0620.0990.050
LibrasMIN0.1530.1380.2380.2040.1780.2710.1640.092
AVG0.1700.1580.2450.2280.2000.2960.1920.129
MAX0.1940.1780.2590.2620.2180.3280.2240.144
HillvalleyMIN0.3410.3130.3760.2610.3180.2940.3740.266
AVG0.3600.3230.3910.2810.3310.3100.3960.298
MAX0.3740.3370.4040.3040.3450.3220.4160.323
MuskMIN0.0630.1090.0760.0200.0180.0880.0700.064
AVG0.0930.1330.0980.0330.0280.1090.0900.080
MAX0.1120.1450.1120.0400.0450.1270.1090.088
CleanMIN0.0670.0730.0840.0750.0200.0470.0750.017
AVG0.0890.0910.0940.1030.0440.0780.0830.027
MAX0.1260.1050.1000.1280.0660.1020.0940.034
SemeionMIN0.1140.1060.1310.0700.1190.0890.0800.063
AVG0.1230.1150.1390.0820.1280.0970.0880.080
MAX0.1330.1230.1450.0960.1350.1100.0930.106
MadelonMIN0.2440.2300.2370.1580.1070.1900.1830.091
AVG0.2680.2620.2510.2010.1210.2090.2040.109
MAX0.2870.2810.2590.2590.1360.2250.2430.125
IsoletMIN0.1610.1510.1780.1640.0650.1360.1340.060
AVG0.1720.1620.1840.1760.0860.1470.1530.075
MAX0.1840.1770.1890.1840.1100.1570.1780.097
LungMIN0.0720.0510.0510.1150.0870.0470.0380.021
AVG0.0930.0720.0810.1220.0950.0820.0710.042
MAX0.1240.0930.0980.1350.1130.1200.1130.063
MLLMIN0.0810.0710.0820.0720.1010.1230.0530.041
AVG0.0900.1000.1010.0960.1250.1470.0740.061
MAX0.1010.1170.1250.1230.1450.1600.0920.082
OvarianMIN0.1100.0610.1010.0710.0620.0410.0510.012
AVG0.1230.0930.1180.0950.0830.0720.0790.020
MAX0.1800.1180.1230.1140.0980.0810.1000.034
CNSMIN0.1180.0840.0710.0880.0800.1170.1010.061
AVG0.1330.1080.0990.1150.1090.1360.1190.086
MAX0.1530.1260.1190.1420.1370.1500.1250.091
Mean RankMIN4.704.705.044.444.635.004.631.19
AVG4.855.265.044.444.675.154.701.15
MAX5.075.114.444.334.375.074.851.63
Final RankMIN55823731
AVG58623741
MAX68423651
Table A5. Classification accuracy on FS problems.
Table A5. Classification accuracy on FS problems.
DatasetsCOAFATAHOAMCOAMSAACOSBOAPLOABCCOA
Aggregation100.00100.00100.00100.00100.0099.3699.36100.00
Banana87.6489.3489.8188.9687.5589.4389.2590.19
Iris100.0096.33100.00100.0099.67100.0090.00100.00
Bupa73.0472.9066.8166.6769.2866.6771.0173.04
Glass69.0569.5273.8175.0076.1968.5775.0076.19
Breastcancer97.1296.7696.4097.8496.4097.8497.5597.12
Lipid74.3173.6274.3175.5275.0074.6672.8477.59
HeartEW90.5685.3774.6385.1987.4190.3788.8992.41
Zoo90.5093.50100.0095.0091.0095.5097.00100.00
Vote96.6797.3699.0896.5596.7896.6796.3297.70
Congress98.8596.7897.1397.7094.0295.4098.0598.85
Lymphography94.1488.9795.5293.1087.9391.3891.7295.86
Vehicle73.4971.9574.3272.4973.4976.4575.5675.98
WDBC94.0796.5592.4893.6395.7592.2193.5496.64
BreastEW96.3798.2399.2998.0596.9997.9696.9999.20
SonarEW95.6185.3798.2993.9096.1098.5495.1298.05
Libras83.7584.7279.3176.5379.5871.3982.6487.78
Hillvalley61.8265.3762.7370.5065.3770.3360.5067.85
Musk92.1187.7993.4798.3299.1692.7494.9598.32
Clean91.0593.4796.4291.0597.6895.7995.4799.26
Semeion91.4291.6091.7094.5690.5394.3495.3194.65
Madelon74.2974.5279.3780.3888.8582.2582.5489.29
Isolet84.3485.0286.9883.2292.9989.1688.2693.99
Lung89.80 92.12 91.20 86.80 89.50 91.09 92.21 95.38
MLL90.01 89.11 88.93 89.38 86.23 83.70 91.80 93.25
Ovarian86.38 89.71 86.98 89.56 90.81 92.01 91.23 97.79
CNS85.38 88.13 89.12 87.30 88.03 84.98 86.89 90.50
Mean Rank5.07 5.11 4.26 4.70 4.70 4.63 4.59 1.44
Final Rank7 8 2 5 5 4 3 1
Table A6. Feature subset size on FS problem.
Table A6. Feature subset size on FS problem.
DatasetsCOAFATAHOAMCOAMSAACOSBOAPLOABCCOA
Aggregation2.002.002.002.002.002.002.002.00
Banana2.002.002.002.002.002.002.002.00
Iris1.002.201.002.001.903.001.001.00
Bupa3.202.902.102.002.203.003.003.00
Glass3.203.003.003.604.103.602.703.20
Breastcancer3.102.802.803.002.003.002.802.20
Lipid1.602.602.202.002.402.401.802.00
HeartEW3.204.302.604.003.704.703.805.00
Zoo5.605.906.504.505.304.806.005.60
Vote1.904.104.302.303.305.003.301.00
Congress1.003.702.001.001.801.004.902.30
Lymphography5.905.806.505.005.406.305.105.40
Vehicle6.505.906.305.706.207.106.606.30
WDBC3.806.605.402.803.105.603.704.70
BreastEW7.608.6011.106.406.109.408.009.20
SonarEW19.9012.3034.9013.3010.8021.9017.6012.20
Libras21.7018.3052.6014.9014.8034.8032.1016.90
Hillvalley16.6011.0055.3015.2019.2043.4040.808.80
Musk36.2038.7064.8029.7034.4072.3073.10108.00
Clean13.4054.10103.2038.0038.6067.8070.0034.50
Semeion116.60100.20163.8083.50109.70118.00116.8080.80
Madelon183.60161.50326.10123.40104.20247.30234.6061.80
Isolet194.10167.20412.10152.20138.70302.20289.80131.70
Lung138.298.3187.6387.830.9210.656.819.43
MLL73.4198.2137.8102.588.449.856.330.8
Ovarian98.270.3889.883.276.487.633.517.8
CNS93.274.8101.581.294.278.259.326.5
Mean Rnak4.044.375.593.113.225.674.152.85
Final Rank46723851
Table A7. Runtime on FS problem.
Table A7. Runtime on FS problem.
DatasetsCOAFATAHOAMCOAMSAACOSBOAPLOABCCOA
Aggregation9.626.017.018.459.788.2010.025.66
Banana11.098.849.9215.2118.2922.2621.928.12
Iris5.496.976.488.748.759.228.974.95
Bupa6.197.958.709.368.619.409.815.78
Glass5.978.6110.269.129.119.439.535.74
Breastcancer6.579.237.3610.309.4710.3010.636.94
Lipid6.889.4910.689.799.0210.1910.126.28
HeartEW8.589.457.069.839.889.759.726.46
Zoo8.659.269.829.639.499.489.336.41
Vote8.109.478.6010.1410.0710.209.886.28
Congress7.469.559.789.819.989.9410.066.03
Lymphography8.249.199.949.359.639.879.466.48
Vehicle10.6210.968.2311.5211.9011.7911.817.81
WDBC9.469.687.4010.3413.0410.0110.866.39
BreastEW10.159.439.4010.9910.8511.0110.937.06
SonarEW8.788.439.289.419.019.769.536.59
Libras9.718.729.4910.0010.178.9911.506.93
Hillvalley9.879.279.6910.5910.8111.5012.196.42
Musk11.6810.3410.3310.9011.2012.2012.966.55
Clean10.199.729.8711.1110.1711.9412.016.59
Semeion25.2421.9922.0925.5324.2027.3727.8718.77
Madelon42.4057.2063.4942.3043.0766.3264.9438.34
Isolet26.6134.4133.6431.5031.3744.5341.5326.89
Lung159.01116.91119.00105.60101.4090.3089.5079.29
MLL198.02210.98198.23141.45138.12158.10110.9890.18
Ovarian181.24178.91198.22138.91145.98156.32129.98101.30
CNS78.2786.4487.6893.21101.2093.0387.8869.00
Mean Rnak3.523.594.485.335.376.336.301.07
Final Rank23456871

References

  1. Wu, Q.; Gu, J. Design and research of robot visual servo system based on artificial intelligence. Agro Food Ind. Hi-Tech 2017, 28, 125–128. [Google Scholar]
  2. Chen, J.; Zhang, M.; Xu, B.; Sun, J.; Mujumdar, A.S. Artificial intelligence assisted technologies for controlling the drying of fruits and vegetables using physical fields: A review. Trends Food Sci. Technol. 2020, 105, 251–260. [Google Scholar] [CrossRef]
  3. Khulal, U.; Zhao, J.; Hu, W.; Chen, Q. Nondestructive quantifying total volatile basic nitrogen (TVB-N) content in chicken using hyperspectral imaging (HSI) technique combined with different data dimension reduction algorithms. Food Chem. 2016, 197, 1191–1199. [Google Scholar] [CrossRef] [PubMed]
  4. Li, H.; Geng, W.; Hassan, M.M.; Zuo, M.; Wei, W.; Wu, X.; Chen, Q. Rapid detection of chloramphenicol in food using SERS flexible sensor coupled artificial intelligent tools. Food Control 2021, 128, 108186. [Google Scholar] [CrossRef]
  5. El-Mesery, H.S.; Qenawy, M.; Ali, M.; Hu, Z.; Adelusi, O.A.; Njobeh, P.B. Artificial intelligence as a tool for predicting the quality attributes of garlic (Allium sativum L.) slices during continuous infrared-assisted hot air drying. J. Food Sci. 2024, 89, 7693–7712. [Google Scholar] [CrossRef] [PubMed]
  6. Wu, P.; Lei, X.; Zeng, J.; Qi, Y.; Yuan, Q.; Huang, W.; Lyu, X. Research progress in mechanized and intelligentized pollination technologies for fruit and vegetable crops. Int. J. Agric. Biol. Eng. 2024, 17, 11–21. [Google Scholar] [CrossRef]
  7. Chen, Q.; Hu, W.; Su, J.; Li, H.; Ouyang, Q.; Zhao, J. Nondestructively sensing of total viable count (TVC) in chicken using an artificial olfaction system based colorimetric sensor array. J. Food Eng. 2016, 168, 259–266. [Google Scholar] [CrossRef]
  8. Li, H.; Kutsanedzie, F.; Zhao, J.; Chen, Q. Quantifying total viable count in pork meat using combined hyperspectral imaging and artificial olfaction techniques. Food Anal. Methods 2016, 9, 3015–3024. [Google Scholar] [CrossRef]
  9. Li, H.; Zhang, B.; Hu, W.; Liu, Y.; Dong, C.; Chen, Q. Monitoring black tea fermentation using a colorimetric sensor array-based artificial olfaction system. J. Food Process. Preserv. 2018, 42, e13348. [Google Scholar] [CrossRef]
  10. Yan, H.; Zhang, C.; Gerrits, M.C.; Acquah, S.J.; Zhang, H.; Wu, H.; Fu, H. Parametrization of aerodynamic and canopy resistances for modeling evapotranspiration of greenhouse cucumber. Agric. For. Meteorol. 2018, 262, 370–378. [Google Scholar] [CrossRef]
  11. Wang, X.; Cai, H.; Zheng, Z.; Yu, L.; Wang, Z.; Li, L. Modelling root water uptake under deficit irrigation and rewetting in Northwest China. Agron. J. 2020, 112, 158–174. [Google Scholar] [CrossRef]
  12. Li, W.; Zhang, C.; Ma, T.; Li, W. Estimation of summer maize biomass based on a crop growth model. Emir. J. Food Agric. (EJFA) 2021, 33, 742–750. [Google Scholar] [CrossRef]
  13. Wang, L.; Lin, M.; Han, Z.; Han, L.; He, L.; Sun, W. Simulating the effects of drought stress timing and the amount irrigation on cotton yield using the CSM-CROPGRO-cotton model. Agronomy 2023, 14, 14. [Google Scholar] [CrossRef]
  14. Ni, J.; Xue, Y.; Zhou, Y.; Miao, M. Rapid identification of greenhouse tomato senescent leaves based on the sucrose-spectral quantitative prediction model. Biosyst. Eng. 2024, 238, 200–211. [Google Scholar] [CrossRef]
  15. Chauhdary, J.N.; Li, H.; Pan, X.; Zaman, M.; Anjum, S.A.; Yang, F.; Azamat, U. Modeling effects of climate change on crop phenology and yield of wheat–maize cropping system and exploring sustainable solutions. J. Sci. Food Agric. 2025, 105, 3679–3700. [Google Scholar] [CrossRef]
  16. Xiao-wei, H.; Zhi-Hua, L.; Xiao-bo, Z.; Ji-yong, S.; Han-ping, M.; Jie-wen, Z.; Holmes, M. Detection of meat-borne trimethylamine based on nanoporous colorimetric sensor arrays. Food Chem. 2016, 197, 930–936. [Google Scholar] [CrossRef]
  17. Liang, Z.; Li, Y.; Xu, L.; Zhao, Z. Sensor for monitoring rice grain sieve losses in combine harvesters. Biosyst. Eng. 2016, 147, 51–66. [Google Scholar] [CrossRef]
  18. Zhou, C.; Yu, X.; Qin, X.; Ma, H.; Yagoub, A.E.A.; Hu, J. Hydrolysis of rapeseed meal protein under simulated duodenum digestion: Kinetic modeling and antioxidant activity. LWT-Food Sci. Technol. 2016, 68, 523–531. [Google Scholar] [CrossRef]
  19. Liu, J.; Liu, X.; Zhu, X.; Yuan, S. Droplet characterisation of a complete fluidic sprinkler with different nozzle dimensions. Biosyst. Eng. 2016, 148, 90–100. [Google Scholar] [CrossRef]
  20. Yuexia, C.; Long, C.; Ruochen, W.; Xing, X.; Yujie, S.; Yanling, L. Modeling and test on height adjustment system of electrically-controlled air suspension for agricultural vehicles. Int. J. Agric. Biol. Eng. 2016, 9, 40–47. [Google Scholar]
  21. Darko, R.O.; Shouqi, Y.; Haofang, Y.; Junping, L.; Abbey, A. Calibration and validation of Aqua Crop for deficit and full irrigation of tomato. Int. J. Agric. Biol. Eng. 2016, 9, 104–110. [Google Scholar]
  22. Wu, X.; Wu, B.; Sun, J.; Yang, N. Classification of apple varieties using near infrared reflectance spectroscopy and fuzzy discriminant c-means clustering model. J. Food Process Eng. 2017, 40, e12355. [Google Scholar] [CrossRef]
  23. Lu, X.; Sun, J.; Mao, H.; Wu, X.; Gao, H. Quantitative determination of rice starch based on hyperspectral imaging technology. Int. J. Food Prop. 2017, 20, S1037–S1044. [Google Scholar] [CrossRef]
  24. Yan, H.; Zhang, C.; Oue, H.; Peng, G.J.; Darko, R.O. Determination of crop and soil evaporation coefficients for estimating evapotranspiration in a paddy field. Int. J. Agric. Biol. Eng. 2017, 10, 130–139. [Google Scholar]
  25. Gao, J.; Qi, H. Soil throwing experiments for reverse rotary tillage at various depths, travel speeds, and rotational speeds. Trans. ASABE 2017, 60, 1113–1121. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Sun, J.; Li, J.; Wu, X.; Dai, C. Quantitative analysis of cadmium content in tomato leaves based on hyperspectral image and feature selection. Appl. Eng. Agric. 2018, 34, 789–798. [Google Scholar] [CrossRef]
  27. Liu, C.; Yu, H.; Liu, Y.; Zhang, L.; Li, D.; Zhang, J.; Sui, Y. Prediction of Anthocyanin Content in Purple-Leaf Lettuce Based on Spectral Features and Optimized Extreme Learning Machine Algorithm. Agronomy 2024, 14, 2915. [Google Scholar] [CrossRef]
  28. Duan, N.; Gong, W.; Wu, S.; Wang, Z. Selection and application of ssDNA aptamers against clenbuterol hydrochloride based on ssDNA library immobilized SELEX. J. Agric. Food Chem. 2017, 65, 1771–1777. [Google Scholar] [CrossRef]
  29. Shi, J.; Zhang, F.; Wu, S.; Guo, Z.; Huang, X.; Hu, X.; Zou, X. Noise-free microbial colony counting method based on hyperspectral features of agar plates. Food Chem. 2019, 274, 925–932. [Google Scholar] [CrossRef]
  30. Yao, K.; Sun, J.; Zhou, X.; Nirere, A.; Tian, Y.; Wu, X. Nondestructive detection for egg freshness grade based on hyperspectral imaging technology. J. Food Process Eng. 2020, 43, e13422. [Google Scholar] [CrossRef]
  31. Jiang, H.; He, Y.; Chen, Q. Determination of acid value during edible oil storage using a portable NIR spectroscopy system combined with variable selection algorithms based on an MPA-based strategy. J. Sci. Food Agric. 2021, 101, 3328–3335. [Google Scholar] [CrossRef] [PubMed]
  32. Xu, M.; Sun, J.; Yao, K.; Wu, X.; Shen, J.; Cao, Y.; Zhou, X. Nondestructive detection of total soluble solids in grapes using VMD-RC and hyperspectral imaging. J. Food Sci. 2022, 87, 326–338. [Google Scholar] [CrossRef] [PubMed]
  33. Liu, X.; Sun, Z.; Zuo, M.; Zou, X.; Wang, T.; Li, J. Quantitative detection of restructured steak adulteration based on hyperspectral technology combined with a wavelength selection algorithm cascade strategy. Food Sci. Technol. Res. 2021, 27, 859–869. [Google Scholar] [CrossRef]
  34. Wang, J.; Guo, Z.; Zou, C.; Jiang, S.; El-Seedi, H.R.; Zou, X. General model of multi-quality detection for apple from different origins by Vis/NIR transmittance spectroscopy. J. Food Meas. Charact. 2022, 16, 2582–2595. [Google Scholar] [CrossRef]
  35. Wang, C.; Luo, X.; Guo, Z.; Wang, A.; Zhou, R.; Cai, J. Influence of the peel on online detecting soluble solids content of pomelo using Vis-NIR spectroscopy coupled with chemometric analysis. Food Control 2025, 167, 110777. [Google Scholar] [CrossRef]
  36. Zhang, B.; Li, H.; Pan, W.; Chen, Q.; Ouyang, Q.; Zhao, J. Dual-color upconversion nanoparticles (UCNPs)-based fluorescent immunoassay probes for sensitive sensing foodborne pathogens. Food Anal. Methods 2017, 10, 2036–2045. [Google Scholar] [CrossRef]
  37. Mao, H.; Hang, T.; Zhang, X.; Lu, N. Both multi-segment light intensity and extended photoperiod lighting strategies, with the same daily light integral, promoted Lactuca sativa L. growth and photosynthesis. Agronomy 2019, 9, 857. [Google Scholar] [CrossRef]
  38. Wang, P.; Li, H.; Hassan, M.M.; Guo, Z.; Zhang, Z.Z.; Chen, Q. Fabricating an acetylcholinesterase modulated UCNPs-Cu2+ fluorescence biosensor for ultrasensitive detection of organophosphorus pesticides-diazinon in food. J. Agric. Food Chem. 2019, 67, 4071–4079. [Google Scholar] [CrossRef]
  39. Zhang, Z.; Yang, M.; Pan, Q.; Jin, X.; Wang, G.; Zhao, Y.; Hu, Y. Identification of tea plant cultivars based on canopy images using deep learning methods. Sci. Hortic. 2025, 339, 113908. [Google Scholar] [CrossRef]
  40. Chen, X.; Hou, T.; Liu, S.; Guo, Y.; Hu, J.; Xu, G.; Liu, W. Design of a Micro-Plant Factory Using a Validated CFD Model. Agriculture 2024, 14, 2227. [Google Scholar] [CrossRef]
  41. Niu, Z.; Huang, T.; Xu, C.; Sun, X.; Taha, M.F.; He, Y.; Qiu, Z. A Novel Approach to Optimize Key Limitations of Azure Kinect DK for Efficient and Precise Leaf Area Measurement. Agriculture 2025, 15, 173. [Google Scholar] [CrossRef]
  42. Wang, X.; Jiang, H.; Mu, M.; Dong, Y. A dynamic collaborative adversarial domain adaptation network for unsupervised rotating machinery fault diagnosis. Reliab. Eng. Syst. Saf. 2024, 255, 110662. [Google Scholar] [CrossRef]
  43. Yan, J.; Cheng, Y.; Zhang, F.; Zhou, N.; Wang, H.; Jin, B.; Wang, M.; Zhang, W. Multi-Modal Imitation Learning for Arc Detection in Complex Railway Environments. IEEE Trans. Instrum. Meas. 2025, 74, 3556896. [Google Scholar] [CrossRef]
  44. Han, D.; Qi, H.; Wang, S.; Hou, D.; Wang, C. Adaptive stepsize forward–backward pursuit and acoustic emission-based health state assessment of high-speed train bearings. Struct. Health Monit. 2024. [Google Scholar] [CrossRef]
  45. Yang, N.; Qian, Y.; EL-Mesery, H.S.; Zhang, R.; Wang, A.; Tang, J. Rapid detection of rice disease using microscopy image identification based on the synergistic judgment of texture and shape features and decision tree–confusion matrix method. J. Sci. Food Agric. 2019, 99, 6589–6600. [Google Scholar] [CrossRef]
  46. Shi, J.; Hu, X.; Zou, X.; Zhao, J.; Zhang, W.; Holmes, M.; Zhang, X. A rapid and nondestructive method to determine the distribution map of protein, carbohydrate and sialic acid on Edible bird’s nest by hyper-spectral imaging and chemometrics. Food Chem. 2017, 229, 235–241. [Google Scholar] [CrossRef] [PubMed]
  47. Cheng, J.; Sun, J.; Yao, K.; Dai, C. Generalized and hetero two-dimensional correlation analysis of hyperspectral imaging combined with three-dimensional convolutional neural network for evaluating lipid oxidation in pork. Food Control 2023, 153, 109940. [Google Scholar] [CrossRef]
  48. Tchabo, W.; Ma, Y.; Kwaw, E.; Zhang, H.; Xiao, L.; Apaliya, M.T. Statistical interpretation of chromatic indicators in correlation to phytochemical profile of a sulfur dioxide-free mulberry (Morus nigra) wine submitted to non-thermal maturation processes. Food Chem. 2018, 239, 470–477. [Google Scholar] [CrossRef]
  49. Sun, L.; Feng, S.; Chen, C.; Liu, X.; Cai, J. Identification of eggshell crack for hen egg and duck egg using correlation analysis based on acoustic resonance method. J. Food Process Eng. 2020, 43, e13430. [Google Scholar] [CrossRef]
  50. Khan, I.A.; Luo, J.; Shi, H.; Zou, Y.; Khan, A.; Zhu, Z.; Huang, M. Mitigation of heterocyclic amines by phenolic compounds in allspice and perilla frutescens seed extract: The correlation between antioxidant capacities and mitigating activities. Food Chem. 2022, 368, 130845. [Google Scholar] [CrossRef]
  51. Li, Y.; Ruan, S.; Zhou, A.; Xie, P.; Azam, S.R.; Ma, H. Ultrasonic modification on fermentation characteristics of Bacillus varieties: Impact on protease activity, peptide content and its correlation coefficient. LWT 2022, 154, 112852. [Google Scholar] [CrossRef]
  52. Sun, J.; Jiang, S.; Mao, H.; Wu, X.; Li, Q. Classification of black beans using visible and near infrared hyperspectral imaging. Int. J. Food Prop. 2016, 19, 1687–1695. [Google Scholar] [CrossRef]
  53. Liu, J.; Abbas, I.; Noor, R.S. Development of deep learning-based variable rate agrochemical spraying system for targeted weeds control in strawberry crop. Agronomy 2021, 11, 1480. [Google Scholar] [CrossRef]
  54. Zareef, M.; Chen, Q.; Hassan, M.M.; Arslan, M.; Hashim, M.M.; Ahmad, W.; Agyekum, A.A. An overview on the applications of typical non-linear algorithms coupled with NIR spectroscopy in food analysis. Food Eng. Rev. 2020, 12, 173–190. [Google Scholar] [CrossRef]
  55. Ma, L.; Yang, X.; Xue, S.; Zhou, R.; Wang, C.; Guo, Z.; Cai, J. “Raman plus X” dual-modal spectroscopy technology for food analysis: A review. Compr. Rev. Food Sci. Food Saf. 2025, 24, e70102. [Google Scholar] [CrossRef]
  56. Askr, H.; Abdel-Salam, M.; Hassanien, A.E. Copula entropy-based golden jackal optimization algorithm for high-dimensional feature selection problems. Expert Syst. Appl. 2024, 238, 121582. [Google Scholar] [CrossRef]
  57. Abdel-Salam, M.; Hu, G.; Çelik, E.; Gharehchopogh, F.S.; El-Hasnony, I.M. Chaotic RIME optimization algorithm with adaptive mutualism for feature selection problems. Comput. Biol. Med. 2024, 179, 108803. [Google Scholar] [CrossRef]
  58. Tijjani, S.; Ab Wahab, M.N.; Noor, M.H.M. An enhanced particle swarm optimization with position update for optimal feature selection. Expert. Syst. Appl. 2024, 247, 123337. [Google Scholar] [CrossRef]
  59. Fang, Y.; Yao, Y.; Lin, X.; Wang, J.; Zhai, H. A feature selection based on genetic algorithm for intrusion detection of industrial control systems. Comput. Secur. 2024, 139, 103675. [Google Scholar] [CrossRef]
  60. Ye, Z.; Luo, J.; Zhou, W.; Wang, M.; He, Q. An ensemble framework with improved hybrid breeding optimization-based feature selection for intrusion detection. Future Gener. Comput. Syst. 2024, 151, 124–136. [Google Scholar] [CrossRef]
  61. Wang, Y.; Ran, S.; Wang, G.G. Role-oriented binary grey wolf optimizer using foraging-following and Lévy flight for feature selection. Appl. Math. Model. 2024, 126, 310–326. [Google Scholar] [CrossRef]
  62. Houssein, E.H.; Oliva, D.; Celik, E.; Emam, M.M.; Ghoniem, R.M. Boosted sooty tern optimization algorithm for global optimization and feature selection. Expert Syst. Appl. 2023, 213, 119015. [Google Scholar] [CrossRef]
  63. Dehghani, M.; Montazeri, Z.; Trojovská, E.; Trojovský, P. Coati Optimization Algorithm: A new bio-inspired metaheuristic algorithm for solving optimization problems. Knowl.-Based Syst. 2023, 259, 110011. [Google Scholar] [CrossRef]
  64. Zhang, Q.; Gao, H.; Zhan, Z.H.; Li, J.; Zhang, H. Growth Optimizer: A powerful metaheuristic algorithm for solving continuous and discrete global optimization problems. Knowl.-Based Syst. 2023, 261, 110206. [Google Scholar] [CrossRef]
  65. El-Kenawy, E.S.M.; Khodadadi, N.; Mirjalili, S.; Abdelhamid, A.A.; Eid, M.M.; Ibrahim, A. Greylag goose optimization: Nature-inspired optimization algorithm. Expert Syst. Appl. 2024, 238, 122147. [Google Scholar] [CrossRef]
  66. Lian, J.; Hui, G. Human evolutionary optimization algorithm. Expert Syst. Appl. 2024, 241, 122638. [Google Scholar] [CrossRef]
  67. SeyedGarmroudi, S.; Kayakutlu, G.; Kayalica, M.O.; Çolak, Ü. Improved Pelican optimization algorithm for solving load dispatch problems. Energy 2024, 289, 129811. [Google Scholar] [CrossRef]
  68. Taheri, A.; RahimiZadeh, K.; Beheshti, A.; Baumbach, J.; Rao, R.V.; Mirjalili, S.; Gandomi, A.H. Partial reinforcement optimizer: An evolutionary optimization algorithm. Expert Syst. Appl. 2024, 238, 122070. [Google Scholar] [CrossRef]
  69. Gao, H.; Zhang, Q.; Bu, X.; Zhang, H. Quadruple parameter adaptation growth optimizer with integrated distribution, confrontation, and balance features for optimization. Expert Syst. Appl. 2024, 235, 121218. [Google Scholar] [CrossRef]
  70. Zhu, F.; Li, G.; Tang, H.; Li, Y.; Lv, X.; Wang, X. Dung beetle optimization algorithm based on quantum computing and multi-strategy fusion for solving engineering problems. Expert Syst. Appl. 2024, 236, 121219. [Google Scholar] [CrossRef]
  71. Dhiman, G.; Kaur, A. STOA: A bio-inspired based optimization algorithm for industrial engineering problems. Eng. Appl. Artif. Intell. 2019, 82, 148–174. [Google Scholar] [CrossRef]
  72. Sallam, K.M.; Elsayed, S.M.; Chakrabortty, R.K.; Ryan, M.J. Improved multi-operator differential evolution algorithm for solving unconstrained problems. In Proceedings of the 2020 IEEE congress on evolutionary computation (CEC), Glasgow, UK, 19–24 July 2020; IEEE: Piscataway, NJ, USA; pp. 1–8. [Google Scholar]
  73. Tian, Z.; Gai, M. Football team training algorithm: A novel sport-inspired meta-heuristic optimization algorithm for global optimization. Expert Syst. Appl. 2024, 245, 123088. [Google Scholar] [CrossRef]
  74. Oladejo, S.O.; Ekwe, S.O.; Mirjalili, S. The Hiking Optimization Algorithm: A novel human-based metaheuristic approach. Knowl.-Based Syst. 2024, 296, 111880. [Google Scholar] [CrossRef]
  75. Jia, H.; Zhou, X.; Zhang, J.; Abualigah, L.; Yildiz, A.R.; Hussien, A.G. Modified crayfish optimization algorithm for solving multiple engineering application problems. Artif. Intell. Rev. 2024, 57, 127. [Google Scholar] [CrossRef]
  76. Cui, J.; Wu, L.; Huang, X.; Xu, D.; Liu, C.; Xiao, W. Multi-strategy adaptable ant colony optimization algorithm and its application in robot path planning. Knowl.-Based Syst. 2024, 288, 111459. [Google Scholar] [CrossRef]
  77. Fu, Y.; Liu, D.; Chen, J.; He, L. Secretary bird optimization algorithm: A new metaheuristic for solving global optimization problems. Artif. Intell. Rev. 2024, 57, 123. [Google Scholar] [CrossRef]
  78. Yuan, C.; Zhao, D.; Heidari, A.A.; Liu, L.; Chen, Y.; Chen, H. Polar lights optimizer: Algorithm and applications in image segmentation and feature selection. Neurocomputing 2024, 607, 128427. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of attack phase. (a). Launch an attack on the tree. (b). Launch an attack on the ground.
Figure 1. Schematic diagram of attack phase. (a). Launch an attack on the tree. (b). Launch an attack on the ground.
Biomimetics 10 00380 g001
Figure 2. Principle diagram of escape phase.
Figure 2. Principle diagram of escape phase.
Biomimetics 10 00380 g002
Figure 3. Algorithm execution flowchart: (a) COA algorithm; (b) ABCCOA algorithm.
Figure 3. Algorithm execution flowchart: (a) COA algorithm; (b) ABCCOA algorithm.
Biomimetics 10 00380 g003
Figure 4. Schematic of adaptive search strategy.
Figure 4. Schematic of adaptive search strategy.
Biomimetics 10 00380 g004
Figure 5. Schematic of balancing factor.
Figure 5. Schematic of balancing factor.
Biomimetics 10 00380 g005
Figure 6. Schematic of centroid guidance strategy.
Figure 6. Schematic of centroid guidance strategy.
Biomimetics 10 00380 g006
Figure 7. Average ranking of algorithms under different population sizes.
Figure 7. Average ranking of algorithms under different population sizes.
Biomimetics 10 00380 g007
Figure 8. Average ranking of algorithms under different strategies.
Figure 8. Average ranking of algorithms under different strategies.
Biomimetics 10 00380 g008
Figure 9. Comparison chart of population diversity.
Figure 9. Comparison chart of population diversity.
Biomimetics 10 00380 g009
Figure 10. Exploration/exploitation ratio analysis plot.
Figure 10. Exploration/exploitation ratio analysis plot.
Biomimetics 10 00380 g010
Figure 11. The boxplot on the CEC2020 test function.
Figure 11. The boxplot on the CEC2020 test function.
Biomimetics 10 00380 g011
Figure 12. The mean rank on the CEC2020 test function.
Figure 12. The mean rank on the CEC2020 test function.
Biomimetics 10 00380 g012
Figure 13. Friedman nonparametric test results on CEC2020 test function.
Figure 13. Friedman nonparametric test results on CEC2020 test function.
Biomimetics 10 00380 g013
Figure 14. Convergence graph on CEC2020 test function.
Figure 14. Convergence graph on CEC2020 test function.
Biomimetics 10 00380 g014
Figure 15. Calculation process of fitness function value for FS problem.
Figure 15. Calculation process of fitness function value for FS problem.
Biomimetics 10 00380 g015
Figure 16. Ranking of fitness values on FS problems.
Figure 16. Ranking of fitness values on FS problems.
Biomimetics 10 00380 g016
Figure 17. The average ranking of classification accuracy on FS problems.
Figure 17. The average ranking of classification accuracy on FS problems.
Biomimetics 10 00380 g017
Figure 18. Average ranking of feature subset size on FS problem.
Figure 18. Average ranking of feature subset size on FS problem.
Biomimetics 10 00380 g018
Figure 19. Average ranking of runtime on FS problem.
Figure 19. Average ranking of runtime on FS problem.
Biomimetics 10 00380 g019
Figure 20. Comprehensive indicator analysis on FS problem.
Figure 20. Comprehensive indicator analysis on FS problem.
Biomimetics 10 00380 g020
Table 1. CEC2020 test function-specific information.
Table 1. CEC2020 test function-specific information.
FunctionsTypeNameValues
CEC2020_F1UnimodalShifted and Rotated Bent Cigar Function100
CEC2020_F2Multi-modalShifted and Rotated Schwefel’s Function1100
CEC2020_F3 Shifted and Rotated unacek bi-Rastrigin Function700
CEC2020_F4 Expanded Rosenbrock’s plus Griewangk’s Function1900
CEC2020_F5HybridHybrid Function 1 (N = 3)1700
CEC2020_F6 Hybrid Function 2 (N = 4)1600
CEC2020_F7 Hybrid Function 3 (N = 5)2100
CEC2020_F8CompositionComposition Function 1 (N = 3)2200
CEC2020_F9 Composition Function 2 (N = 4)2400
CEC2020_F10 Composition Function 3 (N = 5)2500
Table 2. Comparison of algorithm parameter settings on cec2020 test functions.
Table 2. Comparison of algorithm parameter settings on cec2020 test functions.
AlgorithmsTimeParameters Setting
COA [63]2023No Parameters
GGO [65]2024 α = 0 . 99 ,   β = 0 . 01
HEOA [66]2024 α = 4 ,   γ = 1 . 5 ,   ω = 0.2 cos π 2 1 t M a x i t e r
IPOA [67]2024 I { 1 , 2 } ,   F P K = 0.4 ,   F G K = 0.6
PRO [68]2024 τ = F E s M a x F E s ,   S R = e ( 1 τ )
QAGO [69]2024 P 1 = r a n d ( 0.05 , 0.2 ) · N ,   P 2 = G a u s s i a n ( 0.001 , 0.001 ) ,   P 2 = G a u s s i a n ( r a n d ( 0 , 0.3 ) , 0.01 )
QHDBO [70]2024 R = ( cos ( π · ( t / T m a x ) ) + 1 ) · 0.5
STOA [71]2019 C f = 2 ,   R r n d [ 0 , 1 ] ,   u = 1 ,   υ = 1
IMODE [72]2020 D = 2 ,   a r c h _ r a t e = 2.6
Table 3. Detailed information on 27 FS datasets.
Table 3. Detailed information on 27 FS datasets.
TypeNameFeature NumberClassification NumberInstance Size
LowAggregation27788
Banana225300
Iris43150
Bupa62345
Glass97214
Breastcancer92699
Lipid102583
HeartEW132270
MediumZoo167101
Vote162435
Congress162435
Lymphography184148
Vehicle184846
WDBC302569
BreastEW302569
SonarEW602208
HighLibras9015360
Hillvalley1002606
Musk1662476
Clean1672476
Semeion256101593
Madelon50022600
Isolet617261559
Lung12,5335203
MLL12,582372
Ovarian151542253
CNS7129260
Table 4. Comparison of algorithm parameter setting information on FS problems.
Table 4. Comparison of algorithm parameter setting information on FS problems.
AlgorithmsTimeParameters Setting
COA [63]2023No Parameters
FATA [73]2024No Parameters
HOA [74]2024No Parameters
MCOA [75]2024 C 2 = 2 ( F E s / M a x F E s ) ,   C 3 = 3 ,   k = ( 1 + ( F E s / M a x F E s ) 0.5 ) 10
MSAACO [76]2024 α = 1 ,   β = 7
SBOA [77]2024 R L = 0.5 · L e v y ( D i m ) ,   K { 1 , 2 }
PLO [78]2024 W 1 = 2 ( 1 + e 2 ( t / T ) 4 ) 1 ,   W 2 = e ( 2 t / T ) 3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, Q.; Yuan, S.; Fang, Y. Three Strategies Enhance the Bionic Coati Optimization Algorithm for Global Optimization and Feature Selection Problems. Biomimetics 2025, 10, 380. https://doi.org/10.3390/biomimetics10060380

AMA Style

Cao Q, Yuan S, Fang Y. Three Strategies Enhance the Bionic Coati Optimization Algorithm for Global Optimization and Feature Selection Problems. Biomimetics. 2025; 10(6):380. https://doi.org/10.3390/biomimetics10060380

Chicago/Turabian Style

Cao, Qingzheng, Shuqi Yuan, and Yi Fang. 2025. "Three Strategies Enhance the Bionic Coati Optimization Algorithm for Global Optimization and Feature Selection Problems" Biomimetics 10, no. 6: 380. https://doi.org/10.3390/biomimetics10060380

APA Style

Cao, Q., Yuan, S., & Fang, Y. (2025). Three Strategies Enhance the Bionic Coati Optimization Algorithm for Global Optimization and Feature Selection Problems. Biomimetics, 10(6), 380. https://doi.org/10.3390/biomimetics10060380

Article Metrics

Back to TopTop