Next Article in Journal
A Statistical Analysis of the Effect of Fin Design Factors on the Cooling Performance and System Mass of PCM–Fin Structured BTMS for LIB Cell
Previous Article in Journal
Computational Method for Dynamic Analysis of Multibody Systems with Deformable Elements
Previous Article in Special Issue
Magnetic Coil’s Performance Optimization with Nonsmooth Search Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adapted Multi-Strategy Fractional-Order Relative Pufferfish Optimization Algorithm for Feature Selection

School of Electronic and Knowledge Engineering, Tongji University, Shanghai 201804, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(17), 2799; https://doi.org/10.3390/math13172799
Submission received: 9 July 2025 / Revised: 26 August 2025 / Accepted: 28 August 2025 / Published: 31 August 2025
(This article belongs to the Special Issue Advances in Metaheuristic Optimization Algorithms)

Abstract

In the development of artificial intelligence (AI) technology, utilizing datasets for model instruction to achieve higher predictive and reasoning efficacy has become a common technical approach. However, primordial datasets often contain a significant number of redundant features (RF), which can compromise the prediction accuracy and generalization ability of models. To effectively reduce RF in datasets, this work advances a new version of the Pufferfish Optimization Algorithm (POA), termed AMFPOA. Firstly, by considering the knowledge disparities among different groups of members and incorporating the concept of adaptive learning, an adaptive exploration strategy is introduced to enhance the algorithm’s Global Exploration (GE) capability. Secondly, by dividing the entire swarm into multiple subswarms, a three-swarm search strategy is advanced. This allows for targeted optimization schemes for different subswarms, effectively achieving a good balance across various metrics for the algorithm. Lastly, leveraging the historical memory property of Fractional-Order theory and the member weighting of Bernstein polynomials, a Fractional-Order Bernstein exploitation strategy is advanced, which significantly augments the algorithm’s local exploitation (LE) capability. Subsequent experimental results on 23 real-world Feature Selection (FS) problems demonstrate that AMFPOA achieves an average success rate exceeding 87.5% in fitness function value (FFV), along with ideal efficacy rates of 86.5% in Classification Accuracy (CA) and 60.1% in feature subset size reduction. These results highlight its strong capability for RF elimination, establishing AMFPOA as a promising FS method.

1. Introduction

Advancements in AI technology have significantly propelled progress in cutting-edge fields such as education, economy, and healthcare [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18]. Within AI technology, large-scale model instruction has gradually become the focal point of researchers’ attention [19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34]. Utilizing massive amounts of data for model instruction to achieve higher predictive and reasoning efficacy has emerged as an effective method for efficacy enhancement [35,36,37,38,39,40,41,42,43,44,45,46,47,48]. Unfortunately, however, the data used for model instruction often contains a substantial amount of redundant knowledge, which compromises the interpretability and generalization ability of the models [49]. To improve the effectiveness of model instruction, it is often necessary to eliminate RF from the primordial dataset to enable accurate and lightweight model instruction, a process known as feature dimension reduction [50]. Instruction models using datasets that have undergone feature dimension reduction help free up a significant amount of computational resources and augment the reliability of AI technology [51]. Currently, employing metaheuristic-based FS methods to clean RF from the primordial dataset has become an effective technical approach [52]. Nevertheless, as the complexity of datasets continues to increase, these methods exhibit certain limitations. Therefore, to better perform FS on the primordial dataset, this paper aims to advance an efficient metaheuristic algorithm to achieve superior FS efficacy, thereby improving the usability of datasets and the reliability of models.
Metaheuristic algorithms are primarily lightweight computational methods formulated by simulating the objective behaviors of animals in nature. Currently, they are mainly categorized into four types: swarm-based, human-based, physics-based, and evolutionary-based [53]. Among them, typical representatives of swarm-based optimization algorithms include FFO [54], GWO [55], and AVOA [56]. For human-based optimization algorithms, notable examples are WSO [57], MOA [58], and TOA [59]. In the realm of physics-based optimization algorithms, typical ones are EO [60], AOA [61], and BHA [62]. Regarding evolutionary-based optimization algorithms, classic examples include GA [63], ES [64], and DE [65]. Thanks to their efficient search capabilities and relatively low computational costs, researchers have widely applied these algorithms to FS in datasets, aiming to achieve higher CA using fewer features.
In recent years, researchers have advanced numerous optimization algorithm-based FS methods to enhance the efficacy of RF elimination while improving the Classification Accuracy (CA) of feature subsets. For instance, Haouassi et al. introduced a novel binary grasshopper optimization FS algorithm and applied it to the feature subset selection problem. Experimental results demonstrated that their advanced algorithm attained a success rate of over 95% in terms of feature subset size across twenty datasets, outperforming five state-of-the-art comparative algorithms [66]. Mohan et al. advanced a binary teaching–learning-based optimization FS algorithm (FS-BTLBO) for FS. Experimental evidence showed that FS-BTLBO attained higher accuracy with the fewest features on the Wisconsin Diagnostic Breast Cancer (WDBC) dataset, effectively classifying malignant and benign tumors [67]. Mohammed et al. presented a monarch butterfly optimization (MBO) FS algorithm. Results from eighteen benchmark datasets indicated that, compared to four metaheuristic algorithms, MBO attained an average CA of up to 93% across all datasets while significantly reducing the number of selected features [68]. Hu et al. advanced an enhanced version of the black widow optimization FS algorithm (SDABWO) for FS. By introducing three learning strategies, the algorithm’s FS efficacy was enhanced. Results from twelve standard datasets from the UCI repository confirmed that SDABWO can simultaneously improve CA and reduce the dimensionality of the primordial dataset, making it a promising FS approach [69]. Ewees et al. advanced an improved sine–cosine FS algorithm (ISOA) that enhanced efficacy by incorporating Lévy flight and mutation operators. A comparative study on twenty benchmark datasets demonstrated that ISOA attained superior feature subset CA compared to FS methods constructed using other metaheuristic approaches [70]. Mohammed et al. introduced a binary version of the Horse herd Optimization FS Algorithm (HOA) to address the FS problem. They experimentally evaluated three transfer functions and three crossover operators to ensure the FS efficiency of their method. Experimental results on twenty-four real-world FS datasets indicated that it can effectively reduce RF in the primordial dataset, making it a promising FS approach [71]. Pan et al. advanced an improved grey wolf optimization FS algorithm and applied it to the FS problem in high-dimensional data. By amalgamating the ReliefF algorithm and Copula entropy, along with two novel search strategies, the algorithm’s FS efficacy was enhanced. Results on ten high-dimensional, small-sample gene explication datasets showed that the advanced algorithm selected fewer than 0.67% of the features while improving CA, demonstrating good efficacy and robustness in high-dimensional FS [72]. The shortcomings and advantages of related technologies are summarized as Table 1.
The aforementioned facts validate the effectiveness of optimization algorithm-based FS methods in feature dimension reduction and demonstrate that the incorporation of learning strategies can enhance FS efficacy of algorithms. However, despite the promising results attained by existing works in the FS domain, as the dimensionality of primordial datasets increases, current FS methods exhibit certain limitations. For instance, they often fail to comprehensively consider the relationships between RF elimination and the improvement of CA, leading to a compromise in the classification efficacy of datasets. Therefore, this paper aims to advance an FS method that can strike a good balance among multiple metrics to effectively enhance FS efficacy and eliminate RF. Fortunately, the Pufferfish Optimization Algorithm (POA) has been proven to be an optimization algorithm with efficient search capabilities [73], along with advantages such as a simple structure and good application scalability. Nevertheless, as the complexity of datasets increases, POA still faces several challenges, including the tendency to get trapped in locally non-ideal feature subsets and insufficient LE capabilities. To further improve the FS efficacy of the algorithm, this paper introduces an enhanced version of POA, termed AMFPOA, by amalgamating an adaptive exploration strategy, a three-swarm search strategy, and a Fractional-Order Bernstein exploitation strategy. Experimental results on 23 real-world FS problems indicate that AMFPOA exhibits a strong ability to balance the improvement of CA and the elimination of RF. Compared to eight efficient optimization algorithms, AMFPOA demonstrates superior FS efficacy and can be considered a promising FS approach. The main contributions of this paper are as follows:
  • An adaptive exploration strategy is advanced, which effectively augments the algorithm’s GE capability and improves the CA of feature subsets.
  • A three-swarm search strategy is introduced to ensure balance during the algorithm’s operation and enhance the metric trade-off when resolving FS problems.
  • A Fractional-Order Bernstein exploitation strategy is presented, which improves the algorithm’s LE capability when addressing FS problems, enabling effective elimination of RF.
  • By amalgamating the aforementioned three improvement strategies, an enhanced version of the POA, namely AMFPOA, is advanced, which boosts the algorithm’s FS efficacy.
  • Applying AMFPOA to resolve 23 real-world FS problems confirms that AMFPOA is a promising FS method.
The subsequent work plan of this paper is as follows: Section 2 introduces the mathematical model and execution logic of the POA. Section 3 advances an enhanced version of POA, termed AMFPOA, by amalgamating three improvement strategies. Section 4 applies AMFPOA to resolve 23 real-world FS problems, achieving excellent results in RF elimination and CA. Section 5 presents the research conclusions of this paper and outlines the future work plan. For the convenience of subsequent reading, we have summarized all the abbreviated words used in this paper in Table 2.

2. Mathematical Model of the POA

This section primarily introduces the mathematical model and implementation logic of the POA. POA is an optimization algorithm developed by simulating the predatory attack behaviors and defensive behaviors of pufferfish. In POA, the predatory attack on pufferfish mainly simulates the GE stage of the algorithm, while the defensive behaviors of pufferfish to evade predator attacks primarily simulate the LE stage. When utilizing the POA to resolve optimization problems, it is first necessary to initialize a set of initial solution candidates with search capabilities, a process known as the swarm initialization stage. Subsequently, the GE and LE stages of the POA are employed to iteratively refine the initialized swarm, thereby enhancing the quality of the candidate solutions. Below, a detailed description of the swarm initialization stage, GE stage, and LE stage of POA will be provided.

2.1. Swarm Initialization Stage

This section primarily focuses on the mathematical modeling of the swarm initialization stage in the POA. When employing an optimization algorithm to resolve a problem, it is initially necessary to produce an initialized swarm with a certain level of quality. Each member within the swarm represents a solution candidate for the problem to be optimized. The symbolic representation of a swarm X containing N members is shown in Equation (1).
X = X 1 X i X N N × 1 = x 1 , 1 x 1 , j x 1 , d i m x i , 1 x i , j x i , d i m x N , 1 x N , j x N , d i m N × d i m
where X represents the initialized swarm, X i signifies the knowledge of the i t h member, x i , j signifies the j t h dimensional knowledge of the i t h member, d i m indicates the dimensionality of each member, which corresponds to the number of variables in the problem to be optimized, and N represents the swarm size. x i , j is represented using Equation (2).
x i , j = l b j + r · ( u b j l b j )
where r signifies a random number within the interval [0, 1], while l b j and u b j represent the lower and upper boundary constraints, respectively, for the j t h variable of the problem to be optimized. Meanwhile, during the iterative process, the capability of an member to resolve the optimization problem is primarily evaluated using the FFV. The vector of FFV for members in the swarm is explicated as Equation (3).
F = F 1 F i F N N × 1 = F ( X 1 ) F ( X i ) F ( X N ) N × 1
where F represents the vector of FFV for members in the swarm, and F i signifies the FFV corresponding to the i t h member. For a minimization optimization problem, a smaller FFV indicates a higher quality of the member. After swarm initialization, the POA utilizes the GE stage and the LE stage to refine the quality of members. Below, the mathematical models of these two stages will be described in detail.

2.2. GE Stage

This section primarily introduces the mathematical model of the GE stage in the POA. This stage is mainly formulated by simulating the predator’s attack behavior on pufferfish. The core idea is that the predator initiates an attack on the pufferfish, moving in a direction guided by the pufferfish’s position, thereby enabling members within the swarm to “jump” across the global solution space. By abstractly representing the members in the swarm as the predator’s positions, the algorithm facilitates the exploration of the entire solution space by each member. The position update mechanism during the GE stage is explicated by Equation (4).
x i , j P 1 = x i , j + r i , j · ( S P i , j I i , j · x i , j )
where x i , j P 1 represents the updated knowledge of the j t h dimension for the i t h member after the GE stage. r i , j signifies a random number produced within the interval [0, 1]. I i , j is a randomly selected constant from the set {1, 2}. S P i , j indicates the pufferfish that the i t h member selects to attack from the set of pufferfish that are more likely to be attacked. This set is defined using Equation (5).
C P i = { X k : F k < F i   and   k i } ,   where   i = 1 , 2 , , N   and   k { 1 , 2 , , N }
where C P i signifies the set of candidate pufferfish that the i t h member may attack. X k represents an member whose FFV is superior to that of X i , and F k signifies the FFV of X k . Subsequently, the knowledge of the i t h member is preserved using Equation (6).
X i = X i P 1 , F i P 1 < F i X i , e l s e
where X i P 1 represents the updated state of the i t h member after undergoing the GE stage of the POA, and F i P 1 signifies the FFV corresponding to X i P 1 . Members are updated through the GE stage of the POA, which effectively ensures the algorithm’s GE capability and contributes to enhancing its optimization accuracy.

2.3. LE Stage

This section primarily introduces the mathematical model of the LE stage in the POA. This stage is mainly developed by simulating the defensive behavior of pufferfish when they are attacked by predators. The core idea is that, when under attack, a pufferfish inflates itself into a spiky ball. In response to this dangerous situation, the predator flees in the vicinity of the pufferfish, causing its position to be updated around the member, thereby realizing the LE stage of the POA. The position update mechanism during the LE stage is explicated by Equation (7).
x i , j P 2 = x i , j + ( 1 2 r i , j ) · u b j l b j t
where x i , j P 2 represents the updated knowledge of the j t h dimension for the i t h member after undergoing the LE stage. r i , j signifies a random number produced within the interval [0, 1], and t signifies the current iteration count. Subsequently, the knowledge for the i t h member is preserved using Equation (8).
X i = X i P 2 , F i P 2 < F i X i , e l s e
where X i represents the updated state of the i t h member after undergoing the LE stage of the POA, and F i P 2 signifies the FFV corresponding to X i P 2 . The updating of members through the LE stage of the POA effectively ensures the algorithm’s LE capability, thereby contributing to enhancing its optimization accuracy.

2.4. Implementation of the POA

The preceding sections introduced the mathematical models for the swarm initialization stage, GE stage, and LE stage of the POA. In this section, we will mainly elaborate on the execution logic of POA when resolving optimization problems. Specifically, the pseudocode for the execution of POA is presented in Algorithm 1.
Algorithm 1: Pseudo code for POA
Input: Parameters N , d i m , u b , l b , T .
Output: Best solution ( X b e s t ).
1. 
Initialize swarm base on Equation (1) and compute FFV base on Equation (3).
2. 
for  t = 1 : T
3.  
for  i = 1 : N
4.   
exploration stage
5.   
for  j = 1 : d i m
6.     
Compute the j t h dimensional of the i t h member base on Equation (4).
7.   
end for
8.   
Use Equation (6) to preserve member X i .
9.   
Exploitation Stage
10.    
for  j = 1 : d i m
11.      
Compute the j t h dimensional of the i t h member base on Equation (7).
12.    
end for
13.    
Use Equation (8) to preserve member X i .
14.   
end for
15.   
Save the best solution X b e s t .
16.  
end for
17.  
Output the best solution X b e s t .

3. Mathematical Model of the AMFPOA

The primordial POA exhibits deficiencies when addressing the FS problem, including insufficient GE capability, inadequate LE capability, and an imbalance between GE and LE. These shortcomings often lead the algorithm to become trapped in locally non-ideal feature subsets, resulting in reduced redundancy feature elimination and compromised CA of feature subsets. To mitigate these limitations, this section advances an enhanced version of POA, termed AMFPOA, by amalgamating three improvement strategies aimed at enhancing the algorithm’s FS efficacy and improving subset CA. In AMFPOA, firstly, to address the issue of inadequate GE capability in POA, an adaptive exploration strategy is introduced. The core idea behind this strategy is to enhance the algorithm’s GE by considering knowledge disparities among members from different groups and incorporating the concept of adaptive learning. This enables members to effectively explore the solution space, thereby improving the algorithm’s GE capability. Secondly, to tackle the imbalance between GE and LE stages in POA, a three-swarm search strategy is advanced. By considering the diverse characteristics of members, the entire swarm is divided into multiple subswarms. Tailored optimization strategies are then applied to each subswarm based on its specific characteristics, achieving a balance between the two stages of the algorithm. This augments the algorithm’s ability to escape from local optima. Lastly, to address the problem of insufficient LE capability in POA, a Fractional-Order Bernstein exploitation strategy is introduced. Leveraging the historical memory property of Fractional-Order theory and the weighted averaging capability of Bernstein polynomials, this strategy effectively guides members in their exploitation behavior, thereby strengthening the algorithm’s LE capability. As a result, the algorithm achieves higher CA when resolving the FS problem. The following sections will provide a detailed introduction to these three improvement strategies.

3.1. Adaptive Exploration Strategy

The primordial POA demonstrates inadequate GE capabilities when addressing complex FS problems, primarily due to the increase in the number of features in the primordial dataset. This deficiency leads to a reduction in Population Diversity (PD), which hampers the algorithm’s ability to locate regions that may contain the ideal combinations of FS subsets. Consequently, the CA of the subsets is diminished. To address this issue, there is an urgent need to advance a search strategy with efficient exploration capabilities. Zhang et al. [74] pointed out in their work that allowing members to learn from the gaps between different types of members within the swarm can effectively enhance the algorithm’s GE ability, and this viewpoint has been verified. Inspired by this, to strengthen the algorithm’s GE ability in resolving FS problems, this section incorporates this learning concept. At the same time, considering that enhancing the GE efficacy of the algorithm may reduce the convergence degree of the solution set, a learning factor is employed to control the process of gap-based learning. This ensures that while the GE efficacy is improved, the convergence speed is not significantly affected.
This section primarily focuses on introducing the core idea and mathematical model of the adaptive exploration strategy. The main concept behind the adaptive exploration strategy is that members learn from the disparities between themselves and others with different attributes, while also taking into account the members’ learning capabilities and the acceptability of these disparities, in order to enhance the algorithm’s GE capability. The schematic diagram of the adaptive exploration strategy is illustrated in Figure 1. Specifically, in the adaptive exploration strategy, four distinct types of disparities are considered: the disparity between the ideal member and a relatively superior member ( g a p 1 ), the disparity between the ideal member and a relatively inferior member ( g a p 2 ), the disparity between a relatively superior member and a relatively inferior member ( g a p 3 ), and the disparity between two different randomly selected members ( g a p 4 ). These disparities are represented using Equation (9).
g a p 1 = X b e s t X b e t t e r g a p 2 = X b e s t X w o r s e g a p 3 = X b e t t e r X w o r s e g a p 4 = X r a n d 1 X r a n d 2
where X b e s t signifies the ideal member within the swarm. X b e t t e r represents a relatively superior member in the swarm, defined as a randomly selected member from the set of the top five members with the smallest fitness values. Conversely, X w o r s e indicates a relatively inferior member in the swarm, defined as a randomly selected member from the set of the top five members with the largest fitness values. X r a n d 1 and X r a n d 2 denote two distinct random members within the swarm, respectively. Additionally, the acceptability levels of different disparities vary and are represented using Equation (10).
L F k = g a p k k = 1 4 g a p k ,   ( k = 1 , 2 , 3 , 4 )
where L F k represents the acceptability level of the member disparities within the k t h group. Additionally, it is taken into account that members with different characteristics possess varying capabilities for learning from disparities, and the learning capability of a member is explicated using Equation (11).
S F i = F i F m a x ,   ( 1 i N )
where S F i signifies the learning capability of the i t h member; the process of the i t h member learning from the disparities within the k t h group is represented using Equation (12).
K A k = a r c t a n ( 1 t / T ) · S F i · L F k · g a p k ,     ( k = 1 , 2 , 3 , 4 )
where K A k represents the amount of knowledge acquired by the i t h member after learning from the disparities within the k t h group. The a r c t a n ( · ) signifies the arctangent function operation; in the context of the arctangent function, as the variable t increases, the function value decreases non-linearly from π / 4 to 0. This design is primarily intended to simulate the situation where, in the later stages of iteration, the learning of knowledge volume should be gradually reduced. One cannot simply focus on enhancing the algorithm’s GE efficacy while neglecting the member-aggregation ability in the late iteration stage. This approach can effectively strengthen the convergence efficacy during the later stages of iteration. T indicates the Maximum Number (MN) of iterations. Subsequently, the new state produced by the i t h member after learning from the four groups of disparities is represented using Equation (13).
X i n e w = X i + K A 1 + K A 2 + K A 3 + K A 4
where X i n e w represents the new state of the i t h member after being updated through the adaptive exploration strategy. Subsequently, the knowledge of the i t h member is preserved using Equation (14).
X i = X i n e w i f   F i n e w < F i X i o t h e r w i s e
where F i n e w signifies the FFV corresponding to the member X i n e w . By adaptively learning from different knowledge disparities, the member effectively augments the algorithm’s GE capability, thereby improving its ability to eliminate RF.

3.2. Three-Swarm Search Strategy

As the dimensionality of the FS problem increases, the algorithm needs to search through an exponentially growing number of feature subset candidates when resolving the FS problem. This necessitates a well-balanced exploration and exploitation stage in the algorithm, enabling it to effectively locate local regions while further exploring these regions to enhance the CA of the feature subset. When the primordial POA is applied to resolve the FS problem, its CA is compromised due to an inadequate balance between the GE and LE stages. To alleviate this issue, there is an urgent need for us to advance a solution that can enhance the balance between the two stages. Fortunately, Shen et al. [75] introduced the concept of multi-swarm evolution in their work. By dividing the swarm into sub-swarms with distinct characteristics and then enabling each sub-swarm to undergo targeted updates, they effectively improved the balance between the two stages during the algorithm’s execution. Inspired by this idea, this section adopts the same concept of utilizing a three-swarm search strategy to improve the balance between the GE and LE stages. Specifically, by considering the FFV of members in the swarm, they are divided into three sub-swarms; the proportion for swarm division mainly adopts the concept of averaging. Specifically, the entire swarm is divided into three sub-swarms in a ratio of 3:4:3. This approach not only effectively preserves the members’ GE capabilities but also augments their LE abilities, thereby contributing to the balance during the algorithm’s execution process. The set of members ranking in the top 30% in terms of FFV is classified as the exploitation sub-swarm. LE operations are performed on this sub-swarm to enable better tuning of CA within local regions. Secondly, the set of members ranking in the bottom 30% in terms of FFV is classified as the exploration sub-swarm. GE operations are carried out on this sub-swarm to facilitate exploration across a wide range of solution spaces, thereby enhancing the algorithm’s GE capability. In addition, the remaining members are classified as the exploration–exploitation sub-swarm. Both GE and LE behaviors are simultaneously applied to this sub-swarm to achieve a better balance between the GE and LE stages of the algorithm, enhancing its ability to escape from local sub-ideal feature subset traps. This swarm division process is illustrated in Figure 2.
The exploration sub-swarm primarily performs GE operations and updates its positions using Equation (15).
X i n e w = X i + a r c t a n 1 t T · r a n d · ( X r a n d X i )
where r a n d signifies a random number produced within the interval [0, 1], and X r a n d represents a random member in the swarm that is different from X i . Meanwhile, Mirjalili et al. [76] pointed out that the idea of employing a cosine-function-based form for member spiral updating can effectively enhance the algorithm’s LE capability. Therefore, to strengthen the exploitation ability of the exploitation sub-swarm in our case, we adopt this cosine-spiral updating method. Additionally, we control the spiral factor by using the proportion of the iteration number. This ensures that while the spiral strategy augments the LE ability, it does not fall into the local optima. In summary, the exploitation sub-swarm updates its positions using Equation (16).
X i n e w = X b e s t + 1 t T · ( r a n d 1 ) · ( X r a n d X i ) i f   P 0.5 X b e s t + r a n d · cos 2 π · 1 1 + e t T · ( X b e s t X i ) o t h e r w i s e
where P represents a random number within the interval [0, 1]. Subsequently, the exploration–exploitation sub-swarm simultaneously leans towards both GE and LE operations and updates its positions using Equation (17).
X i n e w = E x p l o r a t i o n   p h a s e :   Execution   of   Equation   ( 15 ) r a n d 0.5 E x p l o i t a t i o n   p h a s e :   Execution   of   Equation   ( 16 ) r a n d > 0.5
After conducting multi-swarm search behaviors, the member knowledge is preserved using Equation (14). By updating members through the three-swarm search strategy, a good balance is attained between the GE and LE stages of the algorithm, thereby increasing the probability of the algorithm escaping from local sub-ideal feature subset traps.

3.3. Fractional-Order Bernstein Exploitation Strategy

The primordial POA suffers from insufficient LE capability when resolving the FS problem, resulting in a certain loss in the CA of the feature subsets. To address this, there is an urgent need for us to advance an effective LE strategy to enhance the convergence accuracy of the algorithm. Fortunately, Chen et al. [77] pointed out that applying fractional-order theory to weight historical members can yield members with better representativeness. That is, these weighted members can absorb knowledge from past members and improve their own quality; based on this inspiration, in this section, we partially utilize the former representative fractional-order weighted members for guidance to strengthen the algorithm’s LE capability. Since it has been verified in the literature [77] that using members from the previous three generations can achieve favorable results, we still adhere to this proven idea when weighting the fractional-order members. Moreover, Wang et al. [78] and Zhang et al. [79] indicated that using Bernstein polynomials to weight members with different characteristics can also enhance member quality and strengthen the algorithm’s exploitation efficacy. Therefore, we integrate the above-mentioned ideas based on fractional-order theory and Bernstein polynomials; this paper advances a Fractional-Order Bernstein exploitation strategy to enhance the algorithm’s LE efficacy. Compared to traditional development strategies such as Lévy flight and those based on chaos, this advanced strategy takes into account both the risk of getting trapped in local optima associated with Lévy flight and the irregularity introduced by chaos-based approaches. Consequently, it is better at enhancing the algorithm’s effective exploitation capability rather than promoting a blind or untargeted exploitation process. Specifically, this strategy simultaneously takes into account the historical learning property of the fractional-order learning strategy and the weighted nature of the Bernstein learning strategy, significantly improving the algorithm’s LE capability. The Fractional-Order Bernstein exploitation strategy mainly guides the swarm through fractional-order weighted members and Bernstein weighted members to enhance the CA of feature subsets. Among them, the fractional-order theory enables the enhancement of a member’s LE capability by incorporating its historical knowledge, allowing for full utilization of such historical data. Furthermore, the historical knowledge of members is updated progressively. Compared to members from the previous generation, the overall quality of members in the current iteration tends to be higher. Therefore, the extent to which historical member knowledge is utilized should gradually diminish as the historical time lapse increases. This approach not only further ensures the algorithm’s LE capability but also effectively leverages historical knowledge. Among them, the fractional-order weighted members are calculated using Equation (18).
X i f r a c = 1 1 ! · q · X i t + 1 2 ! · q · ( 1 q ) · X i t 1 + 1 3 ! · q · ( 1 q ) · ( 2 q ) ·   X i t 2 + 1 4 ! · q · ( 1 q ) · ( 2 q ) · ( 3 q ) · X i t 3 + r a n d · ( X i t X r t )
where X i f r a c represents the fractional-order weighted member. X i t signifies the knowledge of the i t h member at the t t h iteration, X i t 1 indicates the knowledge of the i t h member at the ( t 1 ) t h iteration, X i t 2 signifies the knowledge of the i t h member at the ( t 2 ) t h iteration, X i t 3 represents the knowledge of the i t h member at the ( t 3 ) t h iteration, and X r t stands for the knowledge of the rand member at the t t h iteration. The adaptive factor q is calculated using Equation (19).
q = 1 1 + e L · c o s ( 2 · π · L )
where L is represented using Equation (20). The function value of q with respect to the increase in the iteration number t is depicted in Figure 3. As can be observed from Figure 3, the value of q tends to exhibit a chaotic and fluctuating state. Meanwhile, after the iteration process reaches its mid-point, the fluctuations in the value of q become more pronounced. The setting of this value is primarily aimed at mitigating the issue of the algorithm easily falling into local optima caused by fractional-order weighting. By introducing certain fluctuations in the weighting process through q , the risk of getting stuck in local optima can be reduced. Furthermore, these fluctuations should be more significant in the later stages of iteration compared to the early stages. This is precisely the reason behind our design of the value-determination method for q .
L = 1 · t T 2 · r a n d + 1
Moreover, the second-order Bernstein polynomial is explicated as Equation (21), and its graph is visualized in Figure 4.
B 0 , 2 ( p ) = ( 1 p ) 2 B 1 , 2 ( p ) = 2 · p · ( 1 p ) B 2 , 2 ( p ) = p 2
As can be observed from the figure, when the value of p lies within the interval [0, 1], the sum of the three polynomials equals 1. This indicates that they possess the property of member weighting. Therefore, we employ the second-order Bernstein polynomial to weight the ideal member, the superior member, and a random member, generating a Bernstein-weighted member, which is explicated as Equation (22).
X B e r n = B 0 , 2 ( p ) · X b e s t + B 1 , 2 ( p ) · X b e t t e r + B 2 , 2 ( p ) · X r a n d
where X B e r n represents the second-order Bernstein-weighted member. After generating the fractional-order weighted member and the Bernstein-weighted member, the members in the swarm implement the Fractional-Order Bernstein exploitation strategy through Equation (23). This significantly augments the algorithm’s LE capability and effectively ensures the CA of the feature subsets.
X i n e w = X i + 0.5 · 1 t T · ( X i f r a c X i ) + 0.5 · 1 t T · ( X B e r n X i )
Subsequently, the knowledge of the i t h member is preserved using Equation (14). By guiding the swarm with the Fractional-Order Bernstein exploitation strategy, the algorithm not only augments its LE capability but also maintains a certain degree of randomness. This prevents the algorithm from easily getting trapped in local sub-ideal feature subset traps.

3.4. Implementation of the AMFPOA

In response to the shortcomings of the POA when resolving the FS problem, this section advances three improvement strategies to alleviate these deficiencies, resulting in an enhanced version of the POA, referred to as AMFPOA. To visually demonstrate the execution logic of AMFPOA when resolving optimization problems, Figure 5 presents the execution flowchart of AMFPOA.
Below, we will conduct a detailed analysis of the steps involved in employing the AMFPOA to resolve optimization problems, as depicted in Figure 5.
Step 1:
Set the operational parameters, namely the swarm size N and the MN of iterations T . Initialize the current member index i = 1, iteration counter t = 0 , and index j = 1.
Step 2:
Update the ideal solution based on the current state of the swarm.
Step 3:
If r a n d < 0.5 , update the member’s position using Equation (4). Otherwise, update it using Equation (13). Here, r a n d represents a randomly produced number within the range [0, 1].
Step 4:
Preserve the member’s knowledge using either Equation (6) or Equation (14), depending on specific criteria or predefined rules.
Step 5:
Calculate the member’s position knowledge using Equations (15) through (17).
Step 6:
If r a n d < 0.5 , update the member’s position using Equation (7). Otherwise, update it using Equation (23).
Step 7:
Preserve the member’s knowledge again, this time using either Equation (8) or Equation (14), as appropriate.
Step 8:
Increment the iteration counter: t = t + 1. Check if i is equal to N . If i = = N , save the ideal solution obtained during the current iteration. Then, check if t < T where T is the MN of iterations. If t T , reset i = 1 and j = 1, and jump back to Step 2. If t > T , terminate the algorithm iteration and output the ideal solution. If the pre-condition i N holds, increment i = i + 1 and reset j = 1, then jump back to Step 3.

3.5. Time Complexity

This subsection primarily focuses on comparing the time complexity of the AMFPOA and POA. In the primordial POA, there are two main components: swarm initialization and algorithm iteration. Here, we use the evaluation of the FFV as the fundamental step for time-complexity estimation. In the initialization stage, the time complexity is O ( N · D i m ) , where N represents the number of members in the swarm, and D i m signifies the dimensionality of each member. During the iteration process, the time complexity for a single iteration is O ( 2 · N · D i m ) . Therefore, for T iterations, where T represents the MN of iterations of the algorithm, the time complexity is O ( 2 · N · D i m · T ) . Consequently, the overall time complexity of the POA is O ( N · D i m · ( 1 + 2 T ) ) . The initialization process of the improved AMFPOA has the same time complexity as that of the POA, which is O ( N · D i m ) . During the iteration process, the member position-updating method in AMFPOA is combined with the updating method of the primordial POA in an equal-probability manner. Considering the evaluation of the FFV as the basic step, the time complexity of its iteration process remains O ( 2 · N · D i m · T ) . Thus, the time complexity of the AMFPOA is also O ( N · D i m · ( 1 + 2 T ) ) , which is the same as that of the POA. The above analysis is based on taking the evaluation of the FFV as the basic step to calculate the time complexity of both the AMFPOA and POA, and there is no significant difference between them. However, from the perspective of program execution, in the AMFPOA, necessary intermediate variable calculations are required before a member updates its position. This will increase its computational cost. Nevertheless, the difference in computational cost between AMFPOA and POA does not span orders of magnitude and can be considered negligible. We deem this acceptable, as it means sacrificing a small amount of time in exchange for a higher accuracy return.

4. Results and Discussion

This section primarily focuses on evaluating the FS efficacy of the AMFPOA. We conducted experiments on 23 real-world FS datasets and compared the results with eight state-of-the-art algorithms to comprehensively assess the FS efficacy of AMFPOA. The detailed knowledge of the 23 FS datasets is presented in Table 3, and the specific parameters of the eight comparative algorithms are outlined in Table 4. Moreover, to ensure the rationality of the experiments, each experiment was independently and non-repetitively run 30 times. We objectively and comprehensively evaluated the FS efficacy of AMFPOA by analyzing the FFV, CA, feature subset size, and Friedman non-parametric test results obtained from these 30 experimental runs. Additionally, to maintain fairness in the comparative experiments, the swarm size was set to 40, and the MN of iterations was set to 100. All experiments involved in this paper were coded and executed using MATLAB R2021b on a WINDOWS 11 operating system. Below, we will provide a detailed analysis of the FS problem model and the experimental results. Before explaining the experiment, we describe the detailed knowledge of the comparative algorithm.
  • Detailed description of the comparison algorithm:
(1) 
POA [73]: A novel swarm-based optimization algorithm advanced in 2024 demonstrated superior efficacy by being compared with 12 state-of-the-art optimization algorithms on the CEC 2017 test suite and 26 real-world problems. Its execution logic framework for resolving optimization problems remains consistent with that of AMFPOA.
(2) 
ALSHADE [80]: An improved swarm-based optimization algorithm introduced in 2022 attained better efficacy through comparison with six champion algorithms on the CEC 2014, CEC 2018, and unmanned aerial vehicle resource allocation problems. Its execution logic framework for optimization problem-resolving is in line with that of AMFPOA.
(3) 
PLO [81]: A new swarm-based optimization algorithm put forward in 2024 outperformed 17 optimization algorithms on the CEC 2022 and multi-threshold image segmentation problems. Its execution logic framework for resolving optimization problems is consistent with that of AMFPOA.
(4) 
LSHADE [82]: An improved swarm-based optimization algorithm advanced in 2014, which was the champion algorithm of the 2014 CEC competition, showed better efficacy by being compared with multiple champion algorithms on the CEC 2014 problems. Its execution logic framework for optimization problem-resolving is the same as that of AMFPOA.
(5) 
BEGJO [83]: An improved binary swarm-based optimization algorithm introduced in 2024 attained better efficacy through comparison with various optimization algorithms on the FS problem. Its execution logic framework for resolving optimization problems is consistent with that of AMFPOA.
(6) 
IPOA [84]: An improved swarm-based optimization algorithm advanced in 2024 attained better efficacy by being compared with multiple optimization algorithms on the CEC 2017 and scheduling problems. Its execution logic framework for optimization problem-resolving is in line with that of AMFPOA.
(7) 
MCOA [85]: An improved swarm-based optimization algorithm put forward in 2024 outperformed various algorithms on the CEC 2020 problems and the FS problem. Its execution logic framework for resolving optimization problems is consistent with that of AMFPOA.
(8) 
QHDBO [86]: An improved swarm-based optimization algorithm introduced in 2024 demonstrated better efficacy through comparison with multiple algorithms on 37 optimization problems. Its execution logic framework for optimization problem-resolving remains consistent with that of AMFPOA.

4.1. Establishment of the FS Problem Model

This section primarily focuses on establishing the FS problem model. The FS problem aims to reduce RF knowledge in the primordial dataset while enhancing the CA of the feature subset. Specifically, by searching through numerous combinations of features in the primordial dataset, we seek to identify a feature subset that balances both the size of the feature subset and CA, thereby effectively representing the primordial dataset and reducing the cost of model instruction. Consequently, the objective function primarily incorporates metrics for classification error rate and feature subset size, with the fitness function explicated using Equation (24).
min f ( X i ) = λ 1 · e r r o r + λ 2 · R n
where X i represents the i t h feature subset combination, e r r o r signifies the classification error rate when using feature subset X i , n indicates the total number of feature elements in the primordial dataset, and R signifies the number of feature elements in the selected feature subset. λ 1 is a constant within the interval [0, 1], with λ 2 = 1 λ 1 . With regard to the selection of the parameter λ 1 , we consulted the literature [87,88]. In these references, the parameter λ 1 is consistently set to 0.9 to ensure the model’s applicability. This well-chosen parameter can further enhance the efficacy when resolving real-world FS problems, making it more in line with the application requirements in practical scenarios. Therefore, we selected λ 1 = 0.9 in this paper, which is reasonably justified. By considering CA as the most vital efficacy metric, our FS model is predisposed to select features that can substantially improve CA. This approach not only effectively reduces RF but also boosts CA, thereby endowing the model with a certain degree of generalizability during its application.
The FS problem is a classic combinatorial optimization problem involving discrete variables. Therefore, when applying the AMFPOA to resolve the FS problem, it is necessary to convert the continuous real-valued members produced during algorithm iteration into binary (0–1) discrete values. Here, 0 indicates that the corresponding feature is not selected, while 1 signifies that the feature is selected. Below, we provide a detailed description of the numerical conversion and computational steps when using AMFPOA for FS, with the workflow illustrated in Figure 6.
Step 1:
Randomly sample 80% of the data from the primordial dataset to serve as the instruction set, while designating the remaining 20% of the data as the test set.
Step 2:
Convert the continuous real-valued member X i = ( x i , 1 , x i , 2 , x i , j , , x i , d i m ) into a discrete member Y i = ( y i , 1 , y i , 2 , y i , j , , y i , d i m ) using Equation (25). Here, y i , j = 1 indicates that the j t h feature element is selected in the i t h feature subset combination, while y i , j = 0 signifies its exclusion.
y i , j = 1 i f   x i , j < 0.5 0 i f   x i , j 0.5 ,     i = 1 , 2 , , N ,   j = 1 , 2 , , d i m .
Step 3:
Select feature subset elements from the primordial dataset based on the discrete member Y i derived from the continuous-to-discrete conversion.
Step 4:
Compute the CA of the selected feature subset using the K-Nearest Neighbors (KNN) classifier. In this study, K is set to 5 for the KNN algorithm. Simultaneously, 5-fold cross-validation is employed for cross-validation.
Step 5:
Calculate the FFV of the selected feature subset combination using Equation (24).

4.2. Sensitivity Analysis

In this section, a sensitivity analysis is primarily conducted on the settings of the swarm size and the number of iterations for AMFPOA. First, to reasonably set the swarm size, we employ the control variable method. We fix the MN of iterations at 100 and consider five different swarm sizes: 20, 30, 40, 60, and 100. Subsequently, by analyzing the efficacy of the AMFPOA under these various swarm sizes, we determine the ideal swarm size for algorithm execution. The experimental results are illustrated in Figure 7. After establishing an appropriate swarm size, we set the MN of iterations to 50, 80, 100, 150, and 200, respectively, to observe the convergence behavior of the AMFPOA during execution and select a suitable MN of iterations. The experimental results are presented in Figure 8.
Figure 7 displays a bar chart illustrating the average rankings of the AMFPOA when resolving 23 FS problems under different swarm size settings. As evident from the chart, when the swarm size was set to 20 and 30, the AMFPOA attained average rankings of 4.39 and 2.39, respectively. However, when the swarm size was set to 40, it attained an impressive average ranking of 1.17. The primary reason for this phenomenon is that an excessively small swarm size results in too few search agents during the algorithm’s process of exploring feature subsets. Consequently, the algorithm fails to extensively explore the solution space, making it prone to falling into local optima and ultimately compromising the final CA. Furthermore, when the swarm size was set to 60 and 100, the average rankings were 2.57 and 4.48, respectively, which were noticeably inferior to the average ranking attained with a swarm size of 40. This discrepancy can be attributed to the fact that although a larger swarm size enables the algorithm to thoroughly explore the solution space, the excessive number of search agents subsequently hinders effective convergence among members within the swarm, thereby weakening the algorithm’s ultimate search efficacy. In summary, we opted for a swarm size of 40 for subsequent experiments. This choice takes into account both the algorithm’s global search capability and its convergence efficacy, allowing for better optimization of the algorithm’s overall effectiveness.
Figure 8 presents the convergence curves of the AMFPOA on a selection of FS problems. As can be observed from the graph, under most circumstances, the algorithm begins to exhibit a relatively stable convergence trend by the 70th iteration. As iterations progress, the algorithm’s convergence becomes increasingly stable after the 100th iteration. Throughout the optimization process from the 100th to the 200th iteration, the algorithm’s optimization capability remains highly consistent. Therefore, to avoid unnecessary waste of computational resources, we set the swarm size at 100. This configuration not only ensures that the algorithm achieves stable convergence when addressing FS problems but also effectively conserves computational resources. Consequently, we have established the MN of iterations at 100 for subsequent experiments.

4.3. Swarm Diversity Analysis

This section primarily focuses on analyzing the PD of AMFPOA when resolving FS problems. An algorithm with high PD contributes to enhancing its ability to escape from local optimum feature subset traps and ensures a more extensive exploration of the FS solution space. We employ the moment of inertia I C to quantify PD, as depicted in Equation (26). Here, c d represents the degree of dispersion of the swarm from its centroid c during each iteration, where the centroid c is defined as shown in Equation (27). The parameter x i d signifies the value of the d t h dimension of the i t h search agent at the t t h iteration [77].
I C ( t ) = i = 1 N d = 1 D i m ( x i d ( t ) c d ( t ) ) 2
c d ( t ) = 1 D i m i = 1 N x i d ( t )
The specific experimental results are illustrated in Figure 9, where the blue line represents the PD of AMFPOA in resolving FS problems, and the red line indicates the PD of POA in addressing the same FS problems. As can be observed from Figure 9, during the early iterations of the algorithm, the PD corresponding to AMFPOA consistently surpasses that of POA. This implies that, compared to POA, AMFPOA exhibits stronger localization capability towards local solution regions within the solution space during the iterative process, thereby enhancing the probability of the algorithm locating the global optimum region. This is primarily attributed to the adaptive exploration strategy advanced in this paper, which effectively improves the algorithm’s GE capability through learning from member differences. Furthermore, as the iterative process progresses, the PD of AMFPOA remains higher than that of POA even in the later iterations. This is mainly due to the introduction of a three-swarm search strategy during the iteration, which achieves a good balance between the GE stage and the LE stage, thereby enhancing the algorithm’s ability to escape from local optimum feature subsets and increasing its PD. In summary, the introduction of the learning strategy in this paper effectively augments the PD of the algorithm, strengthening its GE capability and its ability to escape from local optimum feature subsets. This contributes to improving the CA of feature subsets and effectively reducing RF in the dataset.

4.4. Exploration/Exploitation Balance Analysis

This section primarily focuses on analyzing the exploration/exploitation balance of AMFPOA when resolving FS problems. An optimization algorithm with superior efficacy should strike a good balance between these two stages. It first locates potential solution regions through the GE stage and then conducts further searches within the identified regions through the LE stage. This approach significantly augments the optimization accuracy during algorithm execution. We utilize Equations (28) and (29) to calculate the percentages of GE and LE of the algorithm, respectively. Here, D i v ( t ) represents the diversity measurement in the dimensional space, which is computed using Equation (30). Specifically, D i v m a x signifies the maximum diversity throughout the entire iterative process, and m e d i a n ( x d ( t ) ) indicates the median value of the d t h dimensional values of members during the t t h iteration [89,90,91].
E x p l o r a t i o n ( % ) = D i v ( t ) D i v m a x × 100
E x p l o i t a t i o n ( % ) = | D i v ( t ) D i v m a x | D i v m a x × 100
D i v ( t ) = 1 D i m d = 1 D i m 1 N i = 1 N m e d i a n ( x d ( t ) ) x i d ( t )
The experimental results are depicted in Figure 10, where the blue line represents the GE rate of AMFPOA, and the red line indicates its LE rate. As can be seen from Figure 10, during the early iterations of the algorithm, the GE stage of AMFPOA dominates. In this stage, the primary focus is on locating potential ideal solution regions to strengthen the algorithm’s ability to explore the entire problem search space. This is largely attributed to the adaptive exploration strategy advanced in this paper, which augments the algorithm’s GE capability by adaptively learning from the differences among various members. As the iterative process progresses, the GE rate gradually declines and subsequently achieves a balance with the LE stage. During this stage, AMFPOA conducts further searches within the potentially ideal solution regions identified in the early iterations, thereby appropriately improving the CA of feature subsets. This is primarily due to the reasonable balance attained between these two stages by the three-swarm search strategy, which employs targeted adjustment strategies for different attribute subswarms. In the later iterations, the LE capability of AMFPOA becomes dominant. In this stage, the algorithm thoroughly explores the solution regions identified in the early iterations to enhance the CA and optimization speed when resolving FS problems. This is mainly attributed to the Fractional-Order Bernstein Exploitation Strategy advanced in this paper, which effectively augments the algorithm’s LE capability by leveraging historical swarm knowledge and the weighted nature of Bernstein polynomials. Additionally, it can be observed that the algorithm still preserves a certain degree of GE capability in the later iterations, which helps ensure its ability to escape from local non-ideal feature subset traps. In summary, the strategies advanced in this paper enhance the algorithm’s GE and LE capabilities from different perspectives, achieving a good balance between the two stages. This effectively improves the CA and optimization efficiency of the algorithm when resolving FS problems.

4.5. Ablation Analysis of Strategy Effectiveness

In this section, ablation experiments are conducted primarily to evaluate the effectiveness of the adaptive search strategy, multi-swarm search strategy, and Fractional-Order Bernstein exploitation strategy incorporated in AMFPOA. Specifically, we define APOA by introducing the adaptive search strategy into the POA, MPOA by incorporating the multi-swarm search strategy, and FPOA by amalgamating the Fractional-Order Bernstein exploitation strategy. Meanwhile, AMFPOA is formed by introducing all three learning strategies into POA. Subsequently, the above-defined algorithms are employed to resolve 23 FS problems to assess the effectiveness of each learning strategy. Here, we employ the Friedman non-parametric test to calculate the corresponding rankings of the algorithms. Specifically, the Friedman test is a non-parametric statistical test method designed to compare the efficacy differences among multiple related samples. In this paper, the steps for calculating the Friedman rankings are as follows:
Step 1:
Suppose we have k algorithms undergoing efficacy comparison, with each algorithm undergoing N N independent and non-repetitive experiments (in this case, k = = 9 and N N = 30 ). Record the FFV of each algorithm in each experiment.
Step 2:
For each algorithm j ( j = 1 , , k ) , calculate its average rank R j across all experiments, as explicated in Equation (31).
R j = 1 N N i = 1 N N r i j
Here, r i j represents the rank of algorithm j in the i t h experiment (where 1 indicates the best efficacy and k indicates the worst).
Step 3:
Calculate the Friedman statistic Q . The statistic Q in the Friedman test is used to assess whether the differences among the algorithms are significant. It is computed as shown in Equation (32).
Q = 12 N N k ( k + 1 ) j = 1 k R j 2 k ( k + 1 ) 2 4
Step 4:
Compare the statistic Q with the critical value to determine whether there are significant differences among the algorithms.
Step 5:
Output the Friedman average ranks R j corresponding to each algorithm and determine whether significant differences exist among the algorithms.
The experimental results of the Friedman non-parametric test for the FFV of the algorithms are presented in Figure 11. As can be seen from the figure, compared to the Friedman ranking of the primordial POA, APOA, MPOA, and FPOA all show advantages in ranking. This result confirms that each of the three learning strategies can enhance FS efficacy of the algorithm from different perspectives, validating the effectiveness of each strategy in improving algorithm efficacy. Moreover, AMFPOA exhibits a better Friedman ranking compared to the other algorithms, indicating that introducing all three learning strategies simultaneously into POA contributes to a further enhancement of the algorithm’s FS efficacy compared to introducing a single learning strategy. The aforementioned experimental results confirm the effectiveness of each learning strategy in promoting the algorithm’s FS efficacy. Notably, AMFPOA, which incorporates all three learning strategies, can further boost the algorithm’s FS efficacy, achieving higher CA and precision.
The above description confirms that the integration of the three learning strategies into the POA can effectively enhance its efficacy in resolving FS problems as a whole. However, the efficacy of each strategy in low-dimensional, medium-dimensional, and high-dimensional FS problems has not been thoroughly validated. Therefore, the following section will provide a detailed discussion on the efficacy improvements of the three enhancements in FS problems of different dimensions. Specifically, by applying APOA, MPOA, and FPOA to resolve 23 FS problems, we analyze the average fitness function value they achieve and summarize their efficacy improvement ratios. The experimental results are presented in Table 5, and the comparative rankings of POA against APOA, MPOA, and FPOA are illustrated in Figure 12.
Firstly, as can be seen from Table 5, when using FPOA to resolve low-dimensional FS problems, its fitness function value improved by 0.53% on the Banana problem compared to POA. Similarly, its efficacy improved by 30.91% on the Iris problem, 9.54% on the Bupa problem, 15.74% on the Glass problem, 14.85% on the Breastcancer problem, 4.17% on the Lipid problem, and 40.23% on the HeartEW problem. By synthesizing the aforementioned experimental results, we found that the introduction of the Fractional-Order Bernstein exploitation strategy into the primordial POA resulted in an average efficacy improvement rate of 14.5%. This is primarily because, when resolving low-dimensional FS problems, the solution space is relatively small, thus requiring the algorithm to possess stronger local exploitation capabilities to effectively explore this solution space and enhance the classification accuracy of the feature subset. The Fractional-Order Bernstein exploitation strategy introduced in this paper, which integrates fractional-order weighted members and Bernstein-weighted members, effectively augments the algorithm’s local exploitation capabilities. The experimental phenomena observed in low-dimensional FS problems confirm this point, validating the improvement effect of the Fractional-Order Bernstein exploitation strategy on local exploitation capabilities when resolving low-dimensional problems. Meanwhile, the advantages of FPOA in low-dimensional FS problems can also be visually observed in Figure 12a.
Secondly, as evident from Table 5, when employing APOA to resolve medium-dimensional FS problems, its fitness function value improved by 7.02% on the Zoo problem compared to POA. Similarly, its efficacy enhanced by 21.32% on the Vote problem, 38.31% on the Congress problem, 18.64% on the Lymphography problem, 13.04% on the Vehicle problem, 22.55% on the WDBC problem, 28.66% on the BreastEW problem, and 67.42% on the SonarEW problem. By synthesizing the aforementioned experimental results, we discovered that the introduction of the adaptive exploration strategy into the primordial POA led to an average efficacy improvement rate of 27.12%. This is primarily because, when addressing medium-dimensional FS problems, as the number of feature elements increases, the search space of the solution region gradually expands, necessitating the algorithm to possess robust global search capabilities to effectively explore the feature subsets. The adaptive exploration strategy introduced in this paper effectively augments the algorithm’s global exploration capabilities by amalgamating knowledge differences among members in the swarm along with adaptability. This enables the algorithm to conduct extensive searches for effective feature subsets as the dimensionality of FS increases, confirming the advantages of the adaptive exploration strategy in improving the algorithm’s global exploration efficacy. Additionally, the advantages of APOA in medium-dimensional FS problems can be visually observed in Figure 12b.
Finally, as can be seen from Table 5, when utilizing MPOA to resolve high-dimensional FS problems, its fitness function value improved by 16.48% on the Libras problem compared to POA. Similarly, its efficacy enhanced by 18.18% on the Hillvalley problem, 24.59% on the Musk problem, 47.34% on the Clean problem, 45.17% on the Semeion problem, 65.64% on the Madelon problem, and 56.71% on the Isolet problem. By synthesizing the aforementioned experimental results, we found that the introduction of the multi-swarm search strategy into the primordial POA resulted in an average efficacy improvement rate of 39.16%, indicating a notably significant enhancement. This is primarily because, when addressing high-dimensional FS problems, the substantial increase in feature elements renders the search space of the solution region exceedingly complex. Consequently, the algorithm may struggle to achieve a favorable trade-off across multiple metrics, leading to the obtained feature subsets being prone to local optima. Therefore, the algorithm is required to possess a strong ability to escape from the trap of local ideal feature subsets, enabling the obtained feature subsets to achieve a better balance across various metrics. The multi-swarm search strategy introduced in this paper integrates swarms with diverse characteristics, allowing different subswarms to undergo targeted update guidance. This effectively strengthens the algorithm’s ability to balance across multiple metrics and augments its capacity to escape from local optima. The experimental results also corroborate this point, demonstrating significant efficacy improvements when resolving high-dimensional FS problems. Additionally, the advantages of MPOA in high-dimensional FS problems can be visually observed in Figure 12c.

4.6. Fitness Function Value Analysis

To evaluate the efficacy of the AMFPOA in resolving FS problems, this section analyzes the FFV obtained by AMFPOA when reducing dimensionality across 23 FS datasets spanning low-dimensional, medium-dimensional, and high-dimensional scenarios. The results are compared with eight state-of-the-art algorithms, with experimental outcomes summarized in Table 6. In Table 6, “MIN”, “AVG”, and “MAX” denote the minimum, average, and maximum FFV obtained by each algorithm over 30 independent non-repetitive trials, respectively. “Mean Rank” represents the algorithm’s average efficacy ranking, while “Final Rank” indicates its ultimate ranking based on the “Mean Rank” metric. Below, we provide a detailed analysis of the obtained FFV.
As shown in Table 6, when resolving low-dimensional FS problems, AMFPOA attained first place in the minimum FFV metric across six FS datasets, with a winning rate of 75%, outperforming the second-ranked ALSHADE by 37.5%. This result demonstrates AMFPOA’s superior capability in locating ideal feature subsets for low-dimensional FS problems, primarily due to the Fractional-Order Bernstein exploitation strategy introduced in this study, which effectively augments the algorithm’s LE efficacy and improves optimization precision. Additionally, in terms of the average FFV, AMFPOA ranked first in eight FS datasets for dimensionality reduction, achieving a 100% success rate over the comparative algorithms. This indicates AMFPOA’s superior solution stability when addressing low-dimensional FS problems. To visually illustrate this stability, Figure 13 presents box plots of the algorithm’s efficacy across 30 independent runs, among them, "+" represents an outlier. The figure reveals that AMFPOA consistently exhibits narrower box widths, reflecting higher solution stability under most conditions. Finally, in the maximum FFV metric, AMFPOA secured first place in seven FS datasets, with an 87.5% success rate, demonstrating its enhanced fault tolerance and reduced error rates in FS problem-resolving. In summary, compared to the benchmark algorithms, AMFPOA exhibits superior optimization efficacy and higher solution stability when resolving low-dimensional FS problems.
As shown in Table 6, when resolving medium-dimensional FS problems, AMFPOA attained first place in the minimum FFV metric across seven FS datasets, with a success rate of 87.5%, demonstrating superior redundancy reduction capability compared to benchmark algorithms. This improvement is primarily attributed to the adaptive exploration strategy advanced in this study, which effectively augments the algorithm’s global search capability, enabling AMFPOA to efficiently explore multiple solution regions and thereby improve the quality of selected FS subsets. Additionally, in terms of the average FFV, AMFPOA ranked first in eight FS datasets for dimensionality reduction, achieving a 100% success rate over comparative algorithms. This outstanding efficacy reflects the enhanced stability of AMFPOA due to its optimized search strategy. As illustrated in Figure 13, AMFPOA consistently exhibits higher solution stability under most conditions, indicating its strong practical applicability. Finally, in the maximum FFV metric, AMFPOA secured first place in five FS datasets, with a 62.5% success rate, outperforming the second-ranked algorithm by approximately 50%. This demonstrates AMFPOA’s superior fault tolerance and improved robustness in FS problem-resolving. In summary, compared to benchmark algorithms, AMFPOA exhibits higher optimization efficacy and greater solution stability when addressing medium-dimensional FS problems, making it a promising FS method for real-world applications.
As shown in Table 6, when addressing high-dimensional FS problems, AMFPOA attained first place in the minimum FFV metric across six FS datasets, with a success rate of 85.7%. This demonstrates its superior capability in efficiently exploring subset combinations compared to benchmark algorithms. This improvement is primarily attributed to the three-swarm search strategy advanced in this study, which effectively augments the algorithm’s ability to escape local non-ideal feature subset traps. Consequently, AMFPOA can better tackle the challenges posed by high-dimensional FS problems, improving CA and enhancing dataset utility. Additionally, in terms of the average FFV, AMFPOA ranked first in seven FS datasets for dimensionality reduction, achieving a 100% success rate over comparative algorithms. This reflects its higher solution stability when resolving high-dimensional FS problems. As illustrated in Figure 13, AMFPOA consistently exhibits greater solution stability under most conditions, ensuring its practical applicability to a certain extent. Finally, in the maximum FFV metric, AMFPOA secured first place in five FS datasets, with a 71.4% success rate, demonstrating a clear advantage over benchmark algorithms and showcasing its stronger fault tolerance in problem-resolving. In summary, compared to benchmark algorithms, AMFPOA exhibits higher optimization efficacy and greater solution stability when addressing high-dimensional FS problems, making it a promising FS method for real-world applications.
The aforementioned analysis of FFV confirms that AMFPOA is an algorithm with robust FS efficacy. However, it must be acknowledged that AMFPOA underperforms compared to benchmark algorithms on certain specific FS datasets, indicating room for further improvement in its efficacy. From a comprehensive perspective, Figure 14 illustrates the average ranking of algorithms across 23 FS datasets. Notably, AMFPOA achieves the lowest bar heights in terms of minimum, average, and maximum FFV, demonstrating its strong overall efficacy. These findings suggest that AMFPOA can be considered a promising FS method when evaluated holistically.

4.7. Wilcoxon Rank Sum Test Analysis

The aforementioned analysis delved into the numerical results of the AMFPOA when resolving FS problems. However, numerical results may occasionally be skewed by outliers, potentially impacting the overall findings. To mitigate this randomness, this section presents a Wilcoxon non-parametric test conducted on the results obtained from 30 independent and non-repetitive experiments, with the significance level set at 0.05. The experimental results are summarized in Table 7. In the last row of the table, “+” indicates that the efficacy of the compared algorithm is significantly superior to that of the AMFPOA, “−” signifies that the compared algorithm’s efficacy is significantly inferior to that of the AMFPOA, and “=” signifies that there is no significant difference in efficacy between the compared algorithm and the AMFPOA. As can be seen from Table 7, when addressing 23 FS problems encompassing low-dimensional, medium-dimensional, and high-dimensional scenarios, the POA, ALSHADE, PLO, LSHADE, BEGJO, and QHDBO algorithms exhibit significantly inferior efficacy compared to the AMFPOA on 22 of these FS problems, representing a ratio of 95.6%. Meanwhile, the IPOA and MCOA algorithms also demonstrate significantly worse efficacy than the AMFPOA across all 23 FS problems. Moreover, out of the 184 tests conducted, the AMFPOA significantly outperforms the competing algorithms in 97.2% of the cases. This evidence confirms that the AMFPOA possesses superior FS efficacy. It also validates that the integration of the three learning strategies advanced in this paper effectively augments the efficacy of the POA.

4.8. CA Analysis

CA is a critical metric in FS problems, as it directly reflects the quality of the selected feature subsets. This section primarily analyzes the CA attained by AMFPOA when resolving FS problems. The experimental results are presented in Table 8, where “Mean Rank” signifies the algorithm’s average ranking in terms of CA, and “Final Rank” represents its average ranking based on the feature subset size metric.
As demonstrated in Table 8, when addressing low-dimensional FS problems, AMFPOA secured first place in seven FS problems, achieving the highest success rate compared to benchmark algorithms. This indicates that AMFPOA effectively eliminates RF while enhancing the utility of primordial low-dimensional datasets. This improvement is primarily attributed to the advanced strategy’s enhancement of LE capability, enabling higher optimization precision in low-dimensional FS problem-resolving. Additionally, for medium-dimensional FS problems, AMFPOA attained the highest CA in seven FS problems, with a success rate of 87.5%. This success stems from the advanced learning strategy, which facilitates comprehensive exploration of solution spaces, allowing AMFPOA to maintain efficient search efficacy despite increasing dataset dimensionality, thereby improving data utility. Furthermore, in high-dimensional FS problems, AMFPOA attained the highest CA in six FS problems, with a success rate of 85.7%. This is largely due to the three-swarm search strategy, which improves algorithmic balance during solution exploration, enabling easier escape from local feature subset traps caused by extreme dimensionality increases and yielding higher-quality feature subsets for primordial dataset representation. Figure 15 illustrates the average CA rankings of algorithms across 23 FS datasets, showing that AMFPOA consistently achieves lower average rankings compared to benchmark algorithms. In summary, the three learning strategies advanced in this study significantly enhance AMFPOA’s search efficacy, enabling superior FS problem-resolving capabilities relative to benchmark algorithms. By effectively improving CA, AMFPOA demonstrates strong potential as a promising FS method.

4.9. Feature Subset Size Analysis

This section primarily analyzes feature subset size, another critical metric in FS problems. The experimental results are presented in Table 9, where “Mean Rank” signifies the algorithm’s average ranking in terms of feature subset size across 23 FS problems, and “Final Rank” represents the algorithm’s overall ranking based on the “Mean Rank” metric. A detailed analysis of the experimental results follows.
As shown in Table 9, when addressing low-dimensional FS problems, all algorithms effectively reduced the number of feature elements. Notably, AMFPOA secured first place in six FS problems, achieving a success rate of 75%, demonstrating superior RF elimination capability compared to benchmark algorithms. This advantage primarily stems from AMFPOA’s enhanced LE capability, enabling effective removal of noisy features. Additionally, for medium-dimensional FS problems, AMFPOA ranked first in five FS problems, with a success rate of 62.5%, outperforming the second-ranked QHDBO by 25%. This efficacy highlights AMFPOA’s superior RF elimination in medium-dimensional scenarios and validates the advanced adaptive exploration strategy, which improves global search efficacy by optimizing feature subsets across the entire solution space, thereby enhancing dataset utility. Furthermore, in high-dimensional FS problems, AMFPOA attained first place in three FS problems (42.8%) and second place in four FS problems, exhibiting stronger feature reduction capability than benchmark algorithms. This confirms that AMFPOA’s three-swarm search strategy enables comprehensive consideration of multiple metrics during exploration, effectively eliminating RF and further improving dataset utility and model reliability. To demonstrate AMFPOA’s overall efficacy, Figure 16 illustrates the algorithm’s average ranking across 23 FS problems, showing that AMFPOA consistently achieves lower average rankings, indicating strong feature dimensionality reduction capability. In summary, the integration of learning strategies in AMFPOA establishes it as a highly efficient FS method, capable of effectively reducing RF in datasets and enhancing the utility of data models.

4.10. Convergence Analysis

The aforementioned discussions have substantiated that the AMFPOA exhibits favorable FS efficacy. However, in practical applications, the convergence efficacy of an algorithm is also of critical importance. An appropriate algorithm should not only possess high convergence accuracy but also demonstrate a satisfactory convergence speed. Therefore, in this section, we conduct an analysis of the convergence curves of the AMFPOA to evaluate its convergence efficacy. The experimental results are presented in Figure 17. Here, the x-axis represents the number of iterations, while the y-axis represents the average FFV.
As can be observed from Figure 17, most algorithms are in a state of stable convergence, making it meaningful to analyze the convergence curves under such stable conditions. In the majority of cases, for instance, when resolving FS problems on datasets like Glass, Breastcancer, Lipid, Vote, and Vehicle, the AMFPOA takes the lead in efficacy after approximately the 20th iteration, demonstrating rapid convergence speed and high convergence accuracy. In a small number of cases, such as when addressing FS problems on the HeartEW, Zoo, and Congress datasets, the AMFPOA achieves a leading position in terms of FFV after around the 40th iteration. This reflects the algorithm’s superior global search capabilities, which enhance its convergence speed as the problem dimensionality increases. As the iteration process progresses, the AMFPOA consistently maintains its leading edge while gradually approaching a state of stable convergence. These advantages are primarily attributable to the three learning strategies advanced in this paper, which effectively strengthen the algorithm’s search ability. Consequently, the AMFPOA attains faster convergence speed and higher convergence accuracy, exhibiting a certain degree of algorithmic applicability.

4.11. Comprehensive Metric Analysis

The preceding section conducted separate analyses of AMFPOA’s efficacy on FS problems across three key metrics: FFV, CA, and feature subset size. These analyses confirmed AMFPOA’s superior feature subset search capability, demonstrating its effectiveness in eliminating RF from raw datasets while significantly improving CA. To evaluate AMFPOA’s efficacy more comprehensively, this section integrates these three metrics into a multi-indicator analysis, visualized through the stacked bar chart shown in Figure 18. As illustrated, AMFPOA exhibits the lowest stacked height across all five aggregated indicators, confirming its strong ability to balance multiple efficacy metrics. This balanced optimization enables AMFPOA to achieve superior overall FS efficacy, establishing it as a promising FS method.

4.12. Extended Experimental Analysis

This section primarily aims to explore and expand the efficacy of the AMFPOA when addressing FS problems with tens of thousands of dimensions. In this regard, we employ seven datasets, each containing over ten thousand feature-related knowledge entries, for experimental evaluation. The specific details of these datasets are presented in Table 10. Similarly, each evaluation experiment is independently and non-repetitively conducted 30 times to gather statistical data on various metrics, including the ideal FFV, the average FFV, and the worst FFV. The experimental results are displayed in Table 11. Here, “Mean Rank” represents the average ranking of an algorithm’s FFV across the seven FS problems, while “Final Rank” indicates the algorithm’s ultimate ranking based on the “Mean Rank” metric.
As can be observed from the table, with regard to the ideal FFV metric, the AMFPOA ranks first across all seven datasets, demonstrating its remarkable ability to search for ideal feature subsets. This also validates that the learning strategies advanced in this paper can still exert a certain positive influence on the algorithm’s efficacy even when the feature dimensionality is extremely high. Moreover, concerning the average FFV metric, the AMFPOA achieves the best values on all seven listed datasets. This showcases the algorithm’s high solution stability. As the feature dimensionality increases, the AMFPOA can still maintain its solution stability. Finally, in terms of the worst FFV metric, the AMFPOA attains the ideal values on five datasets, with a success rate of 71.4%. This indicates that it has better fault-tolerance compared to the competing algorithms. Additionally, to provide a more intuitive demonstration of the AMFPOA’s efficacy in resolving ultra-high-dimensional FS problems, Figure 19 presents the average rankings of the algorithm across the three metrics. It is clearly evident that the AMFPOA significantly outperforms the competing algorithms in all three metrics. This also reflects that the AMFPOA has better applicability and holds advantages in resolving ultra-high-dimensional datasets compared to the competing algorithms.

5. Conclusions and Future Works

This study addresses the limitations of the POA in resolving FS problems, specifically its deficiencies in GE and LE and imbalance between exploration and exploitation, which compromise CA and dataset utility, and advances an enhanced POA variant termed AMFPOA by amalgamating three complementary strategies to improve FS efficacy and subset CA: first, to address POA’s insufficient GE, an adaptive exploration strategy is introduced, which augments GE capability by considering knowledge disparities among different groups of members and incorporating adaptive learning principles, enabling thorough exploration of the feature subset combination space and improving CA; second, to resolve the imbalance between GE and LE, a three-swarm search strategy is advanced, which divides the swarm into multiple subswarms and applies tailored optimization strategies to each, achieving superior multi-metric balance in FS problem-resolving and optimizing feature subset quality; third, to strengthen POA’s LE capability, a Fractional-Order Bernstein exploitation strategy is developed, which leverages the historical memory properties of Fractional-Order theory and the multi-member weighting of Bernstein polynomials to significantly enhance LE, yielding higher CA in FS tasks; experimental results on 23 real-world FS problems demonstrate that AMFPOA achieves an average success rate of over 87.5% in FFV, along with ideal efficacy rates of 86.5% in CA and 60.1% in feature subset size reduction, exhibiting superior FS efficacy compared to benchmark algorithms and establishing itself as a promising FS method.
Based on the work presented in this paper, our follow-up research plans mainly revolve around the following two aspects: (1) The AMFPOA advanced in this paper is a single-objective optimization algorithm. In our subsequent work, we will introduce its corresponding multi-objective version to further expand the algorithm’s application domains. (2) We will apply AMFPOA to resolve more challenging combinatorial optimization problems to further evaluate its efficacy in combinatorial optimization.

Author Contributions

Conceptualization, L.X.; methodology, L.X. and J.L.; software, L.X.; validation, L.X., J.L. and Y.Y.; formal analysis, L.X. and J.L.; investigation, Y.Y.; resources, L.X.; data curation, L.X., J.L. and Y.Y.; writing—primordial draft preparation, L.X. and J.L.; writing—review and editing, Y.Y.; visualization, L.X.; supervision, Y.Y.; project administration, Y.Y.; funding acquisition, Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All experimental data can be obtained by contacting the corresponding author.

Acknowledgments

We greatly appreciate the efforts of the reviewers and editorial team for our article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gomes, B.; Ashley, E.A. Artificial intelligence in molecular medicine. N. Engl. J. Med. 2023, 388, 2456–2465. [Google Scholar] [CrossRef] [PubMed]
  2. Xu, Q.; Cai, J.R.; Zhang, W.; Bai, J.W.; Li, Z.Q.; Tan, B.; Sun, L. Detection of citrus Huanglongbing (HLB) based on the HLB-induced leaf starch accumulation using a home-made computer vision system. Biosyst. Eng. 2022, 218, 163–174. [Google Scholar] [CrossRef]
  3. Wang, X.; Jiang, H.; Zeng, T.; Dong, Y. An adaptive fused domain-cycling variational generative adversarial network for machine fault diagnosis under data scarcity. Knowl. Fusion 2025, 126, 103616. [Google Scholar] [CrossRef]
  4. Wu, Q.; Gu, J. Design and research of robot visual servo system based on artificial intelligence. Agro Food Ind. Hi-Tech 2017, 28, 125–128. [Google Scholar]
  5. Chen, J.; Zhang, M.; Xu, B.; Sun, J.; Mujumdar, A.S. Artificial intelligence assisted technologies for controlling the drying of fruits and vegetables using physical fields: A review. Trends Food Sci. Technol. 2020, 105, 251–260. [Google Scholar] [CrossRef]
  6. Zhu, C.; Hao, S.; Liu, C.; Wang, Y.; Jia, X.; Xu, J.; Wang, W. An Efficient Computer Vision-Based Dual-Face Target Precision Variable Spraying Robotic System for Foliar Fertilisers. Agronomy 2024, 14, 2770. [Google Scholar] [CrossRef]
  7. Zhang, L.; Liao, B.; Liu, D.; Jiang, Q.; Sun, Q. Artificial Intelligence empowered evolution in medicine food homology: Innovations, Challenges, and Future Prospects. Food Biosci. 2025, 69, 106928. [Google Scholar] [CrossRef]
  8. Li, H.; Geng, W.; Hassan, M.M.; Zuo, M.; Wei, W.; Wu, X.; Ouyang, Q.; Chen, Q. Rapid detection of chloramphenicol in food using SERS flexible sensor coupled artificial intelligent tools. Food Control 2021, 128, 108186. [Google Scholar] [CrossRef]
  9. El-Mesery, H.S.; Qenawy, M.; Ali, M.; Hu, Z.; Adelusi, O.A.; Njobeh, P.B. Artificial intelligence as a tool for predicting the quality attributes of garlic (Allium sativum L.) slices during continuous infrared-assisted hot air drying. J. Food Sci. 2024, 89, 7693–7712. [Google Scholar] [CrossRef] [PubMed]
  10. El-Mesery, H.S.; Qenawy, M.; Ali, M.; Rostom, M.; Elbeltagi, A.; Salem, A.; Elwakeel, A.E. Optimization of dried garlic physicochemical properties using a self-organizing map and the development of an artificial intelligence prediction model. Sci. Rep. 2025, 15, 3105. [Google Scholar] [CrossRef]
  11. Chen, Q.; Hu, W.; Su, J.; Li, H.; Ouyang, Q.; Zhao, J. Nondestructively sensing of total viable count (TVC) in chicken using an artificial olfaction system based colorimetric sensor array. J. Food Eng. 2016, 168, 259–266. [Google Scholar] [CrossRef]
  12. Li, H.; Kutsanedzie, F.; Zhao, J.; Chen, Q. Quantifying total viable count in pork meat using combined hyperspectral imaging and artificial olfaction techniques. Food Anal. Methods 2016, 9, 3015–3024. [Google Scholar] [CrossRef]
  13. Li, L.; Xie, S.; Zhu, F.; Ning, J.; Chen, Q.; Zhang, Z. Colorimetric sensor array-based artificial olfactory system for sensing Chinese green tea’s quality: A method of fabrication. Int. J. Food Prop. 2017, 20 (Suppl. 2), 1762–1773. [Google Scholar] [CrossRef]
  14. Li, H.; Zhang, B.; Hu, W.; Liu, Y.; Dong, C.; Chen, Q. Monitoring black tea fermentation using a colorimetric sensor array-based artificial olfaction system. J. Food Process. Preserv. 2018, 42, e13348. [Google Scholar] [CrossRef]
  15. Bai, J.W.; Xiao, H.W.; Ma, H.L.; Zhou, C.S. Artificial neural network modeling of drying kinetics and color changes of ginkgo biloba seeds during microwave drying process. J. Food Qual. 2018, 2018, 3278595. [Google Scholar] [CrossRef]
  16. Dai, C.; Huang, X.; Huang, D.; Lv, R.; Sun, J.; Zhang, Z.; Aheto, J.H. Real-time detection of saponin content during the fermentation process of Tremella aurantialba using a homemade artificial olfaction system. J. Food Process Eng. 2019, 42, e13101. [Google Scholar] [CrossRef]
  17. Yunfeng, X.; Xiliang, Z.; Xiaojia, S.; Jizhang, W.; Jizhan, L.; Zhiguo, L.; Pingping, L. Tensile mechanical properties of greenhouse cucumber cane. Int. J. Agric. Biol. Eng. 2016, 9, 1–8. [Google Scholar]
  18. Jin, Y.; Liu, J.; Xu, Z.; Yuan, S.; Li, P.; Wang, J. Development status and trend of agricultural robot technology. Int. J. Agric. Biol. Eng. 2021, 14, 1–19. [Google Scholar] [CrossRef]
  19. Yüksel, N.; Börklü, H.R.; Sezer, H.K.; Canyurt, O.E. Review of artificial intelligence applications in engineering design perspective. Eng. Appl. Artif. Intell. 2023, 118, 105697. [Google Scholar] [CrossRef]
  20. Liu, S.; Li, H.; Hassan, M.M.; Ali, S.; Chen, Q. SERS based artificial peroxidase enzyme regulated multiple signal amplified system for quantitative detection of foodborne pathogens. Food Control 2021, 123, 107733. [Google Scholar] [CrossRef]
  21. Zareef, M.; Arslan, M.; Hassan, M.M.; Ahmad, W.; Ali, S.; Li, H.; Ouyang, Q.; Wu, X.; Hashim, M.M.; Chen, Q. Recent advances in assessing qualitative and quantitative aspects of cereals using nondestructive techniques: A review. Trends Food Sci. Technol. 2021, 116, 815–828. [Google Scholar] [CrossRef]
  22. Xu, Y.; Hassan, M.M.; Sharma, A.S.; Li, H.; Chen, Q. Recent advancement in nano-optical strategies for detection of pathogenic bacteria and their metabolites in food safety. Crit. Rev. Food Sci. Nutr. 2023, 63, 486–504. [Google Scholar] [CrossRef]
  23. Wang, H.; Gu, J.; Wang, M. A review on the application of computer vision and machine learning in the tea industry. Front. Sustain. Food Syst. 2023, 7, 1172543. [Google Scholar] [CrossRef]
  24. Liang, Z.; Wada, M.E. Development of cleaning systems for combine harvesters: A review. Biosyst. Eng. 2023, 236, 79–102. [Google Scholar] [CrossRef]
  25. El-Mesery, H.S.; Qenawy, M.; Li, J.; El-Sharkawy, M.; Du, D. Predictive modeling of garlic quality in hybrid infrared-convective drying using artificial neural networks. Food Bioprod. Process. 2024, 145, 226–238. [Google Scholar] [CrossRef]
  26. Jing, T.; Tang, Z.; Hao, S.; Shen, C.; Wang, T.; Wang, M. Structure design and rice threshing efficacy of the variable-speed inertial pulley for simulating artificial threshing. Int. J. Agric. Biol. Eng. 2024, 17, 33–40. [Google Scholar]
  27. Ngolong Ngea, G.L.; Yang, Q.; Xu, M.; Ianiri, G.; Dhanasekaran, S.; Zhang, X.; Bi, Y.; Zhang, H. Revisiting the current and emerging concepts of postharvest fresh fruit and vegetable pathology for next-generation antifungal technologies. Compr. Rev. Food Sci. Food Saf. 2024, 23, e13397. [Google Scholar] [CrossRef]
  28. Lakhiar, I.A.; Yan, H.; Zhang, C.; Wang, G.; He, B.; Hao, B.; Han, Y.; Wang, B.; Bao, R.; Rakibuzzaman, M.; et al. A review of precision irrigation water-saving technology under changing climate for enhancing water use efficiency, crop yield, and environmental footprints. Agriculture 2024, 14, 1141. [Google Scholar] [CrossRef]
  29. Cao, Y.; Tang, Z.; Lu, D.; Lin, S. Efficacy Test of Artificial Defoliating Broccoli Conveyor Line and Analysis of Defoliating Broccoli Inflorescences. Agronomy 2024, 14, 1925. [Google Scholar] [CrossRef]
  30. Zhao, S.; Adade, S.Y.S.S.; Wang, Z.; Jiao, T.; Ouyang, Q.; Li, H.; Chen, Q. Deep learning and feature reconstruction assisted vis-NIR calibration method for on-line monitoring of key growth indicators during kombucha production. Food Chem. 2025, 463, 141411. [Google Scholar] [CrossRef]
  31. Ji, T.; Liaqat, F.; Khazi, M.I.; Liaqat, N.; Nawaz, M.Z.; Zhu, D. Lignin biotransformation: Advances in enzymatic valorization and bioproduction strategies. Ind. Crops Prod. 2024, 216, 118759. [Google Scholar] [CrossRef]
  32. Bai, J.; Mujumdar, A.S.; Xiao, H. Ethical and strategic challenges of AI weapons: A call for global action. Int. J. Agric. Biol. Eng. 2024, 17, 293–294. [Google Scholar] [CrossRef]
  33. Wu, P.; Lei, X.; Zeng, J.; Qi, Y.; Yuan, Q.; Huang, W.; Ma, Z.; Shen, Q.; Lyu, X. Research progress in mechanized and intelligent zed pollination technologies for fruit and vegetable crops. Int. J. Agric. Biol. Eng. 2024, 17, 11–21. [Google Scholar]
  34. Deng, J.; Jiang, H.; Chen, Q. Enhancing Fourier Transform Near-infrared Spectroscopy with Explainable Ensemble Learning Methods for Detecting Mineral Oil Contamination in Corn Oil. J. Food Compos. Anal. 2025, 143, 107594. [Google Scholar] [CrossRef]
  35. Jan, Z.; Ahamed, F.; Mayer, W.; Patel, N.; Grossmann, G.; Stumptner, M.; Kuusk, A. Artificial intelligence for industry 4.0: Systematic review of applications, challenges, and opportunities. Expert Syst. Appl. 2023, 216, 119456. [Google Scholar] [CrossRef]
  36. Taha, M.F.; Mao, H.; Zhang, Z.; Elmasry, G.; Awad, M.A.; Abdalla, A.; Mousa, S.; Elwakeel, A.E.; Elsherbiny, O. Emerging technologies for precision crop management towards agriculture 5.0: A comprehensive overview. Agriculture 2025, 15, 582. [Google Scholar] [CrossRef]
  37. Jayan, H.; Min, W.; Guo, Z. Applications of Artificial Intelligence in Food Industry. Foods 2025, 14, 1241. [Google Scholar] [CrossRef]
  38. Li, D.; Chen, Q.; Ouyang, Q.; Liu, Z. Advances of Vis/NIRS and imaging techniques assisted by AI for tea processing. Crit. Rev. Food Sci. Nutr. 2025, 1–19. [Google Scholar] [CrossRef]
  39. Wang, Y.; Zhang, Z.; Jia, W.; Ou, M.; Dong, X.; Dai, S. A review of environmental sensing technologies for targeted spraying in orchards. Horticulturae 2025, 11, 551. [Google Scholar] [CrossRef]
  40. Guan, B.; Zhao, J.; Jin, H.; Lin, H. Determination of rice storage time with colorimetric sensor array. Food Analytical Methods. 2017, 10, 1054–1062. [Google Scholar] [CrossRef]
  41. Xu, Y.; Hassan, M.M.; Kutsanedzie, F.Y.H.; Li, H.H.; Chen, Q.S. Evaluation of extra-virgin olive oil adulteration using FTIR spectroscopy combined with multivariate algorithms. Qual. Assur. Saf. Crops Foods 2018, 10, 411–421. [Google Scholar] [CrossRef]
  42. Zhao, Z.; Jin, M.; Tian, C.; Yang, S.X. Prediction of seed distribution in rectangular vibrating tray using grey model and artificial neural network. Biosyst. Eng. 2018, 175, 194–205. [Google Scholar] [CrossRef]
  43. Xu, Y.; Chen, Q.; Liu, Y.; Sun, X.; Huang, Q.; Ouyang, Q.; Zhao, J. A novel hyperspectral microscopic imaging system for evaluating fresh degree of pork. Korean J. Food Sci. Anim. Resour. 2018, 38, 362. [Google Scholar]
  44. Chen, H.; Chen, B.; Lu, D. A novel method for detection of camellia oil adulteration based on time-resolved emission fluorescence. Sci. Rep. 2018, 8, 13784. [Google Scholar] [CrossRef]
  45. Jia, W.; Zheng, Y.; Zhao, D.A.; Yin, X.; Liu, X.; Du, R. Preprocessing method of night vision image application in apple harvesting robot. Int. J. Agric. Biol. Eng. 2018, 11, 158–163. [Google Scholar] [CrossRef]
  46. Li, Y.; Sun, J.; Wu, X.; Lu, B.; Wu, M.; Dai, C. Grade identification of tieguanyin tea using fluorescence hyperspectra and different statistical algorithms. J. Food Sci. 2019, 84, 2234–2241. [Google Scholar] [CrossRef]
  47. Li, Y.; Sun, J.; Wu, X.; Chen, Q.; Lu, B.; Dai, C. Detection of viability of soybean seed based on fluorescence hyperspectra and CARS-SVM-AdaBoost model. J. Food Process. Preserv. 2019, 43, e14238. [Google Scholar] [CrossRef]
  48. Wu, X.; Zhu, J.; Wu, B.; Zhao, C.; Sun, J.; Dai, C. Discrimination of Chinese liquors based on electronic nose and fuzzy discriminant principal component analysis. Foods 2019, 8, 38. [Google Scholar] [CrossRef] [PubMed]
  49. Entezari, A.; Aslani, A.; Zahedi, R.; Noorollahi, Y. Artificial intelligence and machine learning in energy systems: A bibliographic perspective. Energy Strategy Rev. 2023, 45, 101017. [Google Scholar] [CrossRef]
  50. Jia, W.; Sun, M.; Lian, J.; Hou, S. Feature dimensionality reduction: A review. Complex Intell. Syst. 2022, 8, 2663–2693. [Google Scholar] [CrossRef]
  51. Zhang, J.; Yu, J.; Tao, D. Local deep-feature alignment for unsupervised dimension reduction. IEEE Trans. Image Process. 2018, 27, 2420–2432. [Google Scholar] [CrossRef] [PubMed]
  52. Abdollahzadeh, B.; Gharehchopogh, F.S. A multi-objective optimization algorithm for FS problems. Eng. Comput. 2022, 38, 1845–1863. [Google Scholar] [CrossRef]
  53. Hashim, F.A.; Hussien, A.G. Snake Optimizer: A novel meta-heuristic optimization algorithm. Knowl.-Based Syst. 2022, 242, 108320. [Google Scholar] [CrossRef]
  54. Zervoudakis, K.; Tsafarakis, S. A global optimizer inspired from the survival strategies of flying foxes. Eng. Comput. 2023, 39, 1583–1616. [Google Scholar] [CrossRef]
  55. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey Wolf Optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  56. Abdollahzadeh, B.; Gharehchopogh, F.S.; Mirjalili, S. African vultures optimization algorithm: A new nature-inspired metaheuristic algorithm for global optimization problems. Comput. Ind. Eng. 2021, 158, 107408. [Google Scholar] [CrossRef]
  57. Braik, M.; Ryalat, M.H.; Al-Zoubi, H. A novel meta-heuristic algorithm for resolving numerical optimization problems: Ali Baba and the forty thieves. Neural Comput. Appl. 2022, 34, 409–455. [Google Scholar] [CrossRef]
  58. Matoušová, I.; Trojovský, P.; Dehghani, M.; Trojovská, E.; Kostra, J. Mother optimization algorithm: A new human-based metaheuristic approach for resolving engineering optimization. Sci. Rep. 2023, 13, 10312. [Google Scholar] [CrossRef]
  59. Dehghani, M.; Trojovský, P. Teamwork Optimization Algorithm: A New Optimization Approach for Function Minimization/Maximization. Sensors 2021, 21, 4567. [Google Scholar] [CrossRef]
  60. Faramarzi, A.; Heidarinejad, M.; Stephens, B.; Mirjalili, S. Equilibrium optimizer: A novel optimization algorithm. Knowl.-Based Syst. 2020, 191, 105190. [Google Scholar] [CrossRef]
  61. Hashim, F.A.; Hussain, K.; Houssein, E.H.; Mabrouk, M.S.; Al-Atabany, W. Archimedes optimization algorithm: A new metaheuristic algorithm for resolving optimization problems. Appl. Intell. 2021, 51, 1531–1551. [Google Scholar] [CrossRef]
  62. Hatamlou, A. Black hole: A new heuristic optimization approach for data clustering. Inf. Sci. 2013, 222, 175–184. [Google Scholar] [CrossRef]
  63. Goldberg, D.E.; Holland, J.H. Genetic Algorithms and Machine Learning. Mach. Learn. 1988, 3, 95–99. [Google Scholar] [CrossRef]
  64. Reynolds, R.G. An Introduction to Cultural Algorithms. In Proceedings of the Third Annual Conference on Evolutionary Programming, San Diego, CA, USA, 24–26 February 1994; World Scientific: Singapore, 1994; pp. 131–139. [Google Scholar]
  65. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  66. Hichem, H.; Elkamel, M.; Rafik, M.; Mesaaoud, M.T.; Ouahiba, C. A new binary grasshopper optimization algorithm for FS problem. J. King Saud Univ.-Comput. Knowl. Sci. 2022, 34, 316–328. [Google Scholar] [CrossRef]
  67. Allam, M.; Nandhini, M. Bestl FS using binary teaching learning based optimization algorithm. J. King Saud Univ.-Comput. Knowl. Sci. 2022, 34, 329–341. [Google Scholar] [CrossRef]
  68. Alweshah, M.; Khalaileh, S.A.; Gupta, B.B.; Almomani, A.; Hammouri, A.I.; Al-Betar, M.A. The monarch butterfly optimization algorithm for resolving FS problems. Neural Comput. Appl. 2022, 34, 11267–11281. [Google Scholar] [CrossRef]
  69. Hu, G.; Du, B.; Wang, X.; Wei, G. An enhanced black widow optimization algorithm for FS. Knowl.-Based Syst. 2022, 235, 107638. [Google Scholar] [CrossRef]
  70. Ewees, A.A.; Mostafa, R.R.; Ghoniem, R.M.; Gaheen, M.A. Improved seagull optimization algorithm using Lévy flight and mutation operator for FS. Neural Comput. Appl. 2022, 34, 7437–7472. [Google Scholar] [CrossRef]
  71. Awadallah, M.A.; Hammouri, A.I.; Al-Betar, M.A.; Braik, M.S.; Abd Elaziz, M. Binary Horse herd optimization algorithm with crossover operators for FS. Comput. Biol. Med. 2022, 141, 105152. [Google Scholar] [CrossRef] [PubMed]
  72. Pan, H.; Chen, S.; Xiong, H. A high-dimensional FS method based on modified Gray Wolf Optimization. Appl. Soft Comput. 2023, 135, 110031. [Google Scholar] [CrossRef]
  73. Al-Baik, O.; Alomari, S.; Alssayed, O.; Gochhait, S.; Leonova, I.; Dutta, U.; Dehghani, M. Pufferfish optimization algorithm: A new bio-inspired metaheuristic algorithm for resolving optimization problems. Biomimetics 2024, 9, 65. [Google Scholar] [CrossRef] [PubMed]
  74. Zhang, Q.; Gao, H.; Zhan, Z.H.; Li, J.; Zhang, H. Growth Optimizer: A powerful metaheuristic algorithm for resolving continuous and discrete global optimization problems. Knowl.-Based Syst. 2023, 261, 110206. [Google Scholar] [CrossRef]
  75. Shen, Y.; Zhang, C.; Gharehchopogh, F.S.; Mirjalili, S. An improved whale optimization algorithm based on multi-swarm evolution for global optimization and engineering design problems. Expert Syst. Appl. 2023, 215, 119269. [Google Scholar] [CrossRef]
  76. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  77. Chen, F.; Ye, S.; Wang, J.; Luo, J. Multi-Strategy Improved Binary Secretarial Bird Optimization Algorithm for FS. Mathematics 2025, 13, 668. [Google Scholar] [CrossRef]
  78. Wang, J.; Bao, Z.; Dong, H. An Improved Northern Goshawk Optimization Algorithm for Mural Image Segmentation. Biomimetics 2025, 10, 373. [Google Scholar] [CrossRef]
  79. Zhang, X.; Lin, Q. Three-learning strategy particle swarm algorithm for global optimization problems. Knowl. Sci. 2022, 593, 289–313. [Google Scholar] [CrossRef]
  80. Li, Y.; Han, T.; Zhou, H. A novel adaptive L-SHADE algorithm and its application in UAV swarm resource configuration problem. Knowl. Sci. 2022, 606, 350–367. [Google Scholar] [CrossRef]
  81. Yuan, C.; Zhao, D.; Heidari, A.A. Polar lights optimizer: Algorithm and applications in image segmentation and FS. Neurocomputing 2024, 607, 128427–128472. [Google Scholar] [CrossRef]
  82. Tanabe, R.; Fukunaga, A.S. Improving the search efficacy of SHADE using linear swarm size reduction. In Proceedings of the 2014 IEEE Congress on Evolutionary Computation (CEC), Beijing, China, 6–11 July 2014; IEEE: New Yor, NY, USA, 2014; pp. 1658–1665. [Google Scholar]
  83. Askr, H.; Abdel-Salam, M.; Hassanien, A.E. Copula entropy-based golden jackal optimization algorithm for high-dimensional FS problems. Expert Syst. Appl. 2024, 238, 121582–121607. [Google Scholar] [CrossRef]
  84. SeyedGarmroudi, S.D.; Kayakutlu, G.; Kayalica, M.O. Improved pelican optimization algorithm for resolving load dispatch problems. Energy 2024, 289, 129811–129826. [Google Scholar] [CrossRef]
  85. Jia, H.; Zhou, X.; Zhang, J. Modified crayfish optimization algorithm for resolving multiple engineering application problems. Artif. Intell. Rev. 2024, 57, 127–183. [Google Scholar] [CrossRef]
  86. Zhu, F.; Li, G.; Tang, H. Dung beetle optimization algorithm based on quantum computing and multi-strategy fusion for resolving engineering problems. Expert Syst. Appl. 2024, 236, 121219–121236. [Google Scholar] [CrossRef]
  87. Wu, B.; Luo, J. A Novel Improved Binary Optimization Algorithm and Its Application in FS Problems. Mathematics 2025, 13, 675. [Google Scholar] [CrossRef]
  88. Cao, Q.; Yuan, S.; Fang, Y. Three Strategies Enhance the Bionic Coati Optimization Algorithm for Global Optimization and FS Problems. Biomimetics 2025, 10, 380. [Google Scholar] [CrossRef]
  89. Nadimi-Shahraki, M.H.; Zamani, H. DMDE: Diversity-maintained multi-trial vector differential evolution algorithm for non-decomposition large-scale global optimization. Expert Syst. Appl. 2022, 198, 116895. [Google Scholar] [CrossRef]
  90. Li, Y.; Yang, C.; Zeng, H.; Dong, Z.; An, Z.; Xu, Y.; Wu, H. Frequency-Aligned Knowledge Distillation for Lightweight Spatiotemporal Forecasting. arXiv 2025, arXiv:2507.02939. [Google Scholar]
  91. Li, Y.; Dong, J.; Dong, Z.; Yang, C.; An, Z.; Xu, Y. SRKD: Towards Efficient 3D Point Cloud Segmentation via Structure-and Relation-aware Knowledge Distillation. arXiv 2025, arXiv:2506.17290. [Google Scholar]
Figure 1. Schematic of adaptive exploration strategy.
Figure 1. Schematic of adaptive exploration strategy.
Mathematics 13 02799 g001
Figure 2. Schematic of three-swarm search strategy.
Figure 2. Schematic of three-swarm search strategy.
Mathematics 13 02799 g002
Figure 3. The variation curve of dynamic function q.
Figure 3. The variation curve of dynamic function q.
Mathematics 13 02799 g003
Figure 4. Second-order Bernstein polynomial function curve.
Figure 4. Second-order Bernstein polynomial function curve.
Mathematics 13 02799 g004
Figure 5. Execution flowchart of AMFPOA.
Figure 5. Execution flowchart of AMFPOA.
Mathematics 13 02799 g005
Figure 6. Calculation process of FFV for FS problem.
Figure 6. Calculation process of FFV for FS problem.
Mathematics 13 02799 g006
Figure 7. Average ranking of AMFPOA corresponding to different swarm sizes.
Figure 7. Average ranking of AMFPOA corresponding to different swarm sizes.
Mathematics 13 02799 g007
Figure 8. AMFPOA iterative convergence curve. (a) Breastcancer. (b) Lipid. (c) HeartEW. (d) Zoo. (e) Vote. (f) Congress. (g) Lymphography. (h) Vehicle.
Figure 8. AMFPOA iterative convergence curve. (a) Breastcancer. (b) Lipid. (c) HeartEW. (d) Zoo. (e) Vote. (f) Congress. (g) Lymphography. (h) Vehicle.
Mathematics 13 02799 g008
Figure 9. Swarm diversity of algorithms. (a) Aggregation. (b) Glass. (c) Congress. (d) BreastEW. (e) Hillvalley. (f) Isolet.
Figure 9. Swarm diversity of algorithms. (a) Aggregation. (b) Glass. (c) Congress. (d) BreastEW. (e) Hillvalley. (f) Isolet.
Mathematics 13 02799 g009
Figure 10. Exploration and exploitation ratio of algorithms. (a) Aggregation. (b) Glass. (c) Congress. (d) BreastEW. (e) Hillvalley. (f) Isolet.
Figure 10. Exploration and exploitation ratio of algorithms. (a) Aggregation. (b) Glass. (c) Congress. (d) BreastEW. (e) Hillvalley. (f) Isolet.
Mathematics 13 02799 g010
Figure 11. Comparison of strategy effectiveness of algorithms.
Figure 11. Comparison of strategy effectiveness of algorithms.
Mathematics 13 02799 g011
Figure 12. Ranking of algorithms on different types of FS problems. (a) Low. (b) Medium. (c) High.
Figure 12. Ranking of algorithms on different types of FS problems. (a) Low. (b) Medium. (c) High.
Mathematics 13 02799 g012
Figure 13. Box plots for resolving FS problems. (a) Glass. (b) Breastcancer. (c) Congress. (d) Lymphography. (e) WDBC. (f) Hillvalley. (g) Musk. (h) Madelon. (i) Isolet.
Figure 13. Box plots for resolving FS problems. (a) Glass. (b) Breastcancer. (c) Congress. (d) Lymphography. (e) WDBC. (f) Hillvalley. (g) Musk. (h) Madelon. (i) Isolet.
Mathematics 13 02799 g013
Figure 14. Ranking of FFV.
Figure 14. Ranking of FFV.
Mathematics 13 02799 g014
Figure 15. Average ranking of CA.
Figure 15. Average ranking of CA.
Mathematics 13 02799 g015
Figure 16. Average ranking of feature subset size.
Figure 16. Average ranking of feature subset size.
Mathematics 13 02799 g016
Figure 17. Convergence curves of all algorithms. (a) Glass. (b) Breastcancer. (c) Lipid. (d) HeartEW. (e) Zoo. (f) Vote. (g) Congress. (h) Lymphography. (i) Vehicle.
Figure 17. Convergence curves of all algorithms. (a) Glass. (b) Breastcancer. (c) Lipid. (d) HeartEW. (e) Zoo. (f) Vote. (g) Congress. (h) Lymphography. (i) Vehicle.
Mathematics 13 02799 g017
Figure 18. Comprehensive indicator ranking fill chart.
Figure 18. Comprehensive indicator ranking fill chart.
Mathematics 13 02799 g018
Figure 19. Stacked plot of the ranking of FFV for the FS problem with tens of thousands of dimensions.
Figure 19. Stacked plot of the ranking of FFV for the FS problem with tens of thousands of dimensions.
Mathematics 13 02799 g019
Table 1. Comparison of related technologies.
Table 1. Comparison of related technologies.
TechnologyEfficacyLimitations
Binary grasshopper optimization FS algorithmSuccess rate with feature subset size exceeding 95%There are limitations in calculating costs
Binary teaching–learning-based optimization FS algorithmEffectively distinguished malignant tumors from benign tumorsLack of universality for FS problems
Monarch butterfly optimization FS algorithmThe average CA reaches 93%There are limitations in resolving high-dimensional FS problems
Black widow optimization FS algorithmReduction in RF has been attainedHigh-dimensional FS problems have limitations
Improved sine–cosine FS algorithmHigh CA of feature subsetsThe reduction in RF is not obvious
Binary Horse herd Optimization FS AlgorithmEffectively reduce RF in the primordial datasetIt is difficult to overcome the challenges posed by high-dimensional FS problems
Improved grey wolf optimization FS algorithmIt effectively reduces RFNot widely applicable
Table 2. Abbreviations of phrases involved in the paper.
Table 2. Abbreviations of phrases involved in the paper.
No.Full NameAbbreviation
1Pufferfish Optimization AlgorithmPOA
2FSFS
3Artificial IntelligenceAI
4Flying Fox OptimizationFFO
5Grey Wolf OptimizerGWO
6African Vulture Optimization AlgorithmAVOA
7War Strategy OptimizationWSO
8Mother Optimization AlgorithmMOA
9Teamwork Optimization AlgorithmTOA
10Equilibrium OptimizerEO
11Archimedes Optimization AlgorithmAOA
12Black Hole AlgorithmBHA
13Genetic AlgorithmGA
14Cultural AlgorithmCA
15Differential EvolutionDE
Table 3. Knowledge of 23 FS datasets.
Table 3. Knowledge of 23 FS datasets.
TypeNameFeature NumberInstance Size
LowBanana25300
Aggregation2788
Iris4150
Bupa6345
Breastcancer9699
Glass9214
Lipid10583
HeartEW13270
Congress16435
Vote16435
Zoo16101
Vehicle18846
Lymphography18148
BreastEW30569
WDBC30569
SonarEW60208
HighLibras90360
Hillvalley100606
Musk166476
Clean167476
Semeion2561593
Madelon5002600
Isolet6171559
Table 4. Parameter settings for eight comparative algorithms.
Table 4. Parameter settings for eight comparative algorithms.
AlgorithmsTimeParameter Settings
POA [73]2024 I { 1 , 2 }
ALSHADE [80]2022 N P init = 18 · D ,         N P min = 4 ,         | A | = 2.6 · N P ,         p = 0.11 ,         H = 6 ,         e = 0.5
PLO [81]2024 W 1 = 2 1 + e 2 ( t / T ) 4 1 ,   W 2 = e ( 2 t / T ) 3
LSHADE [82]2014 N P init = 18 · D ,         N P min = 4 ,         | A | = 2.6 · N P ,         p = 0.11 ,         H = 6
BEGJO [83]2024 β = 1.5 , c 1 = 1.5
IPOA [84]2024 F P K = 0.4 ,   F G K = 0.6
MCOA [85]2024 C 2 = 2 F E s M a x F E s
QHDBO [86]2024 R = cos π · t T max + 1 · 0.5
Table 5. Fitness function value of algorithms formed by introducing different strategies to FS problems of varying dimensions.
Table 5. Fitness function value of algorithms formed by introducing different strategies to FS problems of varying dimensions.
DatasetsPOAFPOAPercentageDatasetsPOAAPOAPercentageDatasetsPOAMPOAPercentage
Aggregation0.1000.1000.00%Zoo0.0520.0487.02%Libras0.1700.14216.48%
Banana0.1900.1890.53%Vote0.0480.03821.32%Hillvalley0.3740.30618.18%
Iris0.0550.03830.91%Congress0.0370.02338.31%Musk0.0840.06324.59%
Bupa0.3240.2939.54%Lymphography0.1070.08718.64%Clean0.0720.03847.34%
Glass0.2220.18715.74%Vehicle0.2910.25313.04%Semeion0.1390.07645.17%
Breastcancer0.0610.05214.85%WDBC0.0270.02122.55%Madelon0.2180.07565.64%
Lipid0.2560.2454.17%BreastEW0.0490.03528.66%Isolet0.2150.09356.71%
HeartEW0.1640.09840.23%SonarEW0.1320.04367.42%
Mean increase 14.50% 27.12% 39.16%
Table 6. The FFV when resolving FS problems.
Table 6. The FFV when resolving FS problems.
DatasetsMetricsPOAALSHADEPLOLSHADEBEGJOIPOAMCOAQHDBOAMFPOA
AggregationMIN0.1000.1000.1000.1000.1000.1060.1060.1000.100
AVG0.1000.1000.1000.1000.1000.1060.1060.1000.100
MAX0.1000.1000.1000.1000.1000.1060.1060.1000.100
BananaMIN0.1900.1930.2000.1980.1950.2040.2000.1930.187
AVG0.1900.1930.2000.1980.1950.2040.2000.1930.187
MAX0.1900.1930.2000.1980.1950.2040.2000.1930.187
IrisMIN0.0550.0550.0550.0800.0550.0550.0800.0800.025
AVG0.0550.0550.0550.0810.0550.0550.0810.0800.025
MAX0.0550.0550.0550.0850.0550.0550.0850.0800.025
BupaMIN0.3240.2980.3460.3070.2880.3330.3070.3070.281
AVG0.3240.3090.3470.3100.2890.3400.3090.3070.284
MAX0.3240.3460.3560.3370.2980.3670.3200.3070.298
GlassMIN0.2160.3020.2910.2590.2370.2480.3120.2150.173
AVG0.2220.3250.2930.2620.2570.2510.3210.2220.173
MAX0.2370.3440.3020.2700.3230.2580.3550.2790.173
BreastcancerMIN0.0610.0350.0420.0460.0480.0420.0530.0530.040
AVG0.0610.0500.0420.0470.0520.0450.0580.0530.040
MAX0.0610.0550.0420.0530.0590.0590.0680.0550.046
LipidMIN0.2490.2240.2470.2450.2550.2660.2530.2610.235
AVG0.2560.2480.2490.2590.2640.2740.2590.2640.241
MAX0.2740.2710.2610.2760.2860.2890.2710.2740.253
HeartEWMIN0.1460.1900.1230.1560.1140.0880.1560.0810.056
AVG0.1640.2120.1330.1650.1270.1030.1690.1140.080
MAX0.2080.2400.1550.1900.1550.1560.2140.1910.140
ZooMIN0.0380.0380.0760.0830.0380.0380.0380.0440.031
AVG0.0520.0530.0760.0940.0610.0610.0480.0710.046
MAX0.0760.0760.0760.1280.0950.1080.0630.1150.076
VoteMIN0.0420.0540.0350.0480.0660.0370.0500.0370.033
AVG0.0480.0640.0360.0480.0680.0390.0600.0370.034
MAX0.0540.0770.0390.0480.0790.0440.0700.0370.037
CongressMIN0.0370.0270.0580.0370.0460.0350.0270.0580.017
AVG0.0370.0410.0630.0370.0580.0420.0270.0580.017
MAX0.0370.0600.0680.0370.0790.0500.0270.0580.017
LymphographyMIN0.0900.1570.1010.0810.1260.1150.1410.1460.059
AVG0.1070.2000.1420.0890.1360.1280.1590.1620.066
MAX0.1260.2280.1660.1100.1630.1630.2200.1890.079
VehicleMIN0.2570.2790.2360.2520.2510.2250.2570.2410.230
AVG0.2910.2960.2470.2710.2750.2490.2840.2550.241
MAX0.3100.3150.2630.2940.3110.2630.3050.2730.257
WDBCMIN0.0230.0560.0370.0420.0580.0530.0460.0390.015
AVG0.0270.0650.0510.0670.0720.0640.0620.0400.017
MAX0.0310.0720.0590.0780.0900.0740.0830.0540.039
BreastEWMIN0.0360.0620.0360.0290.0350.0460.0440.0280.021
AVG0.0490.0690.0390.0340.0530.0580.0620.0360.031
MAX0.0590.0800.0460.0410.0710.0700.0840.0480.044
SonarEWMIN0.0890.1060.0300.0130.0280.0250.0120.0540.010
AVG0.1320.1460.0430.0420.0520.0470.0950.0730.020
MAX0.1560.1700.0740.0620.0820.0710.1310.0990.042
LibrasMIN0.1470.1540.1410.1090.1340.2000.2420.1740.109
AVG0.1700.1760.1580.1350.1560.2290.2730.1910.132
MAX0.1870.1870.1800.1520.1730.2630.3240.2080.146
HillvalleyMIN0.3470.3600.3010.3120.3440.3210.3190.2980.276
AVG0.3740.3760.3150.3420.3640.3380.3380.3100.299
MAX0.3890.3840.3340.3830.3730.3630.3560.3230.317
MuskMIN0.0620.0800.0470.0210.0430.0630.0970.0220.020
AVG0.0840.1040.0650.0450.0590.0770.1230.0520.035
MAX0.0980.1190.0900.0630.0840.1110.1490.0830.048
CleanMIN0.0570.0890.0400.0230.0470.0470.0560.0320.020
AVG0.0720.1090.0540.0330.0880.0630.0750.0530.030
MAX0.0910.1270.0680.0450.1090.0880.1010.0670.053
SemeionMIN0.1340.1220.0720.0590.1090.0810.1240.0690.070
AVG0.1390.1270.0820.0780.1150.0870.1370.0780.078
MAX0.1430.1340.0910.0900.1260.0950.1470.0860.090
MadelonMIN0.1860.2280.1020.1070.1890.1840.1230.0990.060
AVG0.2180.2410.1440.1260.2030.2020.1990.1270.079
MAX0.2650.2610.1910.1410.2210.2430.2570.1720.095
IsoletMIN0.1980.1510.0850.0750.1240.1300.1380.1120.070
AVG0.2150.1730.1080.0930.1460.1430.1760.1260.080
MAX0.2280.1860.1300.1050.1650.1610.2040.1450.099
Mean RankMIN5.3916.0874.5224.0875.0435.3916.3484.3911.217
AVG5.3916.5654.3914.2615.4785.3486.6964.3911.000
MAX5.1746.3483.8264.0435.5225.5656.8264.2171.261
Final RankMIN684256931
AVG683275931
MAX582367941
Table 7. Wilcoxon rank sum test results.
Table 7. Wilcoxon rank sum test results.
DatasetsPOAALSHADEPLOLSHADEBEGJOIPOAMCOAQHDBO
Aggregation9.110 × 10−2/=6.000 × 10−2/=9.200 × 10−2/=7.280 × 10−1/=8.290 × 10−2/=1.892 × 10−10/−6.956 × 10−10/−7.120 × 10−2/=
Banana4.108 × 10−4/−4.574 × 10−10/−7.146 × 10−10/−2.550 × 10−10/−7.087 × 10−10/−4.462 × 10−10/−6.548 × 10−10/−2.392 × 10−10/−
Iris7.930 × 10−5/−8.709 × 10−6/−8.512 × 10−6/−3.301 × 10−7/−7.322 × 10−10/−4.947 × 10−6/−5.880 × 10−8/−8.915 × 10−10/−
Bupa4.632 × 10−5/−8.900 × 10−6/−1.999 × 10−10/−9.077 × 10−10/−7.419 × 10−10/−5.243 × 10−10/−7.833 × 10−10/−2.413 × 10−10/−
Glass3.759 × 10−5/−5.281 × 10−7/−3.112 × 10−10/−3.862 × 10−10/−8.152 × 10−10/−1.258 × 10−10/−2.422 × 10−10/−8.418 × 10−10/−
Breastcancer8.523 × 10−10/−4.779 × 10−6/−7.172 × 10−5/−5.753 × 10−5/−7.568 × 10−10/−2.606 × 10−5/−8.336 × 10−6/−6.122 × 10−10/−
Lipid4.106 × 10−10/−1.040 × 10−6/−5.537 × 10−5/−3.546 × 10−5/−9.842 × 10−7/−8.518 × 10−5/−5.125 × 10−5/−2.049 × 10−7/−
HeartEW6.135 × 10−10/−3.467 × 10−6/−4.884 × 10−5/−4.749 × 10−5/−4.925 × 10−9/−7.337 × 10−5/−7.799 × 10−5/−6.806 × 10−7/−
Zoo8.118 × 10−4/−5.376 × 10−5/−3.168 × 10−6/−6.752 × 10−4/−6.953 × 10−7/−4.294 × 10−5/−9.395 × 10−4/−7.198 × 10−7
Vote8.687 × 10−4/−5.312 × 10−5/−4.068 × 10−5/−6.744 × 10−4/−4.834 × 10−7/−4.891 × 10−5/−8.118 × 10−4/−3.323 × 10−7/−
Congress9.206 × 10−5/−5.752 × 10−5/−4.553 × 10−6/−2.470 × 10−4/−2.908 × 10−7/−2.458 × 10−6/−4.939 × 10−4/−8.754 × 10−7/−
Lymphography9.863 × 10−4/−2.135 × 10−5/−4.500 × 10−5/−4.351 × 10−5/−3.919 × 10−7/−4.051 × 10−5/−7.508 × 10−5/−5.405 × 10−7/−
Vehicle8.115 × 10−4/−8.065 × 10−5/−4.072 × 10−5/−8.227 × 10−4/−7.462 × 10−7/−8.248 × 10−5/−8.966 × 10−4/−9.643 × 10−7/−
WDBC5.441 × 10−4/−8.676 × 10−6/−5.874 × 10−10/−8.534 × 10−5/−2.194 × 10−7/−2.735 × 10−10/−8.741 × 10−5/−9.746 × 10−7/−
BreastEW4.337 × 10−4/−1.552 × 10−5/−2.468 × 10−10/−5.237 × 10−5/−9.608 × 10−7/−8.287 × 10−10/−6.717 × 10−5/−7.761 × 10−7/−
SonarEW7.547 × 10−4/−8.412 × 10−5/−8.319 × 10−10/−3.371 × 10−5/−8.322 × 10−7/−9.829 × 10−10/−3.971 × 10−5/−6.279 × 10−7/−
Libras7.958 × 10−4/−4.018 × 10−5/−8.241 × 10−10/−5.606 × 10−5/−1.044 × 10−7/−4.766 × 10−10/−6.978 × 10−5/−4.869 × 10−7/−
Hillvalley7.086 × 10−4/−8.037 × 10−5/−3.034 × 10−10/−9.272 × 10−5/−6.400 × 10−10/−6.108 × 10−10/−4.109 × 10−5/−1.616 × 10−10/−
Musk8.895 × 10−4/−3.520 × 10−5/−6.301 × 10−10/−8.268 × 10−5/−4.725 × 10−10/−8.507 × 10−10/−3.247 × 10−5/−2.298 × 10−10/−
Clean6.316 × 10−4/−8.486 × 10−10/−2.961 × 10−10/−8.470 × 10−5/−5.054 × 10−10/−8.302 × 10−10/−9.154 × 10−5/−3.678 × 10−10/−
Semeion3.121 × 10−10/−2.848 × 10−10/−4.749 × 10−10/−2.417 × 10−10/−2.158 × 10−10/−1.291 × 10−10/−5.210 × 10−10/−8.495 × 10−10/−
Madelon5.407 × 10−7/−7.818 × 10−10/−5.090 × 10−10/−2.216 × 10−10/−8.459 × 10−10/−8.813 × 10−10/−8.294 × 10−10/−8.301 × 10−10/−
Isolet7.445 × 10−10/−9.694 × 10−10/−1.623 × 10−10/−6.925 × 10−5/−7.022 × 10−10/−9.167 × 10−10/−2.646 × 10−5/−6.848 × 10−10/−
+/−/=0/22/10/22/10/22/10/22/10/22/10/23/00/23/00/22/1
Table 8. CA in resolving FS problems.
Table 8. CA in resolving FS problems.
DatasetsPOAALSHADEPLOLSHADEBEGJOIPOAMCOAQHDBOAMFPOA
Aggregation100.00100.00100.00100.00100.0099.3699.36100.00100.00
Banana90.0089.7288.8789.1589.4388.4988.8789.6290.28
Iris96.6796.6796.6796.3396.6796.6796.0096.67100.00
Bupa69.5771.0164.9369.4272.4666.5269.4269.5775.07
Glass79.5269.2973.3375.7176.4375.4867.3877.8685.71
Breastcancer95.6897.7097.8498.3596.9197.7096.4097.2799.21
Lipid75.2674.9175.6073.4573.7970.8673.1972.4175.09
HeartEW85.3778.8988.5284.2690.0092.0483.7090.0093.15
Zoo98.5099.5095.0094.5098.0098.50100.0096.5098.50
Vote96.0994.7197.5995.4095.5296.6794.6096.5598.62
Congress96.5597.3695.1796.5596.6798.7497.7094.2598.85
Lymphography92.0780.0087.5993.1088.9788.9784.8385.1795.52
Vehicle70.9570.8376.2173.2573.3175.7472.3775.5077.16
WDBC97.8895.5895.2293.6394.4294.8794.4296.2898.85
BreastEW98.1495.5897.7098.0598.0597.1795.0498.3298.32
SonarEW88.7886.1098.0597.5698.0598.0592.2093.1799.27
Libras83.7583.0684.7287.6487.2278.7571.8180.2887.64
Hillvalley59.0959.4266.5364.3063.9766.9463.9766.8667.52
Musk93.7991.3795.1697.6898.1196.5389.2696.1198.53
Clean93.5890.9598.6395.8995.1697.6893.7995.6899.26
Semeion89.5989.8194.7294.4792.4895.5089.6996.7396.67
Madelon78.6276.1785.8188.1982.7982.8380.1786.9892.37
Isolet79.4283.9290.2992.0389.0789.3983.8388.3093.41
Mean Rank5.266.134.394.834.434.707.304.481.22
Final Rank782635941
Table 9. The size of feature subsets in resolving FS problems.
Table 9. The size of feature subsets in resolving FS problems.
DatasetsPOAALSHADEPLOLSHADEBEGJOIPOAMCOAQHDBOAMFPOA
Aggregation2.002.002.002.002.002.002.002.002.00
Banana2.002.002.002.002.002.002.002.002.00
Iris1.001.001.001.901.001.001.802.001.00
Bupa3.002.901.952.103.902.302.002.001.90
Glass3.404.404.803.904.002.702.502.002.40
Breastcancer2.002.602.002.902.202.202.302.602.00
Lipid3.302.202.902.002.801.201.801.601.70
HeartEW4.202.803.803.004.804.102.903.102.40
Zoo6.107.705.007.106.907.607.706.305.00
Vote2.102.603.801.004.501.401.902.001.00
Congress1.002.803.101.004.404.901.001.001.00
Lymphography6.403.605.504.806.605.204.105.104.70
Vehicle5.306.106.005.406.305.506.306.306.30
WDBC2.407.502.502.906.605.303.602.002.00
BreastEW9.708.705.505.8010.709.905.106.204.10
SonarEW18.6012.4015.1011.8020.6017.7014.707.207.90
Libras21.4021.0018.4021.0037.0033.8017.2011.8017.10
Hillvalley5.8010.5014.0021.0039.4040.4014.2012.206.60
Musk45.9043.5035.7039.7069.9075.6044.0027.9035.40
Clean28.1046.4029.3043.8073.6070.1032.0024.2024.00
Semeion115.0090.1087.7073.30121.40118.80111.80124.0087.00
Madelon129.10134.0052.9097.40242.90236.20104.3082.9050.30
Isolet182.70175.60130.50131.10295.10291.00188.90126.90125.10
Mean Rank4.655.004.044.177.045.744.303.431.65
Final Rank673498521
Table 10. Knowledge on the FS dataset with tens of thousands of dimensions.
Table 10. Knowledge on the FS dataset with tens of thousands of dimensions.
NameFeature NumberInstance SizeClass
Lung12,5332035
MLL12,582723
Ovarian15,1542532
Arcene10,0002002
RNA-Seq20,5318015
Dorothea100,0008002
CNS7129602
Table 11. FFV for the FS problem with tens of thousands of dimensions.
Table 11. FFV for the FS problem with tens of thousands of dimensions.
DatasetsMetricsPOAALSHADEPLOLSHADEBEGJOIPOAMCOAQHDBOAMFPOA
LungMIN0.1200.1520.0970.1380.1510.1410.1510.1590.094
AVG0.1410.1660.1180.1620.1820.1620.1770.1740.113
MAX0.1700.1700.1360.1860.2160.1880.1930.1880.139
MLLMIN0.3440.3330.3360.3100.2930.3720.3160.2920.270
AVG0.3640.3450.3540.3300.3150.3890.3450.3080.288
MAX0.3890.3610.3680.3460.3290.4080.3640.3370.297
OvarianMIN0.0580.0830.0290.0570.0560.0720.0670.0380.022
AVG0.0890.1060.0500.0700.0850.0890.0920.0470.031
MAX0.1090.1350.0700.0950.1100.1000.1310.0590.042
ArceneMIN0.0480.0890.0340.0460.0850.0540.0500.0360.017
AVG0.0700.1060.0540.0660.0970.0660.0640.0460.030
MAX0.1070.1230.0670.1050.1080.0860.0780.0600.038
RNA-SeqMIN0.1050.1060.0760.0820.1010.0790.1110.0840.062
AVG0.1090.1160.0890.0890.1110.0890.1240.0910.069
MAX0.1150.1250.0980.0980.1260.1070.1320.1000.083
DorotheaMIN0.1950.2040.0960.0960.1910.1810.1540.0960.082
AVG0.2540.2630.1250.1140.2210.2000.2200.1260.098
MAX0.2850.3050.1620.1330.2660.2240.2750.1730.135
CNSMIN0.1430.1630.0910.0830.1390.1370.1590.1110.069
AVG0.1600.1790.1160.0960.1600.1510.1780.1280.077
MAX0.1750.1870.1430.1150.1960.1640.2070.1440.088
Mean RankMIN6.2868.2862.8573.8575.8576.1436.5714.1431.000
AVG6.4298.0003.4293.4296.5715.5716.8573.7141.000
MAX6.2867.1433.2863.4296.8575.7147.2863.7141.286
Final RankMIN792356841
AVG692275841
MAX682375941
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, L.; Lv, J.; Yu, Y. Adapted Multi-Strategy Fractional-Order Relative Pufferfish Optimization Algorithm for Feature Selection. Mathematics 2025, 13, 2799. https://doi.org/10.3390/math13172799

AMA Style

Xu L, Lv J, Yu Y. Adapted Multi-Strategy Fractional-Order Relative Pufferfish Optimization Algorithm for Feature Selection. Mathematics. 2025; 13(17):2799. https://doi.org/10.3390/math13172799

Chicago/Turabian Style

Xu, Lukui, Jiajun Lv, and Youling Yu. 2025. "Adapted Multi-Strategy Fractional-Order Relative Pufferfish Optimization Algorithm for Feature Selection" Mathematics 13, no. 17: 2799. https://doi.org/10.3390/math13172799

APA Style

Xu, L., Lv, J., & Yu, Y. (2025). Adapted Multi-Strategy Fractional-Order Relative Pufferfish Optimization Algorithm for Feature Selection. Mathematics, 13(17), 2799. https://doi.org/10.3390/math13172799

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop