Next Article in Journal
Mapping EEG Metrics to Human Affective and Cognitive Models: An Interdisciplinary Scoping Review from a Cognitive Neuroscience Perspective
Previous Article in Journal
An Enhanced Randomized Dung Beetle Optimizer for Global Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Black-Winged Kite Algorithm for Global Optimization and Fault Detection

1
School of Electronic and Communication Engineering, Shenzhen Polytechnic University, Shenzhen 518055, China
2
Undergraduate School of Artificial Intelligence, Shenzhen Polytechnic University, Shenzhen 518055, China
3
School of Computer Science, University of Liverpool, Liverpool L69 3DR, UK
*
Authors to whom correspondence should be addressed.
Biomimetics 2025, 10(11), 728; https://doi.org/10.3390/biomimetics10110728
Submission received: 14 October 2025 / Revised: 26 October 2025 / Accepted: 29 October 2025 / Published: 1 November 2025
(This article belongs to the Section Biological Optimisation and Management)

Abstract

In the field of industrial fault detection, accurate and timely fault identification is crucial for ensuring production safety and efficiency. Effective feature selection (FS) methods can significantly enhance detection performance in this process. However, the recently proposed Black-winged Kite Algorithm (BKA) tends to suffer from premature convergence and local optima when handling high-dimensional feature spaces. To address these limitations, this paper proposes an improved Black-winged Kite Algorithm (IBKA). This algorithm integrates two novel enhancement mechanisms: First, the Stagnation-Triggered Diversification Mechanism monitors the algorithm’s convergence state and applies mild perturbations to the worst-performing individuals upon detecting stagnation, effectively preventing traps in local optima. Second, the Adaptive Weak Guidance Mechanism employs a conditional elite guidance strategy during the late optimization phase to provide subtle directional guidance to underperforming individuals, thereby improving convergence efficiency. We comprehensively evaluated the proposed IBKA across 26 benchmark functions. Results demonstrate superior performance in solution quality, convergence speed, and robustness compared to the original BKA and other advanced meta-heuristics. Furthermore, fault detection applications on public datasets validate the practical applicability of the binary version of the IBKA (bIBKA), showcasing significant improvements in detection accuracy and reliability. Experimental results confirm that these enhancement mechanisms effectively balance exploration and exploitation capabilities while preserving algorithmic simplicity and computational efficiency.

1. Introduction

In modern industrial production systems, equipment failures not only cause enormous economic losses but may also trigger serious safety accidents [1]. Therefore, accurate and timely fault detection has become a critical technology for ensuring industrial safety and production efficiency [2]. Industrial fault detection is essentially a complex optimization problem [3] that requires finding optimal feature data in high-dimensional feature spaces to achieve accurate identification and classification of fault patterns [4]. Meanwhile, in numerous industrial applications such as engineering design [5], parameter tuning [6], and resource allocation, the solution quality of optimization problems directly affects system performance and economic benefits. However, real-world fault detection problems often exhibit complex characteristics such as high dimensionality, multi-modality, nonlinearity, and strong coupling [7], making traditional mathematical optimization methods inadequate for handling such challenging problems.
Metaheuristic optimization algorithms [8], as a class of intelligent optimization techniques that do not rely on problem-specific mathematical properties, have demonstrated tremendous application potential in fault detection feature selection, classifier parameter optimization, and various engineering optimization problems due to their powerful global search capabilities and good adaptability [9]. These algorithms design effective search strategies to solve complex optimization problems by simulating intelligent phenomena in nature, such as biological evolution [10], swarm behavior [11], and physical processes [12]. In recent years, researchers have proposed numerous metaheuristic algorithms, including Pied Kingfisher Optimizer (PKO) [13], Red Fox Optimizer (RFO) [14], Sine Cosine Algorithm (SCA) [15], Genetic Algorithm (GA) [16], Particle Swarm Optimization (PSO) [17], Differential Evolution (DE) [18], Grey Wolf Optimizer (GWO) [19], and Whale Optimization Algorithm (WOA) [20]. Notably, Pramanik et al. [21] proposed a hybrid algorithm combining GA and GWO to select and reduce feature dimensions. Dhal et al. [22] proposed a hybrid PSO and GWO algorithm, named BGWOPSO, for binary feature selection. Abinayaa [23] proposed a hybrid optimization framework that integrates GWO with DE to address the high-dimensionality and redundancy issues in EEG signal feature selection. These algorithms have achieved significant success in industrial applications.
The Black-winged Kite Algorithm (BKA) [24] is a recently proposed bio-inspired metaheuristic optimization algorithm that achieves global optimization search by simulating the unique hovering hunting behavior and seasonal migration patterns of black-winged kites [25]. BKA abstracts the attacking behavior of black-winged kites into a local fine search mechanism and transforms their migration behavior into a global exploration strategy, theoretically possessing good exploration-exploitation balance capabilities [26]. BKA has demonstrated competitive advantages in handling standard benchmark optimization problems, laying the foundation for its application in fault detection and engineering optimization. However, in complex high-dimensional fault detection scenarios, the original BKA is prone to local optima traps [27], leading to significantly reduced detection accuracy and reliability when dealing with similar fault patterns.
To address the aforementioned problems, this paper proposes an Improved Black-winged Kite Algorithm (IBKA). The algorithm significantly enhances the optimization performance of the original BKA by integrating stagnation-triggered diversification and adaptive weak guidance mechanisms. Specifically, the main technical contributions of this research include:
  • Development of an improved BKA algorithm that integrates stagnation-triggered diversification and adaptive weak guidance mechanisms. The former employs an intelligent stagnation detection strategy to monitor the algorithm’s convergence status in real-time, systematically applying mild random perturbations to the worst-performing individuals to enhance population diversity when consecutive stagnation is detected. The latter adopts a conditionally triggered elite guidance strategy to provide subtle directional guidance from the optimal solution to poorly performing individuals in the later stages of the algorithm. The synergistic effect of these dual mechanisms effectively resolves local optima trap problems while significantly improving convergence efficiency, all while maintaining the algorithm’s exploration capabilities.
  • Comprehensive performance evaluation of IBKA and bIBKA regarding statistical significance and convergence speed. Through comparison with 9 state-of-the-art algorithms (including PSO, GWO, RFO, etc.), the superior performance of the proposed algorithms is validated.
  • Industrial fault detection application where bIBKA is applied to feature selection in fault detection, pioneering the application of bIBKA in the fault detection domain.
The organization of this paper is as follows: Section 2 presents the optimization process of the BKA algorithm. Section 3 provides detailed descriptions of the proposed IBKA and bIBKA. Section 4 presents the results and analysis of IBKA and bIBKA performance on test functions and fault detection datasets.

2. Black-Winged Kite Algorithm

The BKA represents a novel nature-inspired metaheuristic optimization framework that emulates the sophisticated foraging and migratory behaviors exhibited by black-winged kites in their natural habitat [28]. The various biological behaviors of the Black-winged Kite, which form the core of the BKA, are illustrated in Figure 1. These small diurnal raptors, belonging to the family Accipitridae, demonstrate a distinctive hunting strategy characterized by sustained hovering behavior, wherein they maintain stationary flight positions at optimal altitudes while conducting aerial surveillance of the terrain below. Upon prey identification, they execute rapid vertical descents with remarkable precision to capture their quarry, demonstrating exceptional spatial awareness and timing accuracy. Beyond their hunting prowess, black-winged kites exhibit complex seasonal migration patterns that reflect their adaptive response to environmental variability and resource availability. These migratory behaviors encompass long-distance movements between breeding and wintering territories, with individuals demonstrating remarkable navigational capabilities and route optimization strategies. The migration process involves multiple decision-making stages, including departure timing, route selection, and stopover site utilization, all contributing to overall fitness and survival. The BKA algorithm incorporates these biological mechanisms through a dual-phase optimization approach: a hovering phase that simulates aerial surveillance behavior, enabling comprehensive exploration of the solution space, and a diving phase that mimics precision strike mechanisms, facilitating intensive exploitation of promising regions. Additionally, the algorithm integrates migration-inspired mechanisms for population diversity maintenance and global search capabilities, ensuring robust performance across diverse optimization landscapes. This bio-inspired approach leverages the evolutionary advantages of these avian behaviors, providing an effective balance between exploration and exploitation in complex optimization problems.

2.1. Population Initialization

The BKA algorithm starts by randomly generating a population of N candidate solutions and distributing them throughout the search space:
X i ( 0 ) = l b + rand ( 1 , D ) × ( u b l b ) , i = 1 , 2 , , N
where X i ( 0 ) represents the initial position of the i-th kite, D is the problem dimension, l b and u b are the lower and upper bounds of the search space, respectively, and  rand ( 1 , D ) generates a random vector of size D.

2.2. Attacking Behavior

The attacking behavior simulates the precision hunting strategy [29] of black-winged kites. During this phase, each kite performs local search around its current position. The mathematical formulation is:
n ( t ) = 0.05 × exp 2 t T 2
where t is the current iteration, T is the maximum number of iterations, and  n ( t ) is a time-varying parameter that controls the intensity of local search.
The position update mechanism for attacking behavior follows two strategies based on a probability parameter p = 0.9 :
Strategy 1 (with probability p):
X i new = X i ( t ) + n ( t ) × ( 1 + sin ( r ) ) × X i ( t )
Strategy 2 (with probability 1 p ):
X i new = X i ( t ) × ( n ( t ) × ( 2 × rand ( 1 , D ) 1 ) + 1 )
where r is a random number, and  rand ( 1 , D ) generates a random vector.

2.3. Migration Behavior

The migration behavior represents the global exploration capability of the algorithm, inspired by the seasonal movement patterns of black-winged kites. This mechanism enables the algorithm to explore distant regions of the search space and avoid local optima.
The migration factor is calculated as:
m = 2 × sin r + π 2
The Cauchy distribution is employed to generate step sizes for migration:
Cauchy ( x ) = tan ( x 0.5 ) × π
where x U ( 0 , 1 ) is a uniformly distributed random variable.
The position update for migration behavior depends on the fitness comparison:
Case 1: If f ( X i ( t ) ) < f ( X s ( t ) ) (current kite is better than a randomly selected kite):
X i new = X i ( t ) + Cauchy ( D ) × ( X i ( t ) X leader ( t ) )
Case 2: If f ( X i ( t ) ) f ( X s ( t ) ) (current kite is worse than or equal to a randomly selected kite):
X i new = X i ( t ) + Cauchy ( D ) × ( X leader ( t ) m × X i ( t ) )
where X s ( t ) is a randomly selected kite from the population, X leader ( t ) is the best solution found so far, and  Cauchy ( D ) represents the Cauchy-distributed random vector of dimension D.

2.4. Selection Mechanism

A greedy selection mechanism is applied whenever the position is updated, either after an attacking or a migration step:
X i ( t + 1 ) = X i new if f ( X i new ) < f ( X i ( t ) ) X i ( t ) otherwise
where f ( · ) represents the objective function to be minimized.

3. Improved Optimization Algorithm for Black-Winged Kite Algorithm

In this section, we propose an improved Black-winged Kite Algorithm based on the original BKA to solve global optimization problems as shown in Figure 2. Building upon this, we introduce a binary version, named bIBKA, specifically designed for feature selection. Finally, bIBKA is combined with the K-Nearest Neighbor (KNN) [30] classifier and applied to the task of fault detection.

3.1. Enhancement Mechanisms

3.1.1. Stagnation-Triggered Diversification Mechanism

The first enhancement mechanism addresses the premature convergence issue by implementing an intelligent stagnation detection and diversification strategy [31]. This mechanism monitors the algorithm’s convergence status [32] and applies targeted perturbations when stagnation is detected.
Stagnation Detection: The algorithm monitors the improvement of the best fitness value over consecutive iterations. Stagnation is identified when this improvement remains below a predefined threshold for a specified duration:
Stagnation = True if | f best ( t ) f best ( t 1 ) | < ϵ for τ consecutive iterations False otherwise
where f best ( t ) is the best fitness at iteration t, ϵ = 1 × 10 15 is the convergence threshold, and  τ = 30 is the stagnation threshold.
Diversification Strategy: When stagnation is detected and the algorithm has progressed beyond the initial exploration phase ( t > T / 3 ), the mechanism selectively perturbs the worst-performing individuals to enhance population diversity:
N div = max ( 1 , 0.1 × N )
P j = α × ( ub lb ) × ( 2 r 1 )
X j new = X j ( t ) + P j , j W
where N div is the number of individuals to be diversified, α = 0.02 is the perturbation strength factor, r is a random vector, and  W represents the set of worst-performing individuals.

3.1.2. Adaptive Weak Guidance Mechanism

The second enhancement mechanism improves convergence efficiency through a conditionally triggered elite guidance strategy that provides subtle directional guidance to underperforming individuals during the late optimization phase.
Activation Conditions: The adaptive weak guidance mechanism is activated only when all of the following conditions are met:
Guidance Active = True if i > 0.85 N and R < 0.03 and t > 0.6 T False otherwise
where i is the individual index (sorted by fitness), R U ( 0 , 1 ) is a random probability, and the conditions ensure that guidance is applied only to the worst 15% of individuals with a 3% probability during the final 40% of iterations.
Weak Guidance Update: When the guidance mechanism is activated, the position update incorporates a weak directional bias toward the best solution:
D weak = X leader ( t ) X i ( t )
X i new = X i ( t ) + n ( t ) × β × D weak
where D weak is the direction vector toward the leader, and  β = 0.1 is the weak guidance strength factor that ensures the guidance remains subtle and non-disruptive.
A critical design challenge in integrating two enhancement mechanisms is ensuring they do not interfere with each other. Specifically, the potential risk exists that the weak guidance mechanism could prematurely override or undermine the diversification mechanism’s efforts. This conflict is theoretically eliminated through a multi-dimensional complementary design strategy. The mechanisms operate in distinct temporal phases to minimize interference. The Stagnation-Triggered Diversification Mechanism activates during mid-to-late optimization stages ( t > T / 3 ) upon detecting convergence stagnation, with immediate counter reset after intervention. In contrast, the Adaptive Weak Guidance Mechanism engages exclusively during the final refinement phase ( t > 0.6 T ) with ultra-low activation probability (3%). This temporal separation ensures that diversification interventions occur before guidance refinement, with negligible overlap between activation windows. The mechanisms target different subsets of the population to reduce simultaneous intervention risk. Diversification focuses on the worst-performing 10% of individuals, while weak guidance applies to the bottom 15% ( i > 0.85 N ) with stringent probabilistic control. While there exists theoretical overlap between these target ranges, the ultra-low probability of guidance activation ensures minimal simultaneous intervention. The mechanisms employ carefully calibrated intensity levels to prevent interference. Diversification uses mild perturbation strength to escape local optima without disrupting global search patterns. Weak guidance utilizes an even weaker directional bias, ensuring that even in rare concurrent activation scenarios, guidance cannot override or neutralize diversification effects. The substantial difference in intensity levels provides a safety margin against interference. The combination of temporal separation, spatial differentiation, and intensity calibration creates a hierarchical safeguard system that theoretically eliminates the risk of mechanism conflict.

3.2. Detailed Procedure of the IBKA

The complete IBKA algorithm integrates both enhancement mechanisms into the original BKA framework. Algorithm 1 presents the detailed pseudocode.
Algorithm 1 The pseudocode of the IBKA
Require: Population size N, Maximum iterations T, Problem dimension D, Bounds [ lb , ub ]
Ensure: Best solution X best and fitness f best
  1:
Initialize population X i randomly using Equation (1)
  2:
Evaluate fitness of all individuals
  3:
Initialize stagnation counter: stag _ counter = 0
  4:
Initialize last best fitness: f last =
  5:
Set stagnation threshold: τ = 30
  6:
t = 1
  7:
while  t T  do
  8:
    Sort population by fitness and identify leader X leader
  9:
    Update stagnation detection using Equation (10)
10:
    for each individual i = 1 to N do
11:
          Calculate n ( t ) using Equation (2)
12:
          // Attacking Behavior with Adaptive Weak Guidance
13:
          if  p < r  then
14:
               Apply Strategy 1 using Equation (3)
15:
          else
16:
               if Guidance conditions met (Equation (14)) then
17:
                    Apply weak guidance using Equations (15) and (16)
18:
               else
19:
                    Apply Strategy 2 using Equation (4)
20:
               end if
21:
          end if
22:
          Apply boundary handling and selection mechanism
23:
          // Migration Behavior (unchanged)
24:
          Apply migration behavior using Equations (7) or (8)
25:
          Apply boundary handling and selection mechanism
26:
    end for
27:
    // Stagnation-Triggered Diversification
28:
    if  stag _ counter > τ  and  t > T / 3  then
29:
          Calculate N div using Equation (11)
30:
          Identify worst individuals W
31:
          for each j W  do
32:
                Generate perturbation using Equation (12)
33:
                Update position using Equation (13)
34:
                Apply boundary handling and re-evaluate fitness
35:
          end for
36:
          Reset stagnation counter: stag _ counter = 0
37:
    end if
38:
    Update best solution if improved
39:
     t = t + 1
40:
end while
41:
return  X best and f best

3.3. bIBKA for Feature Selection

Given the inherent binary nature of feature selection problems, a discrete variant of the IBKA, termed bIBKA, was developed. This adaptation utilizes a transfer function to map the algorithm’s continuous search solutions into a discrete, binary feature space.

3.3.1. Binary Population Initialization

Unlike continuous optimization, bIBKA initializes the population directly in binary space:
X i , d ( 0 ) = 1 if R i , d > 0.5 0 otherwise
where R i , d U ( 0 , 1 ) is a random number for the d-th dimension of the i-th individual.

3.3.2. Multi-Modal Transfer Functions

bIBKA employs three different transfer functions depending on the operation context:
S-shaped Transfer Function:
T S ( x ) = 1 1 + exp ( 2 x )
V-shaped Transfer Function:
T V ( x ) = erf π 2 x
U-shaped Transfer Function:
T U ( x ) = | sin ( 2 x ) |

3.3.3. Binary Position Update Mechanisms

The binary position update varies according to the behavioral phase:
Enhanced Attacking Behavior: For attacking behavior with probability p:
x cont = X i , d ( t ) + n ( t ) × ( 1 + sin ( r ) ) × X i , d ( t )
X i , d new = 1 if R < T V ( x cont ) 0 otherwise
Adaptive Weak Guidance (Binary Version): When guidance conditions are met:
d weak = X leader , d ( t ) X i , d ( t )
x cont = X i , d ( t ) + 0.1 × n ( t ) × d weak
X i , d new = 1 if R < T V ( x cont ) 0 otherwise
Standard Binary Conversion: For other operations:
x cont = X i , d ( t ) × ( n ( t ) × ( 2 R 1 ) + 1 )
X i , d new = 1 if R < T S ( x cont ) 0 otherwise

3.3.4. Binary Migration Behavior

The migration behavior is adapted for binary space using Cauchy distribution:
Cauchy binary = tan ( ( R 0.5 ) × π ) × 0.1
x migration = X s , d ( t ) + Cauchy binary
X i , d new = 1 if R < T V ( x migration ) 0 otherwise

3.3.5. Binary Stagnation-Triggered Diversification

The diversification mechanism in bIBKA uses bit-flip operations:
X j , d new = 1 X j , d ( t ) if R < ρ flip X j , d ( t ) otherwise
where ρ flip = 0.02 is the bit-flip probability, and j W represents the worst-performing individuals.

3.3.6. Feature Selection Fitness Function

For feature selection problems, the fitness function combines classification accuracy and feature subset size to achieve an optimal balance between performance and complexity:
f ( X i binary ) = α × ( 1 Accuracy ) + ( 1 α ) × | X i binary | D
where α [ 0 , 1 ] is a weighting parameter that balances classification accuracy and feature reduction, Accuracy is the classification accuracy obtained using the selected feature subset, | X i binary | represents the number of selected features (Hamming weight), and D is the total number of available features.
The fitness function is minimized, so ( 1 Accuracy ) represents the classification error rate. The second term | X i binary | D represents the ratio of selected features to total features, encouraging feature reduction. In this work, α = 0.99 is used for all experiments, prioritizing classification accuracy since missing fault instances can result in severe safety consequences, while the small weight ( 1 α = 0.01 ) on feature reduction encourages compact subsets when solutions achieve comparable accuracy.

3.4. Computational Complexity Analysis

The computational complexity of IBKA remains O ( T × N × D ) , identical to the original BKA. The additional operations introduced by the enhancement mechanisms have negligible computational overhead:
  • Stagnation detection: O ( 1 ) per iteration
  • Diversification: O ( 0.1 N × D ) when triggered
  • Adaptive weak guidance: O ( D ) for eligible individuals

4. Algorithm Performance Test and Analysis

4.1. Evaluation Criteria

The effectiveness of the proposed method is quantified using several evaluation metrics. We report the average fitness (Avg) [33], optimal fitness (Best) [34], F1-score, accuracy [35], and the standard deviation (Std) [36] to provide a thorough analysis. The formulas for these metrics can be found in Equations (33)–(37).
Avg = 1 N i = 1 N f i t n e s s i
Best = M a x ( f i t n e s s i )
Std = 1 M 1 i = 1 M f i t n e s s i Avg 2
F1-score = 2 · T P 2 · T P + F P + F N
Accuracy = T P + T N T P + T N + F P + F N
where N is the number of independent runs, M is the total number of runs for standard deviation calculation, f i t n e s s i represents the fitness value of the i-th run, T P , T N , F P , and F N represent true positive, true negative, false positive, and false negative, respectively.

4.2. Experimental Settings

All algorithms were executed on a Windows 11 platform equipped with an Intel Core i5-12600K processor (3.7 GHz) and 32 GB of RAM. In Experiment I, all algorithms are configured with a population size of 30 and executed for 500 iterations in a 30-dimensional search space. For Experiment II, the population size is reduced to 10 and the number of iterations to 100. To ensure the reliability of the results, each algorithm is independently run 20 times per dataset to account for stochastic variation.

4.3. Experiment I: Global Optimization of IBKA Using 26 Benchmark Mathematical Functions

4.3.1. Benchmark Functions

A set of 26 standard benchmark functions is employed to thoroughly evaluate the proposed algorithm’s performance on high-dimensional optimization problems. This suite is designed to test both exploitation and exploration capabilities across various search landscapes. The functions are divided into three categories: unimodal functions (F1–F7) to test convergence rate, multimodal functions (F8–F22) to assess global search and local optima avoidance, and complex composite functions (F23–F26) to evaluate the algorithm’s robustness on intricate landscapes. The complete details for these benchmark functions are summarized in Table 1.

4.3.2. Experiment I: Result and Analysis

The comparative analysis of IBKA against competing algorithms, as detailed in Table 2, Table 3 and Table 4, demonstrates its effectiveness on the benchmark function set. The results show that IBKA consistently achieves favorable performance across three key metrics: Avg, Best, and Std. Regarding the Avg metric, which quantifies the expected optimization performance by computing the mean fitness across all independent runs, IBKA demonstrates superior convergence capability with extremely low average fitness values such as 7.692 × 10 94 on F9 and achieves zero or near-zero values on complex multimodal functions such as F16, F17, F19, and F26, indicating strong exploitation capability and effective global search. For the Best metric, which represents the algorithm’s ultimate optimization capacity, IBKA performs excellently by achieving the theoretical global optimum on functions F16–F19, F24, and F26 and reaching remarkable precision values such as 3.279 × 10 147 on F13 and 1.990 × 10 123 on F14, validating IBKA’s strong exploitation capability and confirming the effectiveness of the Adaptive Weak Guidance Mechanism. Concerning the Std metric, which quantifies solution stability, IBKA exhibits impressive performance with zero standard deviation on multimodal functions F16–F19, F24, and F26, indicating perfect consistency across all 20 runs, and extremely low variance on unimodal functions F1, F3, and F4, validating the effectiveness of the Stagnation-Triggered Diversification Mechanism and confirming that IBKA is highly reliable for practical engineering applications. Table 5 presents a quantitative ranking summary that clearly demonstrates IBKA’s superior performance across all evaluation metrics. IBKA achieves first place on 15 functions for average fitness, 17 for best fitness, and 14 for standard deviation—substantially outperforming all competing algorithms. Critically, IBKA ranks either first or second on 23 out of 26 functions (88.5%) for both average fitness and best fitness, demonstrating exceptional consistency. In stark contrast, the original BKA achieves first place on only 10, 8, and 10 functions respectively, while modern metaheuristics WOA and GWO secure first place on merely 5 and 4 functions for average fitness. Traditional algorithms show particularly poor performance, with GA achieving zero first-place rankings and DE failing to secure any top-three positions across all metrics. These ranking statistics provide compelling quantitative evidence that the proposed stagnation-triggered diversification and adaptive weak guidance mechanisms significantly enhance both solution quality and algorithmic stability, enabling IBKA to consistently outperform state-of-the-art competitors across diverse optimization challenges.
Figure 3 presents boxplot distributions across 26 benchmark functions for performance comparison. IBKA demonstrates consistently lower median fitness values on unimodal functions F1–F4 and multimodal functions F9, F11, F20, indicating superior convergence quality. The narrow interquartile ranges on F3, F4, F16–F19, and F24–F26 confirm minimal variance across 20 runs, while significantly fewer outliers compared to GA and DE on complex functions F10–F15 validate enhanced robustness. Notably, IBKA achieves compressed box distributions on F9, F11, and F20, simultaneously demonstrating solution quality and stability. By contrast, SCA and PSO exhibit wide distributions with scattered outliers, indicating inconsistent performance. These results validate that IBKA maintains superior optimization capability and robustness across diverse problem types.
To rigorously validate the performance improvements, Wilcoxon signed-rank tests [37] were conducted for pairwise comparisons between IBKA and each competitor algorithm, as shown in Table 6. The statistical analysis reveals that IBKA consistently achieves statistically significant superiority across the vast majority of benchmark functions: PKO (25/26), RFO (25/26), GA (24/26), DE (25/26), PSO (24/26), SCA (25/26), GWO (25/26), and WOA (20/26). Remarkably, IBKA demonstrates over 90% superiority coverage for most state-of-the-art metaheuristic algorithms, with the majority exhibiting extremely strong statistical significance (p < 0.005). Even compared to its parent algorithm BKA, IBKA exhibits statistical significance on 15/26 functions, demonstrating substantial improvements over the baseline approach. These statistical tests provide overwhelming evidence that the observed performance gains are not attributable to random variation, but rather represent fundamental algorithmic enhancements introduced by the proposed mechanisms. The consistently high proportion of statistically significant results across diverse optimization landscapes—including unimodal, multimodal, and composite functions—confirms IBKA’s robust and generalizable superiority over existing state-of-the-art algorithms.

4.3.3. Convergence Analysis

To evaluate the convergence performance of the proposed algorithm, Figure 4 presents the convergence curves showing the best-so-far fitness values across iterations for all competing algorithms on 26 benchmark functions. The convergence patterns reveal distinct performance characteristics across different function types.
On unimodal functions, IBKA demonstrates rapid initial convergence followed by sustained refinement. For instance, on F1, IBKA’s curve exhibits steep descent in early iterations, reaching near-optimal regions significantly faster than competing algorithms, and maintaining stable convergence thereafter with minimal fluctuation. Similarly, on F3, IBKA’s curve is positioned substantially lower than BKA’s curve throughout the optimization process, demonstrating IBKA’s superior exploitation capability. In contrast, traditional algorithms such as GA and DE show premature stagnation on these functions, with their curves flattening at considerably higher fitness levels. On multimodal functions, IBKA exhibits robust global search capability. On F16, IBKA’s convergence curve descends sharply and reaches the bottom of the fitness scale, while competitors like PSO and SCA exhibit fluctuating curves that stabilize at visibly higher fitness levels. The smooth, monotonic descent of IBKA’s curve on F18 contrasts sharply with the irregular, oscillating patterns observed in GWO and RFO, indicating IBKA’s effective balance between exploration and exploitation. Notably, on F20, IBKA’s curve shows consistently steeper descent and lower final positioning compared to WOA, which exhibits earlier stagnation. On complex composite functions, IBKA maintains steady convergence without premature stagnation. For F25, IBKA’s curve demonstrates continuous improvement throughout all iterations, while competitors either stagnate early or converge to visibly inferior solutions. The convergence curves collectively validate that IBKA achieves faster convergence speed, superior final solution quality, and enhanced stability across diverse optimization landscapes.

4.4. Experiment II: Fault Detection Using bIBKA

In this experiment, we evaluated the fault detection performance of bIBK algorithm using six diverse fault datasets from different domains.The other compared algorithms are binarized as shown in the Table 7. The datasets comprise: a semiconductor manufacturing fault dataset (DS1, SECOM dataset from UCI Machine Learning Repository), a mechanical system fault dataset (DS2, from Heywhale public dataset repository), and four software fault datasets (DS3-DS6, from NASA Metrics Data Program repository) from different software projects. This heterogeneous collection of datasets enables a comprehensive assessment of bIBK’s fault detection capabilities across various industrial scenarios. Table 8 presents the detailed characteristics of these datasets.

4.4.1. Experiment II: Result and Analysis

Table 9, Table 10, and Table 11 present the experimental results in terms of Avg, Accuracy, and F1-score, respectively. The results demonstrate that bIBKA exhibits superior performance compared to other baseline algorithms across most datasets, achieving notable improvements in all three metrics across semiconductor manufacturing, mechanical system, and software fault detection scenarios. Regarding the Avg metric, which quantifies the feature selection quality by combining classification error rate and feature ratio as defined in Equation (32), bIBKA performs excellently with particularly impressive results on the challenging high-dimensional semiconductor dataset DS1 and achieves consistently low values on DS2, DS3, and DS6, demonstrating superior capability in identifying compact yet informative feature subsets. For the Accuracy metric, which measures the proportion of correctly classified instances, bIBKA achieves strong classification accuracy with 72.10% on DS1 substantially surpassing bBKA, 88.90% on DS2 representing a 6.5% improvement over bBKA, 91.40% on DS3, and 86.90% on DS4, validating that bIBKA’s binary adaptation mechanisms effectively preserve the algorithm’s optimization capability in discrete search spaces. Concerning the F1-score, which provides a balanced evaluation of precision and recall particularly important for fault detection, bIBKA demonstrates excellent performance with substantial improvements over bBKA such as 0.8820 on DS2, 0.6610 on DS3, and 0.6510 on DS4, effectively addressing the precision-recall trade-off and reducing both false alarms and missed detections.

4.4.2. Convergence Analysis

Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 present the convergence curves of bIBKA and competing algorithms across six fault detection datasets, tracking fitness evolution over iterations. The curves reveal dataset-specific convergence characteristics and algorithm performance differences.
On the high-dimensional semiconductor dataset DS1, bIBKA’s convergence curve exhibits rapid initial descent in early iterations, followed by gradual refinement toward the end of optimization. In comparison, bBKA’s curve is positioned noticeably higher throughout the process, while bPKO shows similar initial descent but settles at a slightly higher final level. The steeper initial slope and lower final positioning of bIBKA’s curve demonstrate superior feature selection capability in high-dimensional spaces. On DS2, bIBKA achieves the fastest convergence among all algorithms, with its curve reaching the lower fitness region earlier than competitors, whereas bSCA shows slower descent and stabilizes at a visibly higher level. The smooth, monotonic curves of bIBKA on DS3 and DS6 contrast sharply with the fluctuating, irregular patterns observed in bGA and bDE, indicating stable optimization without oscillation between local optima.
Notably, on DS4, bIBKA’s curve shows continuous improvement throughout all iterations, while bPSO and bWOA exhibit early stagnation with their curves flattening around mid-iterations. On DS5, although bPKO and bRFO’s curves are positioned slightly lower in the final region, bIBKA’s convergence curve demonstrates more consistent, steady descent with smaller iteration-to-iteration variation, suggesting better algorithmic stability. The convergence patterns across all datasets validate that bIBKA effectively balances rapid initial convergence with sustained refinement, achieving both computational efficiency and solution quality in practical fault detection scenarios.

5. Conclusions

This paper presents an improved Black-winged Kite Algorithm and its binary variant for addressing global optimization and fault detection challenges. The proposed Stagnation-Triggered Diversification Mechanism effectively prevents premature convergence by monitoring the population’s convergence state and applying targeted perturbations to stagnant individuals. The Adaptive Weak Guidance Mechanism further enhances optimization efficiency through a conditional elite guidance strategy, which is particularly beneficial during the late optimization phase. Comprehensive evaluations on 26 benchmark functions demonstrate IBKA’s superior performance in solution quality and convergence speed. Furthermore, experimental results across semiconductor manufacturing, mechanical system, and software fault detection scenarios validate that bIBKA achieves significant improvements in detection accuracy and reliability compared to existing algorithms while maintaining computational efficiency.
While these results validate the effectiveness of IBKA under the tested conditions, several important considerations emerge when extending this work to broader industrial contexts. The evaluation on static benchmark functions and fault datasets provides a solid foundation for understanding the algorithm’s core capabilities, though real-world deployments may encounter dynamic fault patterns where the fixed stagnation threshold could benefit from adaptive tuning mechanisms. Additionally, the fault detection datasets inherently exhibit class imbalance characteristic of real-world industrial applications, where normal operations significantly outnumber fault occurrences; in such cases, the reported accuracy values are naturally influenced by the performance of the majority class, while F1-scores provide more balanced indicators of the algorithm’s capability in detecting minority fault classes, which are often the most critical for industrial safety.

Author Contributions

Conceptualization, K.Q. and W.W.; methodology, K.Q., K.W. and R.C.; software, G.L. and J.H.; validation, K.Q. and K.W.; formal analysis, R.C. and G.L.; investigation, K.Q.; resources, W.W.; data curation, J.H.; writing—original draft preparation, K.Q.; writing—review and editing, W.W.; visualization, K.W.; supervision, W.W.; project administration, K.Q. and W.W.; funding acquisition, W.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data that support the findings of this study are available from the corresponding authors upon reasonable request.

Conflicts of Interest

The authors declare that they have no competing interests.

References

  1. Kim, S.; Lee, J.; Kang, C. Analysis of industrial accidents causing through jamming or crushing accidental deaths in the manufacturing industry in South Korea: Focus on non-routine work on machinery. Saf. Sci. 2021, 133, 104998. [Google Scholar] [CrossRef]
  2. Abid, A.; Khan, M.T.; Iqbal, J. A review on fault detection and diagnosis techniques: Basics and beyond. Artif. Intell. Rev. 2021, 54, 3639–3664. [Google Scholar] [CrossRef]
  3. Deng, W.; Xu, J.; Song, Y.; Zhao, H. Differential evolution algorithm with wavelet basis function and optimal mutation strategy for complex optimization problem. Appl. Soft Comput. 2021, 100, 106724. [Google Scholar] [CrossRef]
  4. Kalyani, G.; Janakiramaiah, B.; Karuna, A.; Prasad, L.N. Diabetic retinopathy detection and classification using capsule networks. Complex Intell. Syst. 2023, 9, 2651–2664. [Google Scholar] [CrossRef]
  5. Abdel-Basset, M.; Mohamed, R.; Jameel, M.; Abouhawwash, M. Nutcracker optimizer: A novel nature-inspired metaheuristic algorithm for global optimization and engineering design problems. Knowl.-Based Syst. 2023, 262, 110248. [Google Scholar] [CrossRef]
  6. Ding, N.; Qin, Y.; Yang, G.; Wei, F.; Yang, Z.; Su, Y.; Hu, S.; Chen, Y.; Chan, C.M.; Chen, W.; et al. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nat. Mach. Intell. 2023, 5, 220–235. [Google Scholar] [CrossRef]
  7. Li, Q.; Ke, S.; Cao, J.; Wang, W.; Cai, T.; Zhang, X. Nonlinear dynamic response analysis of offshore new floating pipe-type flexible photovoltaic arrays under wind-wave coupling effects. Ocean. Eng. 2025, 340, 122222. [Google Scholar] [CrossRef]
  8. Akinola, O.O.; Ezugwu, A.E.; Agushaka, J.O.; Zitar, R.A.; Abualigah, L. Multiclass feature selection with metaheuristic optimization algorithms: A review. Neural Comput. Appl. 2022, 34, 19751–19790. [Google Scholar] [CrossRef] [PubMed]
  9. Tang, J.; Liu, G.; Pan, Q. A review on representative swarm intelligence algorithms for solving optimization problems: Applications and trends. IEEE/CAA J. Autom. Sin. 2021, 8, 1627–1643. [Google Scholar] [CrossRef]
  10. Rabanus-Wallace, M.T.; Hackauf, B.; Mascher, M.; Lux, T.; Wicker, T.; Gundlach, H.; Baez, M.; Houben, A.; Mayer, K.F.; Guo, L.; et al. Chromosome-scale genome assembly provides insights into rye biology, evolution and agronomic potential. Nat. Genet. 2021, 53, 564–573. [Google Scholar] [CrossRef] [PubMed]
  11. Cai, H.; Wang, H.; Bei, Z.; Zhou, D.; Gao, H. Biomimetic swarm fission driven algorithm with preassigned target subgroup size. Bioinspiration Biomimetics 2025, 20, 026021. [Google Scholar] [CrossRef] [PubMed]
  12. Tu, Z.; Liu, W.; Wang, J.; Qiu, X.; Huang, J.; Li, J.; Lou, H. Biomimetic high performance artificial muscle built on sacrificial coordination network and mechanical training process. Nat. Commun. 2021, 12, 2916. [Google Scholar] [CrossRef] [PubMed]
  13. Bouaouda, A.; Hashim, F.A.; Sayouti, Y.; Hussien, A.G. Pied kingfisher optimizer: A new bio-inspired algorithm for solving numerical optimization and industrial engineering problems. Neural Comput. Appl. 2024, 36, 15455–15513. [Google Scholar] [CrossRef]
  14. Braik, M.; Al-Hiary, H. Rüppell’s fox optimizer: A novel meta-heuristic approach for solving global optimization problems. Clust. Comput. 2025, 28, 1–77. [Google Scholar] [CrossRef]
  15. Abualigah, L.; Diabat, A. Advances in sine cosine algorithm: A comprehensive survey. Artif. Intell. Rev. 2021, 54, 2567–2608. [Google Scholar] [CrossRef]
  16. Zhou, J.; Hua, Z. A correlation guided genetic algorithm and its application to feature selection. Appl. Soft Comput. 2022, 123, 108964. [Google Scholar] [CrossRef]
  17. Gad, A.G. Particle swarm optimization algorithm and its applications: A systematic review. Arch. Comput. Methods Eng. 2022, 29, 2531–2561. [Google Scholar] [CrossRef]
  18. Ghosh, A.; Jana, N.D.; Mallik, S.; Zhao, Z. Designing optimal convolutional neural network architecture using differential evolution algorithm. Patterns 2022, 3, 100567. [Google Scholar] [CrossRef]
  19. Liu, Y.; As’arry, A.; Hassan, M.K.; Hairuddin, A.A.; Mohamad, H. Review of the grey wolf optimization algorithm: Variants and applications. Neural Comput. Appl. 2024, 36, 2713–2735. [Google Scholar] [CrossRef]
  20. Nadimi-Shahraki, M.H.; Zamani, H.; Asghari Varzaneh, Z.; Mirjalili, S. A systematic review of the whale optimization algorithm: Theoretical foundation, improvements, and hybridizations. Arch. Comput. Methods Eng. 2023, 30, 4113–4159. [Google Scholar] [CrossRef] [PubMed]
  21. Pramanik, R.; Pramanik, P.; Sarkar, R. Breast cancer detection in thermograms using a hybrid of GA and GWO based deep feature selection method. Expert Syst. Appl. 2023, 219, 119643. [Google Scholar] [CrossRef]
  22. Dhal, P.; Azad, C. A multi-objective feature selection method using Newton’s law based PSO with GWO. Appl. Soft Comput. 2021, 107, 107394. [Google Scholar] [CrossRef]
  23. Abinayaa, S.; Sridhar, S.S. An Enhanced Bio-inspired GWO–DE Technique for Efficient Feature Selection in the EEG-RSVP Paradigm. Int. J. Comput. Intell. Syst. 2025, 18, 1–33. [Google Scholar] [CrossRef]
  24. Wang, J.; Wang, W.C.; Hu, X.X.; Qiu, L.; Zang, H.F. Black-winged kite algorithm: A nature-inspired meta-heuristic for solving benchmark functions and engineering problems. Artif. Intell. Rev. 2024, 57, 98. [Google Scholar] [CrossRef]
  25. Duman, S.; Kahraman, H.T.; Sonmez, Y.; Guvenc, U.; Kati, M.; Aras, S. A powerful meta-heuristic search algorithm for solving global optimization and real-world solar photovoltaic parameter estimation problems. Eng. Appl. Artif. Intell. 2022, 111, 104763. [Google Scholar] [CrossRef]
  26. Kantaros, A.; Ganetsos, T.; Pallis, E.; Papoutsidakis, M. Biomimetic Additive Manufacturing: Engineering Complexity Inspired by Nature’s Simplicity. Biomimetics 2025, 10, 453. [Google Scholar] [CrossRef] [PubMed]
  27. Fu, Z.; Li, Z.; Li, Y.; Chen, H. MICFOA: A Novel Improved Catch Fish Optimization Algorithm with Multi-Strategy for Solving Global Problems. Biomimetics 2024, 9, 509. [Google Scholar] [CrossRef] [PubMed]
  28. Bernard, A.; Perales, F. Is migration a learned behavior? Understanding the impact of past migration on future migration. Popul. Dev. Rev. 2021, 47, 449–474. [Google Scholar] [CrossRef]
  29. Hansen, M.J.; Domenici, P.; Bartashevich, P.; Burns, A.; Krause, J. Mechanisms of group-hunting in vertebrates. Biol. Rev. 2023, 98, 1687–1711. [Google Scholar] [CrossRef]
  30. Zhang, S. Challenges in KNN classification. IEEE Trans. Knowl. Data Eng. 2021, 34, 4663–4675. [Google Scholar] [CrossRef]
  31. Palomino-Segura, M.; Sicilia, J.; Ballesteros, I.; Hidalgo, A. Strategies of neutrophil diversification. Nat. Immunol. 2023, 24, 575–584. [Google Scholar] [CrossRef] [PubMed]
  32. Lau, E.S.; Oakley, T.H. Multi-level convergence of complex traits and the evolution of bioluminescence. Biol. Rev. 2021, 96, 673–691. [Google Scholar] [CrossRef]
  33. Schirner, M.; Deco, G.; Ritter, P. Learning how network structure shapes decision-making for bio-inspired computing. Nat. Commun. 2023, 14, 2963. [Google Scholar] [CrossRef]
  34. Brown, J.J.; Pascual, M.; Wimberly, M.C.; Johnson, L.R.; Murdock, C.C. Humidity–The overlooked variable in the thermal biology of mosquito-borne disease. Ecol. Lett. 2023, 26, 1029–1049. [Google Scholar] [CrossRef]
  35. Lekshmy, S.; Sridhar, K.P. Analyzing the performance of a bio-sensor integrated improved blended learning model for accurate pneumonia prediction. Results Eng. 2024, 22, 102063. [Google Scholar] [CrossRef]
  36. Kadlec, D.; Sainani, K.L.; Nimphius, S. With great power comes great responsibility: Common errors in meta-analyses and meta-regressions in strength & conditioning research. Sports Med. 2023, 53, 313–325. [Google Scholar] [PubMed]
  37. Zhao, S.; Zhang, T.; Cai, L.; Yang, R. Triangulation topology aggregation optimizer: A novel mathematics-based meta-heuristic algorithm for continuous optimization and engineering applications. Expert Syst. Appl. 2024, 238, 121744. [Google Scholar] [CrossRef]
Figure 1. Various Biological Behaviors of the Black-winged Kite.
Figure 1. Various Biological Behaviors of the Black-winged Kite.
Biomimetics 10 00728 g001
Figure 2. Flowchart of the IBKA.
Figure 2. Flowchart of the IBKA.
Biomimetics 10 00728 g002
Figure 3. Boxplot comparison of optimization algorithms on 26 benchmark test functions.
Figure 3. Boxplot comparison of optimization algorithms on 26 benchmark test functions.
Biomimetics 10 00728 g003
Figure 4. Convergence curves on each benchmark function.
Figure 4. Convergence curves on each benchmark function.
Biomimetics 10 00728 g004aBiomimetics 10 00728 g004bBiomimetics 10 00728 g004c
Figure 5. Convergence curves on DS1.
Figure 5. Convergence curves on DS1.
Biomimetics 10 00728 g005
Figure 6. Convergence curves on DS2.
Figure 6. Convergence curves on DS2.
Biomimetics 10 00728 g006
Figure 7. Convergence curves on DS3.
Figure 7. Convergence curves on DS3.
Biomimetics 10 00728 g007
Figure 8. Convergence curves on DS4.
Figure 8. Convergence curves on DS4.
Biomimetics 10 00728 g008
Figure 9. Convergence curves on DS5.
Figure 9. Convergence curves on DS5.
Biomimetics 10 00728 g009
Figure 10. Convergence curves on DS6.
Figure 10. Convergence curves on DS6.
Biomimetics 10 00728 g010
Table 1. The benchmark Functions.
Table 1. The benchmark Functions.
Function NameFormulaRange f min
F1 (Sphere) f 1 ( x ) = i = 1 d i m x i 2 [ 100 , 100 ] 0
F2 (Schwefel 2.22) f 2 ( x ) = i = 1 d i m | x i | + i = 1 d i m | x i | [ 10 , 10 ] 0
F3 (Schwefel 1.2) f 3 ( x ) = i = 1 d i m ( j = 1 i x j ) 2 [ 10 , 10 ] 0
F4 (Schwefel 2.21) f 4 ( x ) = max i { | x i | } [ 10 , 10 ] 0
F5 (Step) f 5 ( x ) = i = 1 d i m | x i + 0.5 | 2 [ 10 , 10 ] 0
F6 (Quartic) f 6 ( x ) = i = 1 d i m i x i 4 + r a n d [ 1.28 , 1.28 ] 0
F7 (Exponential) f 7 ( x ) = exp ( 0.5 i = 1 d i m x i ) [ 10 , 10 ] 0
F8 (SumPower) f 8 ( x ) = i = 1 d i m | x i | i + 1 [ 1 , 1 ] 0
F9 (SumSquare) f 9 ( x ) = i = 1 d i m i x i 2 [ 10 , 10 ] 0
F10 (Rosenbrock) f 10 ( x ) = i = 1 d i m 1 [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] [ 10 , 10 ] 0
F11 (Zakharov) f 11 ( x ) = i = 1 d i m x i 2 + ( D 2 i = 1 d i m x i 2 ) + ( D 2 i = 1 d i m x i 4 ) [ 5.12 , 5.12 ] 0
F12 (Trid) f 12 ( x ) = ( x 1 1 ) 2 + i = 2 d i m i ( 2 x i 2 x i 1 ) 2 [ 5 , 5 ] 0
F13 (Elliptic) f 13 ( x ) = i = 1 d i m ( 10 6 ) i 1 d i m 1 x i 2 [ 100 , 100 ] 0
F14 (Cigar) f 14 ( x ) = x 1 2 + 10 6 i = 2 d i m x i 2 [ 100 , 100 ] 0
F15 (Tablet) f 15 ( x ) = 10 6 x 1 2 + i = 2 d i m x i 6 [ 10 , 10 ] 0
F16 (Rastrigin) f 16 ( x ) = i = 1 d i m [ x i 2 10 cos ( 2 π x i ) ] + 10 d i m [ 5.12 , 5.12 ] 0
F17 (NCRastrigin) f 17 ( x ) = i = 1 d i m y i , where y i = x i 2 10 cos ( 2 π x i ) + 10 | x i | < 0.5 ( r o u n d ( 2 x i ) 2 ) 2 10 cos ( 2 π r o u n d ( 2 x i ) 2 ) + 10 o t h e r w i s e [ 5.12 , 5.12 ] 0
F18 (Ackley) f 18 ( x ) = 20 exp ( 0.2 1 d i m i = 1 d i m x i 2 ) exp ( 1 d i m i = 1 d i m cos ( 2 π x i ) ) + 20 + e [ 20 , 20 ] 0
F19 (Griewank) f 19 ( x ) = i = 1 d i m x i 2 4000 i = 1 d i m cos ( x i i ) + 1 [ 600 , 600 ] 0
F20 (Alpine) f 20 ( x ) = i = 1 d i m | x i sin ( x i ) + 0.1 x i | [ 10 , 10 ] 0
F21 (Penalized1) f 21 ( x ) = π d i m [ 10 sin 2 ( π ( 1 + x 1 + 1 4 ) ) + i = 1 d i m 1 ( ( x i + 1 4 ) 2 ( 1 + 10 sin 2 ( π ( 1 + x i + 1 + 1 4 ) ) ) ) + ( x d i m + 1 4 ) 2 ] + i = 1 d i m u ( x i , 10 , 100 , 4 ) [ 10 , 10 ] 0
F22 (Penalized2) f 22 ( x ) = 0.1 [ sin 2 ( 3 π x 1 ) + i = 1 d i m 1 ( x i 1 ) 2 ( 1 + sin 2 ( 3 π x i + 1 ) ) + ( x d i m 1 ) 2 ( 1 + sin 2 ( 2 π x d i m ) ) ] + i = 1 d i m u ( x i , 5 , 100 , 4 ) [ 5 , 5 ] 0
F23 (Levy) f 23 ( x ) = i = 1 d i m 1 ( x i 1 ) 2 ( 1 + sin 2 ( 3 π x i + 1 ) ) + sin 2 ( 3 π x 1 ) + ( x d i m 1 ) 2 ( 1 + sin 2 ( 2 π x d i m ) ) [ 2 , 2 ] 0
F24 (Weierstrass) f 24 ( x ) = i = 1 d i m [ k = 0 k m a x a k cos ( 2 π b k ( x i + 0.5 ) ) ] d i m k = 0 k m a x a k cos ( 2 π b k · 0.5 ) [ 1 , 1 ] 0
F25 (Solomon) f 25 ( x ) = 1 cos ( 2 π i = 1 d i m x i 2 ) + 0.1 i = 1 d i m x i 2 [ 20 , 20 ] 0
F26 (Bohachevsky) f 26 ( x ) = i = 1 d i m 1 [ x i 2 + 2 x i + 1 2 0.3 cos ( 2 π x i ) 0.4 cos ( 4 π x i + 1 ) + 0.7 ] [ 5 , 5 ] 0
Table 2. Comparison of the average fitness of IBKA against other algorithms.
Table 2. Comparison of the average fitness of IBKA against other algorithms.
Func.MetricsIBKABKAPKORFOGAWOAGWOSCADEPSO
F1Avg 6.639 × 10 83 1.146 × 10 90 8.242 × 10 02 1.052 × 10 02 8.460 × 10 + 03 3.778 × 10 73 1.045 × 10 27 4.182 × 10 + 00 3.501 × 10 + 04 3.294 × 10 01
F2Avg 3.417 × 10 38 9.804 × 10 41 1.899 × 10 02 4.522 × 10 01 3.125 × 10 + 01 8.230 × 10 52 1.104 × 10 16 1.756 × 10 02 2.608 × 10 + 05 1.090 × 10 + 00
F3Avg 9.341 × 10 93 5.482 × 10 73 2.197 × 10 + 01 7.722 × 10 01 1.859 × 10 + 02 4.534 × 10 + 02 2.406 × 10 07 8.312 × 10 + 01 7.598 × 10 + 02 1.786 × 10 + 01
F4Avg 1.204 × 10 38 4.041 × 10 46 5.659 × 10 01 5.383 × 10 02 7.171 × 10 + 00 4.797 × 10 + 00 1.327 × 10 07 3.689 × 10 + 00 8.370 × 10 + 00 7.207 × 10 01
F5Avg 1.469 × 10 + 00 1.898 × 10 + 00 7.628 × 10 04 1.060 × 10 02 8.524 × 10 + 01 8.383 × 10 02 6.835 × 10 01 4.996 × 10 + 00 3.401 × 10 + 02 2.451 × 10 03
F6Avg 6.068 × 10 04 2.693 × 10 04 4.907 × 10 02 3.213 × 10 02 2.899 × 10 01 4.543 × 10 03 1.758 × 10 03 1.492 × 10 01 4.150 × 10 + 01 4.049 × 10 02
F7Avg 4.097 × 10 41 1.988 × 10 32 7.175 × 10 66 7.175 × 10 66 2.942 × 10 37 7.175 × 10 66 1.850 × 10 58 1.709 × 10 40 2.889 × 10 41 3.834 × 10 54
F8Avg 6.166 × 10 117 6.657 × 10 112 8.385 × 10 10 3.602 × 10 07 3.109 × 10 05 4.089 × 10 108 1.060 × 10 97 3.951 × 10 05 2.703 × 10 01 1.503 × 10 16
F9Avg 7.692 × 10 94 1.451 × 10 77 6.693 × 10 03 2.183 × 10 01 8.148 × 10 + 02 1.302 × 10 75 2.313 × 10 28 2.602 × 10 + 00 4.170 × 10 + 03 5.472 × 10 02
F10Avg 2.793 × 10 + 01 2.779 × 10 + 01 3.357 × 10 + 01 2.819 × 10 + 01 2.923 × 10 + 03 2.790 × 10 + 01 2.696 × 10 + 01 1.079 × 10 + 03 1.207 × 10 + 06 5.297 × 10 + 01
F11Avg 3.216 × 10 81 6.104 × 10 79 2.692 × 10 02 1.562 × 10 01 1.184 × 10 + 02 4.962 × 10 75 8.784 × 10 29 8.712 × 10 + 00 1.373 × 10 + 04 2.037 × 10 02
F12Avg 6.973 × 10 01 7.134 × 10 01 7.568 × 10 01 8.626 × 10 01 9.968 × 10 + 01 6.671 × 10 01 6.667 × 10 01 4.219 × 10 + 01 4.085 × 10 + 04 3.189 × 10 + 00
F13Avg 7.366 × 10 130 6.674 × 10 129 1.835 × 10 80 2.845 × 10 05 4.267 × 10 + 04 0.000 × 10 + 00 0.000 × 10 + 00 1.380 × 10 94 2.380 × 10 + 03 1.907 × 10 25
F14Avg 2.567 × 10 108 2.736 × 10 107 1.332 × 10 46 4.045 × 10 03 3.243 × 10 + 04 1.524 × 10 101 9.967 × 10 208 1.723 × 10 66 2.248 × 10 + 03 1.388 × 10 20
F15Avg 1.043 × 10 121 5.494 × 10 123 2.414 × 10 66 1.704 × 10 04 5.229 × 10 + 02 8.836 × 10 130 3.992 × 10 262 1.712 × 10 86 3.461 × 10 + 01 1.650 × 10 27
F16Avg 0.000 × 10 + 00 0.000 × 10 + 00 3.055 × 10 + 01 2.620 × 10 + 01 2.530 × 10 + 02 8.527 × 10 15 7.319 × 10 + 00 4.556 × 10 + 01 3.416 × 10 + 02 4.807 × 10 + 01
F17Avg 0.000 × 10 + 00 0.000 × 10 + 00 4.829 × 10 + 01 3.323 × 10 + 01 1.566 × 10 + 02 0.000 × 10 + 00 8.402 × 10 + 00 7.327 × 10 + 01 3.138 × 10 + 02 4.846 × 10 + 01
F18Avg 8.882 × 10 16 8.882 × 10 16 6.578 × 10 03 9.205 × 10 02 1.675 × 10 + 01 4.619 × 10 15 9.379 × 10 14 3.771 × 10 01 1.651 × 10 + 01 4.428 × 10 01
F19Avg 0.000 × 10 + 00 0.000 × 10 + 00 7.410 × 10 02 1.106 × 10 02 7.466 × 10 + 01 0.000 × 10 + 00 3.726 × 10 03 1.081 × 10 + 00 3.067 × 10 + 02 4.346 × 10 01
F20Avg 2.012 × 10 48 1.693 × 10 40 2.028 × 10 02 1.180 × 10 01 2.152 × 10 + 01 1.754 × 10 43 5.267 × 10 04 1.118 × 10 + 00 4.738 × 10 + 01 2.840 × 10 02
F21Avg 1.198 × 10 01 6.094 × 10 02 4.232 × 10 05 1.766 × 10 04 1.496 × 10 + 00 1.101 × 10 02 3.793 × 10 02 6.608 × 10 01 1.218 × 10 + 01 5.707 × 10 02
F22Avg 1.820 × 10 + 00 1.650 × 10 + 00 4.319 × 10 03 3.137 × 10 03 2.944 × 10 + 00 1.782 × 10 01 3.770 × 10 01 2.325 × 10 + 00 1.298 × 10 + 01 6.696 × 10 04
F23Avg 2.209 × 10 + 00 2.513 × 10 + 00 1.724 × 10 01 3.980 × 10 02 4.616 × 10 01 3.270 × 10 01 6.565 × 10 01 9.914 × 10 + 00 1.346 × 10 + 01 1.679 × 10 01
F24Avg 0.000 × 10 + 00 0.000 × 10 + 00 0.000 × 10 + 00 0.000 × 10 + 00 3.689 × 10 + 01 0.000 × 10 + 00 5.607 × 10 + 00 9.873 × 10 + 00 1.924 × 10 + 01 6.941 × 10 01
F25Avg 4.975 × 10 03 4.977 × 10 03 5.981 × 10 01 3.570 × 10 01 3.718 × 10 + 01 1.841 × 10 01 2.637 × 10 01 3.732 × 10 01 1.428 × 10 + 02 9.612 × 10 01
F26Avg 0.000 × 10 + 00 0.000 × 10 + 00 1.530 × 10 03 4.362 × 10 01 9.334 × 10 + 01 0.000 × 10 + 00 0.000 × 10 + 00 2.118 × 10 01 2.713 × 10 + 02 7.696 × 10 02
Note: Bold values indicate the best performance for the respective metric among all compared algorithms.
Table 3. Comparison of the Best fitness of IBKA against other algorithms.
Table 3. Comparison of the Best fitness of IBKA against other algorithms.
Func.MetricsIBKABKAPKORFOGAWOAGWOSCADEPSO
F1Best 1.324 × 10 105 5.801 × 10 103 4.791 × 10 05 6.270 × 10 03 3.873 × 10 + 03 1.295 × 10 84 3.367 × 10 29 1.040 × 10 02 2.967 × 10 + 04 3.189 × 10 02
F2Best 8.445 × 10 54 8.501 × 10 53 1.249 × 10 04 2.732 × 10 01 2.453 × 10 + 01 2.124 × 10 57 1.853 × 10 17 7.622 × 10 04 8.193 × 10 + 01 3.600 × 10 02
F3Best 3.105 × 10 106 2.873 × 10 104 8.909 × 10 01 1.017 × 10 01 8.913 × 10 + 01 2.265 × 10 + 02 6.433 × 10 11 2.685 × 10 + 01 4.744 × 10 + 02 8.146 × 10 + 00
F4Best 5.784 × 10 54 9.505 × 10 54 3.157 × 10 01 4.093 × 10 02 6.472 × 10 + 00 9.728 × 10 01 1.512 × 10 08 2.067 × 10 + 00 7.623 × 10 + 00 4.564 × 10 01
F5Best 5.119 × 10 01 7.545 × 10 01 1.027 × 10 05 6.866 × 10 03 3.924 × 10 + 01 2.520 × 10 02 1.056 × 10 04 3.838 × 10 + 00 2.581 × 10 + 02 4.556 × 10 04
F6Best 6.952 × 10 05 1.287 × 10 05 1.169 × 10 02 1.965 × 10 02 1.265 × 10 01 2.142 × 10 04 6.834 × 10 04 3.335 × 10 02 2.542 × 10 + 01 1.936 × 10 02
F7Best 7.175 × 10 66 7.175 × 10 66 7.175 × 10 66 7.175 × 10 66 2.606 × 10 44 7.175 × 10 66 9.680 × 10 66 2.425 × 10 50 6.684 × 10 51 7.175 × 10 66
F8Best 1.570 × 10 131 7.619 × 10 134 3.173 × 10 20 3.479 × 10 08 2.728 × 10 07 3.042 × 10 133 2.338 × 10 104 3.842 × 10 14 7.721 × 10 02 4.829 × 10 19
F9Best 5.810 × 10 107 2.218 × 10 106 2.808 × 10 06 1.047 × 10 01 2.069 × 10 + 02 2.284 × 10 86 2.734 × 10 30 1.390 × 10 03 3.154 × 10 + 03 5.084 × 10 03
F10Best 2.605 × 10 + 01 2.624 × 10 + 01 2.712 × 10 + 01 3.619 × 10 + 00 1.355 × 10 + 03 2.703 × 10 + 01 2.539 × 10 + 01 3.081 × 10 + 01 7.679 × 10 + 05 2.202 × 10 + 01
F11Best 4.110 × 10 103 7.191 × 10 106 1.354 × 10 05 7.924 × 10 02 2.928 × 10 + 01 4.582 × 10 87 1.173 × 10 30 7.322 × 10 02 7.136 × 10 + 03 1.427 × 10 03
F12Best 6.667 × 10 01 6.667 × 10 01 6.667 × 10 01 7.719 × 10 01 2.915 × 10 + 01 6.667 × 10 01 6.667 × 10 01 9.603 × 10 01 2.856 × 10 + 04 6.791 × 10 01
F13Best 3.279 × 10 147 1.697 × 10 143 1.162 × 10 136 1.120 × 10 08 2.934 × 10 + 01 0.000 × 10 + 00 0.000 × 10 + 00 1.640 × 10 111 2.499 × 10 01 9.457 × 10 35
F14Best 1.990 × 10 123 1.455 × 10 119 8.096 × 10 75 6.298 × 10 05 2.941 × 10 + 02 4.051 × 10 136 7.981 × 10 254 1.912 × 10 78 2.691 × 10 + 00 3.128 × 10 26
F15Best 6.686 × 10 138 2.667 × 10 137 1.019 × 10 103 2.354 × 10 08 8.774 × 10 03 9.099 × 10 157 1.729 × 10 291 2.485 × 10 102 1.261 × 10 02 4.669 × 10 31
F16Best 0.000 × 10 + 00 0.000 × 10 + 00 1.359 × 10 01 1.720 × 10 + 00 2.082 × 10 + 02 0.000 × 10 + 00 5.684 × 10 14 1.391 × 10 01 2.856 × 10 + 04 6.791 × 10 01
F17Best 0.000 × 10 + 00 0.000 × 10 + 00 2.051 × 10 + 00 1.554 × 10 + 01 1.198 × 10 + 02 0.000 × 10 + 00 2.309 × 10 14 2.049 × 10 + 01 2.764 × 10 + 02 2.817 × 10 + 01
F18Best 8.882 × 10 16 8.882 × 10 16 3.918 × 10 04 7.339 × 10 02 1.579 × 10 + 01 8.882 × 10 16 6.839 × 10 14 2.216 × 10 02 1.474 × 10 + 01 3.652 × 10 02
F19Best 0.000 × 10 + 00 0.000 × 10 + 00 3.807 × 10 04 9.767 × 10 04 3.797 × 10 + 01 0.000 × 10 + 00 0.000 × 10 + 00 3.004 × 10 02 2.148 × 10 + 02 1.318 × 10 01
F20Best 9.425 × 10 57 9.291 × 10 55 1.799 × 10 03 2.461 × 10 02 1.593 × 10 + 01 2.230 × 10 58 1.107 × 10 17 6.620 × 10 03 3.770 × 10 + 01 1.006 × 10 02
F21Best 1.536 × 10 02 1.590 × 10 02 3.645 × 10 07 1.012 × 10 04 4.377 × 10 01 2.757 × 10 03 1.221 × 10 02 4.944 × 10 01 8.609 × 10 + 00 1.150 × 10 05
F22Best 7.389 × 10 01 9.429 × 10 01 7.311 × 10 06 1.154 × 10 03 1.798 × 10 + 00 2.702 × 10 02 1.962 × 10 04 1.909 × 10 + 00 7.657 × 10 + 00 2.499 × 10 05
F23Best 2.002 × 10 01 1.378 × 10 01 3.320 × 10 02 1.748 × 10 02 2.120 × 10 01 4.879 × 10 02 1.153 × 10 01 6.814 × 10 + 00 1.001 × 10 + 01 1.378 × 10 02
F24Best 0.000 × 10 + 00 0.000 × 10 + 00 0.000 × 10 + 00 0.000 × 10 + 00 3.316 × 10 + 01 0.000 × 10 + 00 2.452 × 10 + 00 5.504 × 10 + 00 1.424 × 10 + 01 2.619 × 10 04
F25Best 1.076 × 10 105 1.684 × 10 102 3.980 × 10 01 9.950 × 10 02 2.239 × 10 + 01 1.084 × 10 85 9.950 × 10 02 9.950 × 10 02 9.561 × 10 + 01 3.980 × 10 01
F26Best 0.000 × 10 + 00 0.000 × 10 + 00 1.018 × 10 06 2.398 × 10 01 6.316 × 10 + 01 0.000 × 10 + 00 0.000 × 10 + 00 1.126 × 10 04 1.996 × 10 + 02 3.650 × 10 03
Note: Bold values indicate the best performance for the respective metric among all compared algorithms.
Table 4. Comparison of the standard deviation of IBKA against other algorithms.
Table 4. Comparison of the standard deviation of IBKA against other algorithms.
Func.MetricsIBKABKAPKORFOGAWOAGWOSCADEPSO
F1Std 2.969 × 10 92 5.008 × 10 90 1.542 × 10 01 3.374 × 10 03 2.685 × 10 + 03 1.588 × 10 72 1.703 × 10 27 8.610 × 10 + 00 4.179 × 10 + 03 2.171 × 10 01
F2Std 1.527 × 10 37 4.385 × 10 40 3.013 × 10 02 9.557 × 10 02 3.715 × 10 + 00 2.020 × 10 51 6.881 × 10 17 2.463 × 10 02 1.124 × 10 + 06 3.067 × 10 + 00
F3Std 4.144 × 10 92 2.452 × 10 72 1.587 × 10 + 01 7.992 × 10 01 4.940 × 10 + 01 1.110 × 10 + 02 6.246 × 10 07 4.367 × 10 + 01 1.293 × 10 + 02 1.057 × 10 + 01
F4Std 5.383 × 10 48 1.807 × 10 45 2.318 × 10 01 9.362 × 10 03 3.700 × 10 01 2.817 × 10 + 00 1.716 × 10 07 9.022 × 10 01 3.578 × 10 01 1.284 × 10 01
F5Std 6.241 × 10 01 1.074 × 10 + 00 1.187 × 10 03 2.544 × 10 03 2.312 × 10 + 01 4.421 × 10 02 3.705 × 10 01 1.266 × 10 + 00 5.128 × 10 + 01 1.838 × 10 03
F6Std 1.153 × 10 03 2.397 × 10 04 2.874 × 10 02 1.112 × 10 02 1.357 × 10 01 4.015 × 10 03 9.171 × 10 04 1.190 × 10 01 8.014 × 10 + 00 1.171 × 10 02
F7Std 1.832 × 10 40 8.889 × 10 32 2.163 × 10 81 2.163 × 10 81 1.267 × 10 36 2.163 × 10 81 8.119 × 10 58 7.544 × 10 40 1.276 × 10 40 1.715 × 10 53
F8Std 1.971 × 10 116 2.935 × 10 111 1.703 × 10 09 3.094 × 10 07 4.075 × 10 05 1.498 × 10 107 3.341 × 10 97 7.848 × 10 05 1.505 × 10 01 3.289 × 10 16
F9Std 1.992 × 10 93 6.490 × 10 77 9.592 × 10 03 6.251 × 10 02 2.644 × 10 + 02 4.516 × 10 75 3.054 × 10 28 4.693 × 10 + 00 5.266 × 10 + 02 6.672 × 10 02
F10Std 7.830 × 10 01 8.081 × 10 01 9.450 × 10 + 00 5.811 × 10 + 00 1.533 × 10 + 03 4.236 × 10 01 9.049 × 10 01 1.876 × 10 + 03 2.869 × 10 + 05 3.359 × 10 + 01
F11Std 1.438 × 10 80 2.730 × 10 78 4.210 × 10 02 5.102 × 10 02 7.589 × 10 + 01 2.166 × 10 74 1.056 × 10 28 1.208 × 10 + 01 3.501 × 10 + 03 1.719 × 10 02
F12Std 9.424 × 10 02 1.141 × 10 01 1.741 × 10 01 7.032 × 10 02 8.568 × 10 + 01 2.922 × 10 04 6.939 × 10 05 1.401 × 10 + 02 9.810 × 10 + 03 4.967 × 10 + 00
F13Std 3.124 × 10 129 2.118 × 10 128 8.208 × 10 80 4.649 × 10 05 6.840 × 10 + 04 0.000 × 10 + 00 0.000 × 10 + 00 5.442 × 10 94 7.539 × 10 + 03 7.322 × 10 25
F14Std 7.131 × 10 108 1.208 × 10 106 5.828 × 10 46 5.979 × 10 03 7.381 × 10 + 04 6.816 × 10 101 0.000 × 10 + 00 5.571 × 10 66 5.282 × 10 + 03 5.749 × 10 20
F15Std 4.062 × 10 121 2.450 × 10 122 1.080 × 10 65 2.076 × 10 04 7.779 × 10 + 02 3.951 × 10 129 0.000 × 10 + 00 7.597 × 10 86 8.285 × 10 + 01 3.863 × 10 27
F16Std 0.000 × 10 + 00 0.000 × 10 + 00 2.548 × 10 + 01 1.429 × 10 + 01 2.172 × 10 + 01 2.082 × 10 14 2.289 × 10 + 01 5.314 × 10 + 01 1.781 × 10 + 01 1.509 × 10 + 01
F17Std 0.000 × 10 + 00 0.000 × 10 + 00 2.623 × 10 + 01 1.476 × 10 + 01 1.974 × 10 + 01 0.000 × 10 + 00 7.261 × 10 + 00 2.637 × 10 + 01 2.476 × 10 + 01 1.360 × 10 + 01
F18Std 0.000 × 10 + 00 0.000 × 10 + 00 5.143 × 10 03 1.328 × 10 02 4.428 × 10 01 2.438 × 10 15 2.050 × 10 14 6.254 × 10 01 6.987 × 10 01 5.254 × 10 01
F19Std 0.000 × 10 + 00 0.000 × 10 + 00 9.158 × 10 02 8.461 × 10 03 2.337 × 10 + 01 0.000 × 10 + 00 7.359 × 10 03 6.772 × 10 01 4.769 × 10 + 01 2.341 × 10 01
F20Std 5.415 × 10 48 7.573 × 10 40 3.678 × 10 02 2.546 × 10 01 3.135 × 10 + 00 7.844 × 10 43 5.168 × 10 04 1.681 × 10 + 00 4.975 × 10 + 00 1.686 × 10 02
F21Std 1.753 × 10 01 5.684 × 10 02 7.734 × 10 05 4.137 × 10 05 8.716 × 10 01 9.238 × 10 03 2.309 × 10 02 1.424 × 10 01 2.094 × 10 + 00 7.870 × 10 02
F22Std 6.721 × 10 01 3.934 × 10 01 1.041 × 10 02 2.836 × 10 03 7.975 × 10 01 1.126 × 10 01 2.252 × 10 01 1.935 × 10 01 2.345 × 10 + 00 2.442 × 10 03
F23Std 3.105 × 10 + 00 2.746 × 10 + 00 1.373 × 10 01 1.596 × 10 02 2.402 × 10 01 2.997 × 10 01 6.129 × 10 01 1.682 × 10 + 00 2.208 × 10 + 00 2.234 × 10 01
F24Std 0.000 × 10 + 00 0.000 × 10 + 00 0.000 × 10 + 00 0.000 × 10 + 00 1.913 × 10 + 00 0.000 × 10 + 00 2.040 × 10 + 00 1.743 × 10 + 00 2.590 × 10 + 00 8.867 × 10 01
F25Std 2.225 × 10 02 2.225 × 10 02 2.492 × 10 01 1.754 × 10 01 1.067 × 10 + 01 2.148 × 10 01 1.524 × 10 01 2.211 × 10 01 2.573 × 10 + 01 2.992 × 10 01
F26Std 0.000 × 10 + 00 0.000 × 10 + 00 1.854 × 10 03 2.305 × 10 01 1.861 × 10 + 01 0.000 × 10 + 00 0.000 × 10 + 00 4.199 × 10 01 4.120 × 10 + 01 1.475 × 10 01
Note: Bold values indicate the best performance for the respective metric among all compared algorithms.
Table 5. Performance Ranking Summary of Algorithms Across 26 Benchmark Functions.
Table 5. Performance Ranking Summary of Algorithms Across 26 Benchmark Functions.
AlgorithmAvgBestStd
1st2nd3rd1st2nd3rd1st2nd3rd
IBKA158217611473
BKA10748951064
PKO312103312
RFO235124234
GA000000100
WOA543454443
GWO421322532
SCA002001002
DE000000000
PSO114113123
Note: Numbers indicate how many functions (out of 26) each algorithm ranked 1st/2nd/3rd. For minimization problems, lower values indicate better performance.
Table 6. Statistical Test Results.
Table 6. Statistical Test Results.
Func.BKAPKORFOGAWOAGWOSCADEPSO
F1 2.853 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 1.235 × 10 07 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F2 4.353 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 2.416 × 10 07 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F3 8.920 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F4 1.794 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F5 2.616 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 3.293 × 10 05 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F6 2.977 × 10 01 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 4.680 × 10 05 6.674 × 10 06 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F7 9.023 × 10 01 9.029 × 10 08 9.029 × 10 08 4.654 × 10 07 9.029 × 10 05 1.993 × 10 02 1.659 × 10 05 7.709 × 10 05 7.761 × 10 03
F8 6.750 × 10 01 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 4.320 × 10 03 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F9 1.428 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F10 5.428 × 10 01 3.750 × 10 04 2.041 × 10 05 6.796 × 10 08 6.359 × 10 01 2.343 × 10 03 6.796 × 10 08 6.796 × 10 08 1.782 × 10 03
F11 2.564 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 2.218 × 10 07 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F12 8.817 × 10 01 1.250 × 10 05 9.748 × 10 05 6.796 × 10 08 1.807 × 10 05 4.680 × 10 05 9.173 × 10 08 6.796 × 10 08 2.960 × 10 07
F13 2.229 × 10 02 3.416 × 10 07 6.796 × 10 08 6.796 × 10 08 8.007 × 10 09 8.007 × 10 09 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F14 3.554 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.868 × 10 04 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F15 4.964 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 7.948 × 10 07 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F16 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09 8.036 × 10 02 7.803 × 10 09 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09
F17 1.000 × 10 + 00 8.007 × 10 09 7.992 × 10 09 8.007 × 10 09 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09
F18 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09 7.675 × 10 07 7.732 × 10 09 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09
F19 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09 1.000 × 10 + 00 1.980 × 10 02 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09
F20 4.903 × 10 02 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 8.357 × 10 04 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F21 5.428 × 10 01 6.796 × 10 08 6.796 × 10 08 1.657 × 10 07 1.918 × 10 07 4.320 × 10 03 1.235 × 10 07 6.796 × 10 08 6.389 × 10 02
F22 5.250 × 10 01 6.796 × 10 08 6.796 × 10 08 1.190 × 10 04 6.796 × 10 08 7.898 × 10 08 7.718 × 10 03 6.796 × 10 08 6.796 × 10 08
F23 7.150 × 10 01 2.960 × 10 07 6.796 × 10 08 1.190 × 10 04 1.104 × 10 05 7.718 × 10 03 2.356 × 10 06 7.948 × 10 07 1.918 × 10 07
F24 1.000 × 10 + 00 1.000 × 10 + 00 1.000 × 10 + 00 8.007 × 10 09 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09
F25 4.903 × 10 01 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 3.288 × 10 05 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08 6.796 × 10 08
F26 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09 1.000 × 10 + 00 1.000 × 10 + 00 8.007 × 10 09 8.007 × 10 09 8.007 × 10 09
Note: Bold values indicate significant difference from IBKA.
Table 7. Binarization Methods for Comparison Algorithms.
Table 7. Binarization Methods for Comparison Algorithms.
AlgorithmBinarization Method
bPSOSigmoid transfer function
bDEThreshold method ( θ = 0.5 )
bGWOModified sigmoid function
bWOAThreshold method ( θ = 0.5 )
bSCAThreshold method ( θ = 0.5 )
bGASigmoid transfer function
bBKASigmoid transfer functionn
bIBKAAdaptive multi-transfer functions (proposed: Section 3.3.2)
bPKOThreshold method ( θ = 0.5 )
bRFOSigmoid transfer function
Table 8. Summary of Fault Detection Datasets.
Table 8. Summary of Fault Detection Datasets.
DatasetDomainFeaturesInstancesClass Distribution
DS1Semiconductor5911567Normal: 93.4%, Faulty: 6.6%
DS2Mechanical9943Normal: 69.14%, Faulty: 30.86%
DS3Software213301Normal: 68.95%, Faulty: 31.05%
DS4Software213301Normal: 69.34%, Faulty: 30.66%
DS5Software213301Normal: 68.62%, Faulty: 31.38%
DS6Software213302Normal: 69.05%, Faulty: 30.95%
Table 9. Average Avg of feature subsets obtained by bIBKA versus other competing algorithms.
Table 9. Average Avg of feature subsets obtained by bIBKA versus other competing algorithms.
DatasetsbIBKAbBKAbPKObDEbSCAbPSObGWObWOAbGAbRFO
DS10.72100.63700.71500.67200.68100.67280.66360.67200.31500.6720
DS20.09600.09600.09800.10000.11900.10200.10600.11300.09900.0960
DS30.07400.08100.07400.08500.07900.08100.08400.08300.07900.0751
DS40.11200.11700.11400.11900.12200.12300.12100.12100.12000.1200
DS50.14300.15200.13800.14800.14000.15000.15200.16700.14900.1380
DS60.05200.05700.05500.05700.05800.05700.05700.05600.05600.0560
Note: Bold values indicate the best performance for the respective metric among all compared algorithms.
Table 10. Average Accuracy of feature subsets obtained by bIBKA versus other competing algorithms.
Table 10. Average Accuracy of feature subsets obtained by bIBKA versus other competing algorithms.
DatasetsbIBKAbBKAbPKObDEbSCAbPSObGWObWOAbGAbRFO
DS10.72100.63700.71500.67200.68100.67280.66360.67200.31500.6720
DS20.88900.82400.88200.88800.87700.88800.88200.86100.88800.8880
DS30.91400.90400.90900.90900.90900.90000.90400.90900.90900.9040
DS40.86900.82800.84800.86800.82800.85800.82800.83800.85800.8380
DS50.77200.76000.77100.78600.77600.77600.79800.78800.77600.7710
DS60.94800.93800.94000.94500.94000.94300.94300.94300.94300.9400
Note: Bold values indicate the best performance for the respective metric among all compared algorithms.
Table 11. Average F1-score of feature subsets obtained by bIBKA versus other competing algorithms.
Table 11. Average F1-score of feature subsets obtained by bIBKA versus other competing algorithms.
DatasetsbIBKAbBKAbPKObDEbSCAbPSObGWObWOAbGAbRFO
DS10.58890.57000.58800.58800.61800.58800.58600.58800.63000.5880
DS20.88200.81400.86300.87200.87000.88000.87500.85300.88200.8820
DS30.66100.59400.64500.61600.63100.60200.59400.65900.63100.6480
DS40.65100.56400.57900.64300.51400.59700.55300.56400.62400.5640
DS50.63600.58800.62600.63100.62600.62200.66000.63300.62600.6260
DS60.70100.58200.57200.67700.57200.60300.60300.59100.61600.4840
Note: Bold values indicate the best performance for the respective metric among all compared algorithms.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qi, K.; Wei, K.; Cheng, R.; Liang, G.; Hu, J.; Wu, W. An Improved Black-Winged Kite Algorithm for Global Optimization and Fault Detection. Biomimetics 2025, 10, 728. https://doi.org/10.3390/biomimetics10110728

AMA Style

Qi K, Wei K, Cheng R, Liang G, Hu J, Wu W. An Improved Black-Winged Kite Algorithm for Global Optimization and Fault Detection. Biomimetics. 2025; 10(11):728. https://doi.org/10.3390/biomimetics10110728

Chicago/Turabian Style

Qi, Kun, Kai Wei, Rong Cheng, Guangmin Liang, Jiashun Hu, and Wangyu Wu. 2025. "An Improved Black-Winged Kite Algorithm for Global Optimization and Fault Detection" Biomimetics 10, no. 11: 728. https://doi.org/10.3390/biomimetics10110728

APA Style

Qi, K., Wei, K., Cheng, R., Liang, G., Hu, J., & Wu, W. (2025). An Improved Black-Winged Kite Algorithm for Global Optimization and Fault Detection. Biomimetics, 10(11), 728. https://doi.org/10.3390/biomimetics10110728

Article Metrics

Back to TopTop