Next Article in Journal
An AAM-Based Identification Method for Ear Acupoint Area
Next Article in Special Issue
Biomimetics Design of Tooth Root Zone at Cylindrical Gears Profile
Previous Article in Journal / Special Issue
Reptile Search Algorithm Considering Different Flight Heights to Solve Engineering Optimization Design Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards an Optimal KELM Using the PSO-BOA Optimization Strategy with Applications in Data Classification

1
School of Intelligent Manufacturing and Electronic Engineering, Wenzhou University of Technology, Wenzhou 325035, China
2
Intelligent Information Systems Institute, Wenzhou University, Wenzhou 325035, China
3
Taishun Research Institute, Wenzhou University of Technology, Wenzhou 325035, China
*
Author to whom correspondence should be addressed.
Biomimetics 2023, 8(3), 306; https://doi.org/10.3390/biomimetics8030306
Submission received: 24 June 2023 / Revised: 9 July 2023 / Accepted: 9 July 2023 / Published: 12 July 2023

Abstract

:
The features of the kernel extreme learning machine—efficient processing, improved performance, and less human parameter setting—have allowed it to be effectively used to batch multi-label classification tasks. These classic classification algorithms must at present contend with accuracy and space–time issues as a result of the vast and quick, multi-label, and concept drift features of the developing data streams in the practical application sector. The KELM training procedure still has a difficulty in that it has to be repeated numerous times independently in order to maximize the model’s generalization performance or the number of nodes in the hidden layer. In this paper, a kernel extreme learning machine multi-label data classification method based on the butterfly algorithm optimized by particle swarm optimization is proposed. The proposed algorithm, which fully accounts for the optimization of the model generalization ability and the number of hidden layer nodes, can train multiple KELM hidden layer networks at once while maintaining the algorithm’s current time complexity and avoiding a significant number of repeated calculations. The simulation results demonstrate that, in comparison to the PSO-KELM, BBA-KELM, and BOA-KELM algorithms, the PSOBOA-KELM algorithm proposed in this paper can more effectively search the kernel extreme learning machine parameters and more effectively balance the global and local performance, resulting in a KELM prediction model with a higher prediction accuracy.

1. Introduction

Data classification is one of the most important research hotspots in the field of high-tech at present. It uses certain features to discriminate or classify a group of objects. The information involved in data classification often has the characteristics of high dimensions, many influencing factors, and complex relationships [1]. It is often difficult to effectively determine its laws by human thinking alone, and it needs to be completed through certain mathematical methods with the help of computers. How to discover more and more valuable, associated information from this complex data information, find its internal laws, and establish a model that can better reflect the actual characteristics of the research object, be easily integrated with prior knowledge, and be adaptable to large-scale data-processing requirements are gradually becoming the focus of current data classification [2]. Multi-label classification, or MLC, or the study of a thing according to many class label ideas, has become particularly significant in order to address the inadequacies of conventional single-label classification. The multi-label data flow exhibits the features of vast speed and idea drift in the sphere of practical application, making the conventional multi-label classification algorithm unable to directly address such issues. Designing a reliable multi-label data flow classification approach has therefore emerged as a crucial and difficult challenge due to the need to process these new data fast while having limited time and memory as well as to adjust the concept drift in the data flow environment [3].
In engineering applications, learning for data classification is very desirable. Traditional learning techniques, such as artificial neural networks (ANN) and support vector machines (SVM), appear helpless in the face of the twin needs of quicker training speed and greater learning accuracy [4,5]. Huang et al.’s extreme learning machine (ELM) is a single-layer feedforward network learning technique (single-hidden layer feedforward neural networks, SLFNs) that does not need changing the hidden layer neuron network settings. It offers quick training and effective learning outcomes, which are important characteristics. Due to this, the distributed extreme learning machine (D-ELM) avoids reading all samples into memory at once by partitioning matrix operations, but it also resolves the issue of memory shortage while training vast amounts of sample data [6,7]. Every time they run, they only train an ELM network with a certain amount of hidden layer nodes. Similar to cross-validation, they do not take into account the generalization capacity under various training and test set divisions. They want to improve their generalization skills and accuracy. Numerous runs of the solution to must be performed. When comparing the generalization capacity or the network performance of various hidden layer node numbers, it is impossible to ensure that the hidden layer network parameters are consistent, making it difficult to accurately and intuitively learn their impact on the model [8]. This is due to the random generation of hidden layer network parameters.

1.1. Problem Description and Research Motivation

In recent years, many scholars have applied neural network-based algorithms to data classification research, such as BP neural network, discrete Hopfield network, support vector machine, self-organizing network, fuzzy neural network, and generalized neural network, and achieved many results. Since most neural networks use the gradient descent method, there are often shortcomings, such as slow training speed, easy to fall into local minimum, and learning rate sensitivity. Therefore, it is necessary to explore a fast-training speed, accurate optimal solution, and good generalization The performance of the training algorithm is the main goal of improving the performance of the neural network.
The KELM model relies on both feature selection and parameter optimization, and these two processes are complementary and cooperative. To avoid overfitting and lower the computational cost of the training model, feature selection chooses the most pertinent and discriminative feature subset from the original feature space and eliminates redundant and unimportant features [9]. A proper parameter setup can significantly enhance the KELM model’s classification performance and yield superior classification outcomes. These two elements are taken into account during the design and are optimized simultaneously to increase the KELM model’s capacity for generalization [10].
Due to its strong search capabilities, the meta-heuristic algorithm has received considerable attention recently. Numerous study findings have established that this algorithm is superior to conventional approaches for solving optimization issues [11]. Researchers working on this project have to date suggested a few optimization strategies for parameter and feature selection. For instance, Alcin et al. introduced the genetic algorithm (GA) to KELM model in 2014 after using the GA method to improve the sparse output weight vector of the KELM model. The kernel KELM model’s parameters were optimized using the particle swarm optimization (PSO) technique by Bin Li et al. in the same year [12,13].
The primary issues with KELM’s categorization model are:
The selection of pertinent parameters is a significant issue for both the KELM study model and the SVM model, although there are not many studies on this topic [14].
It is common to conduct feature selection and parameter optimization independently. In essence, the two cooperate and support one another. The best model cannot be assured if they are optimized individually [15].
A novel classification method for kernel extreme learning machines (PSOBOA-KELM) based on the modified PSO butterfly optimization algorithm is suggested in order to solve the aforementioned issues. This technique concurrently performs parameter optimization of the kernel ELM model and feature selection based on the enhanced butterfly algorithm (PSOBOA) to increase KELM’s generalization performance [16,17]. The particle swarm optimization algorithm is introduced at the same time to improve the GSA algorithm’s performance in optimization by addressing the latter stages of iteration’s slow convergence speed and weak local search ability, and the chaos control strategy is intended to broaden the group’s diversity. The PSOBOA-KELM technique suggested in this work simultaneously maximizes the number of hidden layer nodes and generalization ability. It saves a vast amount of time by dividing training into segments based on samples. Thus, potential solutions should be investigated.

1.2. Contribution

Compared with the traditional method, the butterfly optimization algorithm optimized by particle swarm optimization proposed in this paper comprehensively considers issues, such as improving the classification accuracy of the algorithm and the generalization performance of KELM. The main contributions of this paper are as follows:
  • Characterize the swarm intelligence optimization algorithm/butterfly optimization algorithm and classify the current data classification methods.
  • Propose a novel data classification method of kernel extreme learning machine based on the butterfly optimization algorithm optimized by particle swarm optimization (PSOBOA-ELM).
  • Provide extensive simulation results to demonstrate the use and efficiency of the proposed data classification method.
  • Evaluate the performance of the proposed algorithms by comparing them with the data classification methods of other algorithms.
The remainder of this paper is organized as follows: Section 2 discusses the related work. Section 3 describes the basic principles of kernel extreme learning machine. Section 4 describes the principles of the butterfly optimization algorithm optimized using particle swarm optimization. Section 5 describes the implement steps of the proposed algorithm design idea of the kernel extreme learning machine method based on the butterfly optimization algorithm optimized by particle swarm optimization. Section 6 provides the parameters and simulation results that validate the performance of the proposed algorithm. Section 7 concludes the paper.

2. Related Work

Data categorization is frequently employed as the fundamental processing technique in contemporary intelligent data processing. Machine learning is a powerful tool for achieving the objective of data processing since it makes use of data sets to create classification models with great generalization capabilities.
Support vector machines are at present being used by some academics to handle multi-instance learning challenges. The conventional SVM is particularly sensitive to noisy points and singular points in the sample. The CA-SVM based sentiment analysis model that Cyril et al. suggested uses automatic learning to read Twitter datasets, analyze them, and extract features to provide a list of phrases [18]. More characteristics in the input electrocardiosignal (ECG) signal were classified using the SVM model and weighted kernel function approach by Varatharajan et al. At present, the existing multi-label classification methods mainly include: batch processing methods and online learning methods. Among them, the batch processing method defaults to the one-time arrival of each training and testing data set, and uses problem transformation and algorithm self-adaptation to solve multi-label classification problems based on all existing information. The extreme learning machine (ELM) proposed by Huang and its improved algorithm have the characteristics of high speed and high efficiency, avoiding the cumbersome iterative learning process [19]. The random setting of learning parameters caused by the iterative learning of the traditional feed-forward neural network can easily encounter problems, such as the local minima, while improving the algorithm can further improve the classification accuracy. Therefore, related research based on (kernel) extreme learning machines has been widely applied to multi-label classification problems, and a series of results has been achieved [20]. However, due to the characteristics of massive and fast data streams emerging in the practical application field, it is difficult to obtain them all at once. At the same time, when new data arrives, these batch-processing algorithms continue to retrain new data and discard old models, resulting in a large loss of effective historical data [21]. Therefore, learning models that can handle data stream environments are receiving more and more attention.
Ensemble techniques have been one of the most significant advancements in machine learning over the past ten years. In actuality, the kernel function is utilized as a crucial theoretical tool in the data preprocessing since the data in the data set are linearly indistinguishable, making it necessary to build an appropriate data classification procedure. The goal of the kernel function is to discover the classification hyperplane of the low-dimensional, indistinguishable data in the new high-dimensional space by nonlinear transformation, allowing for the separation of the data. The construction and parameter selection of the kernel function are at present its key areas of emphasis. To address the multi-class unbalanced data classification problem, Zhang et al. suggested a support vector machine (SVM) technique based on a proportional kernel and proposed a scaling kernel function, which employs weighting factors to compute its parameters. The issue of skewed distribution-induced classifier performance reduction has a high degree of generalizability [22]. To discriminate between various kinds of ground objects, Chen et al. presented a novel hybrid kernel function SVM point cloud classification technique, and they created a Gaussian and polynomial hybrid kernel function to increase the classification accuracy [23]. Xie et al. calculated the similarity of samples with several unknown attributes using the characteristics of kernel functions. To compute the kernel function and solve additive kernel singular values, an effective technique was also shown [24]. Zhang et al. developed a novel conformal function to scale the kernel matrix of ODM in order to increase the separability of the training data in the feature space. They also presented a kernel modified ODM kernel function (KMODM) to remove the unbalanced data classification approach [25]. When tackling small sample, nonlinear, and high dimensional problems, it demonstrates several distinct benefits and may be successfully applied to various machine learning disciplines.
Although it has been demonstrated that the Ada Boost approach, which uses a neural network as the basic classifier, has a high generalization performance, training is still not without its challenges. Diversity has undergone extensive research as a crucial component in the generalization performance of classifiers, and certain approaches to assess diversity have also been presented. In order to overcome the limitations of fixed representations, Deng et al. used deep learning to perform large-scale task-driven feature learning from big data. They also demonstrated its utility in image classification, high-frequency financial data prediction, and brain MRI And how well these three duties may be divided [26]. Saritas et al. assessed the classification performance of a Bayesian classifier and an artificial neural network applied to nine inputs and one output and compared the findings [27]. By fusing morphological neurons with perceptrons, Gerardo et al. suggested two novel hybrid neural architectures. They then evaluated them using 25 low-dimensional standard data sets and a large data set. The suggested approach achieved improved accuracy while using fewer learning parameters [28]. In contrast to conventional techniques and other state-of-the-art techniques, Wu et al. used a convolutional recurrent neural network (CRNN) to learn more discriminative features for hyperspectral data classification, using recurrent layers to further extract spectral context information from features generated by convolutional layers. The suggested technique offers improved classification performance for hyperspectral data classification when compared to deep learning methods [29].
At present, some achievements have used the sliding window technology to apply extreme learning machines to solve the multi-label classification of data streams, but this method does not consider the problem of class label correlation between multiple labels and concept drift in the data stream environment. On the other hand, some researchers pointed out that, when dealing with data streams, it is necessary to consider the model to make accurate predictions under limited time and memory and include solutions to the problem of concept drift. These requirements pose more challenges to the classification of multi-label data streams. Most of the multi-label classification algorithms in the data flow environment use problem transformation to convert classification into a series of stable learning tasks. Although this method can be applied to a certain extent, it ignores the correlation between labels. At the same time, it does not take into account the high-speed and changeable characteristics of the newly arrived data, and the implicit concept drift problem is also difficult to solve by the problem transformation method. One of the challenging issues in data mining is data categorization, a topic that has attracted considerable attention from both domestic and international scholars studying artificial intelligence. It is required to fundamentally optimize the imbalanced data in order to address the issue of unbalanced data categorization. At present, the outcomes of academic research are improving yearly. Nearly 900 scholarly publications on unbalanced data categorization were published between May 2018 and 2022, a significant increase over the preceding ten years.

3. Kernel Extreme Learning Machine

The extreme learning machine consists of three parts: input layer, hidden layer, and output layer. For a given training sample, continuously optimize the input weights and bias values between the connected input layer and the hidden layer, and maintain them unchanged during the training process. Assume a training sample set of { x i , c i } , i = 1 , 2 , …, N is given, where xi is the input value of the training sample and ci is the corresponding output value. Let the limit learning machine have h hidden layer nodes, the network output is f, and g(*) is the activation function; then, the input and output model of the limit learning machine can be expressed by Formula (1) [30].
f ( x j ) = i = 1 h β i g ( ω i × x j + b i ) , j = 1 , 2 , N
In the formula, the output weight of the input node and the i-th hidden layer node are represented by β i . The input weight of the i-th hidden node and the input node are represented by ω i . The offset value of the i-th hidden node is represented by bi.
H = [ g ( ω 1 × x 1 + b 1 )     g ( ω L × x 1 + b L ) g ( ω 1 × x 2 + b 1 ) g ( ω L × x 2 + b L ) g ( ω 1 × x N + b 1 ) g ( ω 1 × x N + b 1 )   ] N × L β = [ β 1 β 2 β L ] L × 1 , T = [ c 1 c 2 c L ] L × 1
The output weight can be expressed by Formula (3).
β ^ = H T
In the equation, H is the inverse of matrix H.
Replace the hidden layer in the extreme learning machine with the idea of kernel function mapping in the support vector machine. Then, the kernel extreme learning machine can be expressed by Formula (4) [31].
min P E L M = 1 2 β 2 + C 2 i = 1 N ξ i 2 s . t . h ( x i ) β = c i ξ i
Therefore, the input and output model of the kernel extreme learning machine is as shown in Formula (5).
f ( x ) = h ( x ) β = h ( x ) H T ( I C + H H T ) 1 T
Define the extreme learning machine kernel matrix as Formula (6) [32].
Ω E L M = H H T Ω E L M i , j = h ( x i ) h ( x j ) = K ( x i , x j )
The corresponding input–output model can be expressed as shown in Formula (7).
f ( x ) = [ K ( x , x 1 ) K ( x , x 2 ) K ( x , x N ) ] T ( I C + Ω E L M ) 1 T
The feature mapping h ( x ) of the hidden layer is unknown in kernel limit learning machines, but it is usually calculated using the kernel K ( μ , ν ) ( K ( μ , ν ) = exp ( γ μ ν 2 ) ) to reduce the impact of poor classification results caused by the unreasonable setting of the number of hidden layer nodes (the dimension of feature space).
As a result, kernel ELM has the benefit of effective ELM SVM classification. The number of hidden layer nodes need not be predetermined because KELM determines the hidden layer mapping kernel function in the form of an inner product by introducing a kernel function. As a result, the generalization of the KELM-based electric load forecasting model results in significantly increased capacity and stability.

4. Butterfly Optimization Algorithm Optimized by PSO

4.1. Butterfly Optimization Algorithm

A novel meta-heuristic algorithm called the butterfly optimization algorithm (BOA) mimics the foraging and courting behavior of butterflies. It has a high resilience and global convergence ability while solving complicated functions [33,34]. The BOA algorithm has two crucial inputs: switching probability and scent. The switching probability determines the likelihood that the butterfly will select one of two movement modes, either global or local, and the smell stands in for the quality of the particular butterfly’s current position [35]. The butterfly colony is first dispersed at random in the solution space, and the butterflies that have a strong scent draw other individual butterflies to them. The aim optimization is accomplished by consistently updating the butterfly colony’s location [36]. Each butterfly in the butterfly optimization algorithm has a distinct scent and perception ability, and the strength of smell perception varies between individuals. Formula (8) illustrates how strongly other butterflies can smell a person among them:
f(x) = cIa
Among them, f(x) represents the odor intensity function; c represents the sensory shape coefficient; I represents the stimulus intensity, that is, the fitness value of the function; and a represents the intensity coefficient, and the value is in [0, 1].
The sensory shape coefficient c can theoretically take any value within [0, ∞), and its calculation is shown in Formula (9):
ct+1 = ct + [0.025/(ct·Tmax)]
Among them, the initial value of c is 0.01, and Tmax is the maximum number of iterations of the algorithm. The BOA algorithm determines the global search and local search of the algorithm according to the switching probability p, and the position update formula is shown in Formula (10):
x i t + 1 = { x i t + ( r 2 g * x i t ) f i , p < r a n d x i t + ( r 2 x j t x k t ) f , p r a n d
Among them, g* is the best position of all butterflies in the current iteration; xtj and xtk represent the spatial positions of the j-th butterfly and the k-th butterfly in the t-th iteration, respectively; the value of r is a random number of [0,1] number; and fi is the fitness value of the i-th butterfly.

4.2. Particle Swarm Optimization Algorithm

A swarm intelligence optimization system called particle swarm mimics how birds fly while looking for food in a multidimensional search environment. Particle position and velocity are the two key aspects of the PSO method optimization [37,38]. Each one of them is referred to as a particle, and each particle’s initial position and velocity in the search space are initialized at random [39]. The particles’ positions and velocities are updated in accordance with Formulas (11) and (12):
v i t + 1 = ω v i t + c 1 r a n d 1 ( p b e s t x i t ) + c 2 r a n d 2 ( g b e s t x i t )
x i t + 1 = x i t + v i t + 1
Among them, vti and vit+1 represent the velocities of the i-th particle at the t and t + 1 iterations, respectively; and pbest and gbest represent the initial and global optimal positions of particles, respectively. Generally, the hyperparameter c1 = c2 = 2; rand1 and rand2 are random numbers of (0, 1); and ω represents the inertia weight coefficient.

4.3. Butterfly Optimization Algorithm Optimized by PSO (PSOBOA)

(1) Algorithm population initialization
Assume that, in the D-dimensional search space, the greedy strategy is used to generate a new race to generate the initial solution expression, which is shown in Formula (13):
X i = L b + ( U b L b ) O
Among them, Xi represents the spatial position of the i-th butterfly (i = 1, 2, 3,…, N) in the butterfly population, and N represents the number of initial solutions. Lb and Ub represent the upper and lower bounds of the search space, respectively; and O represents a matrix of random numbers with elements (0, 1).
(2) Algorithmic Global Search
The global search phase of butterfly optimization algorithm optimized by PSO (PSOBOA) can be expressed by Formulas (14) and (15):
X i t = ω X i t 1 + ( r 2 g b e s t ω X i t 1 ) f i
X i t + 1 = X i t + ω V i t + C 1 r 1 ( p b e s t X i t ) + C 2 r 2 ( g b e s t X i t )
Among them, ω represents the inertia weight coefficient, and Vti and Vit+1 represent the velocities of the i-th particle at time t and t + 1, respectively. The hyperparameter C1 = C2 = 2, and the values of r1 and r2 are random numbers (0, 1).
(3) Local search of algorithms
The local search stage of the PSOBOA algorithm can be represented by Equations (16) and (17):
X i t = ω X i t 1 + ( r 2 × X k t 1 ω X j t 1 ) f i
X i t + 1 = X i t + ω V i t + C 1 r 1 ( p b e s t X i t ) + C 2 r 2 ( g b e s t X i t )
Among them, Xkt−1 and XJt−1 are the positions of the k-th and j-th butterflies randomly selected from the solution space of the t − 1 iteration, respectively; and ω represents the inertia weight coefficient. C1 = C2 = 2, r1, and r2 are random numbers with the values of (0, 1).
(4) Control strategy
Chaos theory has many applications in swarm intelligence optimization algorithm, such as chaos population initialization and chaos adjustment strategy of control parameters. Logistic mapping is a classical chaotic mapping method in chaos theory, and its expression is shown in Formula (18):
z l + 1 = μ z l ( 1 z l )
Among them, l represents the number of iterations of the chaotic map and μ is the chaotic parameter, and its value is in [0, 4]. The chaotic sequence of logistic mapping is (0, 1); when μ = 4, the mapping produces a chaotic phenomenon.
The Lyapunov index is an important index to distinguish the characteristics of chaos. The larger the maximum Lyapunov exponent of the chaotic map, the more obvious its chaotic characteristics and the higher the degree of chaos. The index expression is shown in Formula (19):
λ = lim n 1 n h i = 0 n h 1 ln | f ( z i ) |
where λ represents the Lyapunov exponent; f ( ) represents the first derivative of the chaotic mapping function; and nh represents the number of iterations of the chaotic mapping.
The expression of the sensory shape coefficient c in the PSOBOA algorithm is shown in Formula (20):
c ( t ) = 4 c ( t 1 ) ( 1 c ( t 1 ) )
The inertia weight coefficient ω has a direct impact on the particle flight speed of the PSO algorithm and can adjust the global search and local search capabilities of the algorithm. In this paper, an adaptive adjustment strategy was adopted, as shown in Formula (21):
ω ( t ) = ω max ( ( ω max ω min ) T i ) / T max
Among them, ωmax = 0.9, ωmin = 0.2, and Tmax is the maximum number of iterations of the algorithm.
(5) Algorithm complexity analysis
Assuming that the number of populations of the algorithm is N, the dimension of the search space is D, and the maximum number of searches is Tmax, the complexity of PSOBOA includes: the initialization complexity of the population O(ND), the fitness value calculation complexity O(ND), the global and the position update complexity of local search O(N2logN), the fitness value sorting complexity of the algorithm O(N2), and the control parameter update complexity of the algorithm O(ND). Then, the complexity of PSOBOA algorithm is shown in Formula (22):
O ( H P S B A ) = O ( N D ) + O ( T max ) O ( N D + N 2 log N + N 2 + N D )
The algorithm time complexity of PSOBOA is shown in Formula (23):
O ( B O A ) = O ( N D ) + O ( T max ) O ( N 2 log N + N + N D )

5. Data Classification of KELM Based on PSOBOA Algorithm

The regularization coefficient C and the kernel function parameter S of the kernel extreme learning machine were optimized using the particle swarm optimization butterfly technique, which raises the network’s classification recognition accuracy. We created the data categorization mathematical model after obtaining the optimal parameters. The following are the precise PSOBOA-KELM steps:
(1)
Establish a classification database, divide it into training and testing sets, and normalize them.
(2)
Train KELM using the training set as the input vector.
(3)
The PSOBOA algorithm was used to optimize the regularization coefficient C and kernel function S in KELM. Select the optimal C and S and reconstruct KELM.
(4)
Train the KELM algorithm optimized by PSOBOA again and compare the results.
(5)
Determine whether the termination conditions are met. If satisfied, exit the loop and output the prediction result. Otherwise, it is recalculated.
(6)
Input the test set into the optimized KELM and output the prediction results.
According to the above steps, the flow chart of the PSOBOA-KELM algorithm is shown in Figure 1.

6. Algorithm Simulation and Result Analysis

6.1. Benchmark Function Test

In order to test the performance of the PSOBOA algorithm, eight test functions were used for testing, and it was compared with the particle swarm optimization algorithm (PSO), crow search algorithm (CSA), binary bat algorithm (BBA), and butterfly optimization algorithm (butterfly optimization algorithm, BOA). The improved PSOBOA algorithm was compared and analyzed. The eight test functions in CEC2017 were all evaluated as minimization problems, which were divided into multimodal test functions, mixed functions, and composite functions. The test functions are shown in Table 1. For fair comparison, the solution dimension of all test functions was 30, the population size was set to 30, the search space was all [−100, 100], all algorithms were run independently on each test function 30 times, and the maximum number of iterations for each run was for 100.
In this paper, the results of PSO, CSA, BBA, BOA, and PSOBOA algorithms independently running 30 times on eight test functions were counted. The iterative calculation results of the test functions of the five algorithms are shown in Figure 2.
It can be seen from Figure 2 that, when solving the test function, the optimization results of the BBA, BOA, and PSOBOA algorithms are not much different, but they are all significantly better than the PSO algorithm and CSA algorithm. When solving the multi-peak test function, although the CSA algorithm achieves better results on the two test functions, according to the average ranking of the five algorithms on the multi-peak test function, the PSOBOA algorithm is better than the other four algorithms, the convergence speed is faster. When solving the mixed function, the PSOBOA algorithm achieved the best results on the test functions. When solving the composite function, the optimization effect of the PSOBOA algorithm is not significant compared with the BBA algorithm and BOA algorithm, but it can be seen from the comprehensive mean and standard deviation that the PSOBOA algorithm has high optimization and stable results.
At the same time, it can be seen from the experimental data in Figure 2 that for F3, F4, F6, and F8, PSOBOA has the strongest optimization performance, which is obviously better than PSO, CSA, BOA, and BBA, and F1, F2, F3, and F4 can directly find the optimal value of 0. For F7, the optimization performance of PSOBOA and BOA is almost the same, the average of optimization is slightly better than BBA, and the effect of PSO is the worst. For F6, the optimization performance of PSOBOA algorithm is obviously better than that of PSO, CSA, BOA and BBA, and the optimization stability is the best. The above analysis shows that the overall optimization ability of PSOBOA is better than that of PSO, CSA, BOA, and BBA.

6.2. Simulation Environment Construction

The proposed algorithm was tested using the database’s standard classification data set, and a number of comparisons and experiments were conducted with conventional PSO-KELM [40], BBA-KELM [41], BOA-KELM, and other algorithms in order to confirm its viability and effectiveness. Windows 10 (64-bit), MATLAB 2020b, a 12th Gen Intel(R) Core(TM) i9-12900 CPU running at 3.20 GHz, and 32G of RAM were used as the simulation experiment setting.
We evaluated the classification performance of this method, which are classification accuracy (ACC), sensitivity (SEN), specificity (SPE), precision(PRE), and F-measure, which are defined as follows:
Accuracy is the proportion of the total number of correct predictions. Use the following methods to determine it:
A C C = T P + T N T P + T N + F N + F P × 100 %
Sensitivity is an index used to measure the classifier’s recognition of abnormal records, and is also often expressed as the TP rate.
S E N = T P T P + F N × 100 %
Specificity is often used to estimate the ability of a classification model to identify normal examples, which is also often expressed as the TN rate.
S P E = T N T N + F P × 100 %
Precision is the correct proportion of positive instances of prediction, calculated using:
P R E = T P T P + F P × 100 %
Among them, TP, FP, TN, and FN represent true positive, false positive, true negative, and false positive, respectively.
Lewis and Gale proposed the F-measure in 1994, which is defined as follows:
F = ( β 2 + 1 ) × Pr e c i s i o n × S e n s i t i v i t y β 2 × Pr e c i s i o n + S e n s i t i v i t y
In Equation (28), there is a value from 0 to infinity to control the weights assigned to the precision and sensitivity. If all positive instances are classified incorrectly, any classifier evaluated using the above will have a metric of 0. In this experiment, the β value was set to 1.

6.3. Algorithm Test Comparison and Result Analysis

In order to verify the effectiveness of the proposed method, this part experiments on the PSOBOA-KELM algorithm on seven classification data sets, which are BreastEW, CongressEW, Hepatitis, JPNdata, Parkinson, SpectEW, and Wdbc. The data sets are from the UCI Machine Learning Library (http://archive.ics.uci.edu/mL/datasets, accessed on 1 October 2022). These data sets are mainly divided into binary classification problems, multi-classification problems, and regression fitting problems. The Breastcancer dataset has 699 data, including 9 features and two categories; the Parkinson dataset has 195 data, including 23 features and two categories; the BreastEW dataset has 569 data, including 30 features and two categories; and the Dermatology dataset has 358 data, including 35 features and six categories. The experiments selected seven real datasets widely used for multi-label classification. The learning factor (c1 = c2 = 2) in the important parameters of the particle swarm optimization algorithm was the inertia weight factor w1 = 0.9 and w2 = 0.4. Table 2 summarizes the data size, attribute dimension, number of tags, and cardinality of the seven datasets. The specific description information of the dataset is shown in Table 2.
The data set had to be preprocessed before the experiment, and certain features were missing. These records were averaged in this experiment to guarantee the accuracy of the sample data. To reduce the gap between the eigenvalues and prevent the larger eigenvalues from adversely affecting the smaller eigenvalues, we normalized each eigenvalue to the [−1, 1] interval. The normalized calculation formula is:
x = ( x min a max a min a ) 2 1
where x is the original value of the data, x is the normalized value, max a is the maximum value in feature a, and max a is the minimum value in feature a.
At the same time, in order to obtain an unbiased estimate of the algorithm’s generalization accuracy, k-fold CV is generally used to evaluate the classification accuracy. In this method, all test sets are independent, which can improve the reliability of the results. In this study, the k value was set to 10, that is, each experimental data set was divided into 10 subsets, one of which was taken as the test set each time, and the rest was used as the training set, and then the average value of 10 experiments was calculated as the result of the ten-fold crossover. Each of the above classification experiments was run independently 20 times to ensure the stability of the algorithm.
The parameter settings of the contrast swarm intelligent optimization algorithm involved in this paper are shown in Table 3.
From the results of Table 4 and Table 5, it can be seen that the method proposed in this paper is accurate in accuracy, precision, F-measure, sensitivity, specificity, and MCC. The indicator performs significantly better than other comparative feature selection methods.
For the accuracy indicator, the PSOBOA-KELM feature selection method proposed in this paper has an accuracy rate of 96.49%, 96.56%, 87.87%, 83.96%, and 90% on BreastEW, CongressEW, hepatitisfulldata, JPNdata, Parkinson, SpectEW, and wdbc data sets, respectively. Compared with the PSO-KELM, BBA-KELM, and BOA-KELM feature selection methods, the method proposed in this paper has the highest accuracy rate. For example, in the BreastEW dataset, the accuracy of the method proposed in this paper is 0.91% higher than the PSO-KELM method, 0.91% higher than the BBA-KELM method, and 0.03% higher than the BOA-KELM method. For the precision indicator, on the BreastEW, CongressEW, hepatitisfulldata, JPNdata, Parkinson, SpectEW, and wdbc data sets, the accuracy of the PSOBOA-KELM feature selection method proposed in this paper is 95.98%, 100%, 100%, 78.89%, 92.86%, 87.5%, and 100%, respectively. Compared with the PSO-KELM, BBA-KELM, and BOA-KELM feature selection methods, the proposed method has the highest accuracy. For example, in the BreastEW dataset, the accuracy of the method proposed in this paper is 1.31% higher than the PSO-KELM method, 1.46% higher than the BBA-KELM method, and 1.31% higher than the BOA-KELM method. For the F-measure index, on the BreastEW, CongressEW, hepatitisfulldata, JPNdata, Parkinson, SpectEW, and wdbc data sets, the F-measure of the PSOBOA-KELM feature selection method proposed in this paper is 97.3%, 97.10%, 70.83%, 84.03%, 93.75%, 33.33%, and 96.4%, respectively. Compared with the PSO-KELM, BBA-KELM, and BOA-KELM feature selection methods, the proposed method’s F-measure works better. For example, in the CongressEW dataset, the F-measure value of the method proposed in this paper is 2.76% higher than the PSO-KELM method, 4.79% higher than the BBA-KELM method, and 0.95% higher than the BOA-KELM method.
For the sensitivity index, on the BreastEW, CongressEW, hepatitisfulldata, JPNdata, Parkinson, SpectEW, and wdbc data sets, the sensitivity values of the PSOBOA-KELM feature selection method proposed in this paper are 100%, 94.37%, 58.33%, 93.75%, 100%, 20%, and 93.07%, respectively. Compared with the PSO-KELM, BBA-KELM, and BOA-KELM feature selection methods, the proposed method has a higher sensitivity value. For example, in the CongressEW dataset, the sensitivity value of the method proposed in this paper is 1.92% higher than the PSO-KELM method, 1.78% higher than the BBA-KELM method, and 1.78% higher than the BOA-KELM method. Compared with the PSO-KELM, BBA-KELM, and BOA-KELM feature selection methods, the method proposed in this paper has a higher specificity value. For example, in the BreastEW dataset, the specificity value of the method proposed in this paper is 1.96% higher than the PSO-KELM method, 2.38% higher than the BBA-KELM method, and 2.38% higher than the BOA-KELM method. For the MCC index, the PSOBOA-KELM feature selection method proposed in this paper has the MCC values of 92.58%, 93.16%, 66.39%, 68.1%, and 72.81% on the BreastEW, CongressEW, hepatitisfulldata, JPNdata, Parkinson, SpectEW, and wdbc data sets, respectively. Compared with the PSO-KELM, BBA-KELM, and BOA-KELM feature selection methods, the method proposed in this paper has a higher MCC value. For example, in the wdbc data set, the MCC value of the method proposed in this paper is 3.64% higher than the PSO-KELM method, 5.61% higher than the BBA-KELM method, and 1.93% higher than the BOA-KELM method.
In addition, in order to compare the performance of these four algorithms more intuitively, as shown in Figure 3, the performance evaluation indicators of these four methods are compared in detail. At the same time, the calculation and simulation time consumption of the four algorithms in the seven data sets are also presented, as shown in Figure 4.
According to the experimental findings, the PSOBOA-KELM technique has an acceptable performance in terms of classification, and the calculation and simulation times are not too long. It may choose an acceptable and constrained feature subset, and its classification performance is noticeably better than that of comparable approaches. The algorithm also performs well when it comes to the challenge of classifying various data sets. It takes a fair amount of time and produces an excellent classification accuracy. The comparison of eight datasets is shown in Table 6 and Table 7. In addition, in order to compare the performance of these four algorithms more intuitively, the performance evaluation indicators of these four methods are compared in detail, as shown in Figure 5. At the same time, the calculation and simulation time consumption of the four algorithms in the seven data sets are also presented, as shown in Figure 6.
The four algorithms were tested in the Australian, Breastcancer, Dermatology, HeartEW, Diabetes, Glass, Heart, and Vote8 data sets regarding accuracy, precision, F-measure, sensitivity, specificity, MCC, and other six indicators, and achieved a good classification performance. The calculation and simulation time were also relatively short for the PSOBOA-KELM method. It may choose an acceptable and constrained feature subset, and its classification performance is noticeably better than that of comparable approaches. In addition to achieving an improved classification accuracy, the algorithm also performs well while classifying data from various data sets.
In addition, the simulation experiment comparison of the Sinc function was conducted. The four algorithms were compared by fitting the Sinc function. The expression of the Sinc function is as follows:
f ( x ) = { sin ( x ) x , x 0 0 , x = 0  
We set to generate 2000 [−10, 10] uniformly distributed data sets x, calculated 2000 data sets { x i , f ( x i ) } , i = 1 , 2 , 3 , , 2000 , and then generated 2000 [−0.2, 0.2] uniformly distributed noise ε. Let the training set be { x i , f ( x i ) + ε i } , i = 1 , 2 , 3 , , 2000 and then generate another set of 2000 data sets { y i , f ( y i ) } , i = 1 , 2 , 3 , , 2000 as the test set. In addition, the root-mean-square error (RMSE), mean absolute error (MAE), and relative standard deviation (RSD) were used as the evaluation indicators for error analysis. The calculation formulas of the three indicators are as follows:
R M S E = 1 N i = 1 N ( y ( i ) y ( i ) ) 2
M A E = 1 N i = 1 N | y ( i ) y ( i ) | = 1 N i = 1 N | e i |
R S D = i = 1 N ( y ( i ) y ¯ ( i ) ) 2 i = 1 N ( y ( i ) y ¯ ( i ) ) 2
where parameter y ( i ) represents the measured value, y ( i ) represents the predicted value, parameter N is the number of samples, parameter e i = y ( i ) y ( i ) is the absolute error, and the numerator and denominator of RSD are in the form of standard deviation. The comparison of Sinc function fitting results is shown in Table 8.
It can be seen from Table 8 that, calculated by the PSO-KELM algorithm, the index values of RMSE and MAE are the largest, and the index value of RSD is closer to the smallest, and the performance of the test results is poor. The index value is even smaller, and the performance of the test result is average. The RMSE and MAE index values of the BOA-KELM algorithm are smaller, the RSD index values are closer to larger, and the test results have a better performance. The PSOBOA-KELM algorithm has the smallest RMSE and MAE index values, the RSD index value is closer to the largest, and the test results have the best performance. It shows that the error of the PSOBOA-KELM model is relatively smaller, and the prediction accuracy is better than that of the PSO-KELM, BBA-KELM, and BOA-KELM algorithms. At the same time, this can also be known from the data change trend in Table 2, which indicates that the PSOBOA-KELM algorithm has the best performance, and optimizing the KELM regularization parameter C and kernel function S can improve the prediction accuracy of the KELM model.

7. Conclusions

The model selection problem of kernel extreme learning machines was investigated in this paper utilizing an enhanced butterfly technique that is based on particle swarm optimization (PSOBOA). This study compared the PSO-KELM, BBA-KELM, and BOA-KELM approaches in-depth to the proposed PSOBOA-KELM model. To assess the model’s performance, we used six indicators: accuracy, precision, F-measure, sensitivity, specificity, and MCC. According to the experimental findings, PSOBOA-KELM can swiftly converge to the best solution inside the search space. The model may combine the benefits of the PSOBOA and KELM models and has a good optimization performance thanks to the inclusion of the original butterfly optimization method in the particle swarm search approach. Better performance, fewer algorithm parameters, and quick search times are some of its features. The performance as well as the classification accuracy have both dramatically increased.
Research on the issue of vast sample sparsity in a high-dimensional space will be the next stage, followed by the in-depth study of the categorization of large sample data by logically making use of useful historical data.

Author Contributions

Writing—review and editing, Y.Y.; Methodology, L.C.; data curation, H.C.; software, Y.C.; formal analysis, Z.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Zhejiang Province under grant LY23F010002; in part by Taishun Research institute of Wenzhou University of Technology under grant 202201001; the Wenzhou basic scientific research project under grant R20210030; the Industrial Science and Technology Project of Yueqing City under grant 2022G007; Wenzhou Association for Science and Technology under grant kjfw36; in part by the major scientific and technological innovation projects of the Wenzhou Science and Technology Plan under grant ZG2021021; the school-level scientific research projects of the Wenzhou University of Technology under grants ky202201 and ky202209; the general scientific research projects of the Provincial Department of Education under grant Y202250103; the teaching reform research project of Wenzhou University of Technology under grant 2022JG12; the Wenzhou intelligent image processing and analysis key laboratory construction project under grant 2021HZSY007105; and the research project of the university laboratory work of Zhejiang Province under grant YB202128.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that they have no conflict of interest regarding the publication of this paper.

References

  1. Geng, Z.; Wang, Y. Automated design of a convolutional neural network with multi-scale filters for cost-efficient seismic data classification. Nat. Commun. 2020, 11, 3311. [Google Scholar] [CrossRef] [PubMed]
  2. Cao, L.; Wang, Z.; Wang, Z.; Wang, X.; Yue, Y. An Energy-Saving and Efficient Deployment Strategy for Heterogeneous Wireless Sensor Networks Based on Improved Seagull Optimization Algorithm. Biomimetics 2023, 8, 231. [Google Scholar] [CrossRef] [PubMed]
  3. Antunes, M.; Gomes, D.; Aguiar, R.L. Towards IoT data classification through semantic features. Future Gener. Comput. Syst. 2018, 86, 792–798. [Google Scholar] [CrossRef] [Green Version]
  4. Cao, L.; Chen, H.; Chen, Y.; Yue, Y.; Zhang, X. Bio-Inspired Swarm Intelligence Optimization Algorithm-Aided Hybrid TDOA/AOA-Based Localization. Biomimetics 2023, 8, 186. [Google Scholar] [CrossRef] [PubMed]
  5. Xu, M.; Cao, L.; Lu, D.; Hu, Z.; Yue, Y. Application of Swarm Intelligence Optimization Algorithms in Image Processing: A Comprehensive Review of Analysis, Synthesis, and Optimization. Biomimetics 2023, 8, 235. [Google Scholar] [CrossRef]
  6. Abid, M.; Mezghani, N.; Mitiche, A. Knee joint biomechanical gait data classification for knee pathology assessment: A literature review. Appl. Bionics Biomech. 2019, 2019, 7472039. [Google Scholar] [CrossRef] [Green Version]
  7. Li, Z.; Tang, J. Semi-supervised local feature selection for data classification. Sci. China Inf. Sci. 2021, 64, 192108. [Google Scholar] [CrossRef]
  8. Bai, Y.; Cao, L.; Chen, B.; Chen, Y.; Yue, Y. A Novel Topology Optimization Protocol Based on an Improved Crow Search Algorithm for the Perception Layer of the Internet of Things. Biomimetics 2023, 8, 165. [Google Scholar] [CrossRef]
  9. Bertini, R.B., Jr.; do Carmo Nicoletti, M. An iterative boosting-based ensemble for streaming data classification. Inf. Fusion 2019, 45, 66–78. [Google Scholar] [CrossRef]
  10. Lahmiri, S.; Bekiros, S.; Giakoumelou, A.; Bezzina, F. Performance assessment of ensemble learning systems in financial data classification. Intell. Syst. Account. Financ. Manag. 2020, 27, 3–9. [Google Scholar] [CrossRef]
  11. Adiwijaya; Wisesty, U.N.; Lisnawati, E.; Aditsania, A.; Kusumo, D.S. Dimensionality reduction using principal component analysis for cancer detection based on microarray data classification. J. Comput. Sci. 2018, 14, 1521–1530. [Google Scholar] [CrossRef] [Green Version]
  12. Too, J.; Mirjalili, S. General learning equilibrium optimizer: A new feature selection method for biological data classification. Appl. Artif. Intell. 2021, 35, 247–263. [Google Scholar] [CrossRef]
  13. Krawczyk, B.; Koziarski, M.; Woźniak, M. Radial-based oversampling for multiclass imbalanced data classification. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 2818–2831. [Google Scholar] [CrossRef]
  14. Bangyal, W.H.; Ahmad, J.; Rauf, H.T. Optimization of neural network using improved bat algorithm for data classification. J. Med. Imaging Health Inform. 2019, 9, 670–681. [Google Scholar] [CrossRef]
  15. Cao, L.; Wang, Z.; Yue, Y. Analysis and Prospect of the Application of Wireless Sensor Networks in Ubiquitous Power Internet of Things. Comput. Intell. Neurosci. 2022, 2022, 9004942. [Google Scholar] [CrossRef] [PubMed]
  16. Tao, X.; Li, Q.; Guo, W.; Ren, C.; Li, C.; Liu, R.; Zou, J. Self-adaptive cost weights-based support vector machine cost-sensitive ensemble for imbalanced data classification. Inf. Sci. 2019, 487, 31–56. [Google Scholar] [CrossRef]
  17. Sreejith, S.; Nehemiah, H.K.; Kannan, A. Clinical data classification using an enhanced SMOTE and chaotic evolutionary feature selection. Comput. Biol. Med. 2020, 126, 103991. [Google Scholar] [CrossRef]
  18. Cyril, C.P.D.; Beulah, J.R.; Subramani, N.; Mohan, P.; Harshavardhan, A.; Sivabalaselvamani, D. An automated learning model for sentiment analysis and data classification of Twitter data using balanced CA-SVM. Concurr. Eng. 2021, 29, 386–395. [Google Scholar] [CrossRef]
  19. Wang, J.; Lu, S.; Wang, S.H.; Zhang, Y. A review on extreme learning machine. Multimed. Tools Appl. 2022, 81, 41611–41660. [Google Scholar] [CrossRef]
  20. Zheng, W.; Shu, H.; Tang, H.; Zhang, H. Spectra data classification with kernel extreme learning machine. Chemom. Intell. Lab. Syst. 2019, 192, 103815. [Google Scholar] [CrossRef]
  21. Yahia, S.; Said, S.; Zaied, M. Wavelet extreme learning machine and deep learning for data classification. Neurocomputing 2022, 470, 280–289. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Fu, P.; Liu, W.; Chen, G. Imbalanced data classification based on scaling kernel-based support vector machine. Neural Comput. Appl. 2014, 25, 927–935. [Google Scholar] [CrossRef]
  23. Chen, C.; Li, X.; Belkacem, A.N.; Qiao, Z. The Mixed Kernel Function SVM-Based Point Cloud Classification. Int. J. Precis. Eng. Manuf. 2019, 20, 737–747. [Google Scholar] [CrossRef]
  24. Xie, Z.; Xu, Y.; Hu, Q. Uncertain data classification with additive kernel support vector machine. Data Knowl. Eng. 2018, 117, 87–97. [Google Scholar] [CrossRef]
  25. Zhang, X.; Wang, D.; Zhou, Y.; Chen, H.; Cheng, F.; Liu, M. Kernel modified optimal margin distribution machine for imbalanced data classification. Pattern Recognit. Lett. 2019, 125, 325–332. [Google Scholar] [CrossRef]
  26. Deng, Y.; Ren, Z.; Kong, Y.; Bao, F.; Dai, Q. A hierarchical fused fuzzy deep neural network for data classification. IEEE Trans. Fuzzy Syst. 2016, 25, 1006–1012. [Google Scholar] [CrossRef]
  27. Saritas, M.M.; Yasar, A. Performance analysis of ANN and Naive Bayes classification algorithm for data classification. Int. J. Intell. Syst. Appl. Eng. 2019, 7, 88–91. [Google Scholar] [CrossRef] [Green Version]
  28. Hernández, G.; Zamora, E.; Sossa, H.; Téllez, G.; Furlán, F. Hybrid neural networks for big data classification. Neurocomputing 2020, 390, 327–340. [Google Scholar] [CrossRef]
  29. Wu, H.; Prasad, S. Convolutional Recurrent Neural Networks for Hyperspectral Data Classification. Remote Sens. 2017, 9, 298. [Google Scholar] [CrossRef] [Green Version]
  30. Xi, B.; Li, J.; Li, Y.; Song, R.; Sun, W.; Du, Q. Multiscale Context-Aware Ensemble Deep KELM for Efficient Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5114–5130. [Google Scholar] [CrossRef]
  31. Chen, P.; Zhao, X.; Zhu, Q. A novel classification method based on ICGOA-KELM for fault diagnosis of rolling bearing. Appl. Intell. 2020, 50, 2833–2847. [Google Scholar] [CrossRef]
  32. Luo, F.; Liu, G.; Guo, W.; Chen, G.; Xiong, N. ML-KELM: A kernel extreme learning machine scheme for multi-label classification of real time data stream in SIoT. IEEE Trans. Netw. Sci. Eng. 2021, 9, 1044–1055. [Google Scholar] [CrossRef]
  33. Arora, S.; Singh, S. Butterfly optimization algorithm: A novel approach for global optimization. Soft Comput. 2019, 23, 715–734. [Google Scholar] [CrossRef]
  34. Mortazavi, A.; Moloodpoor, M. Enhanced Butterfly Optimization Algorithm with a New fuzzy Regulator Strategy and Virtual Butterfly Concept. Knowl. -Based Syst. 2021, 228, 107291. [Google Scholar] [CrossRef]
  35. Sultan, Y.B.; Rıza, Y.A.; İsa, A.E.; Hammoudi, A.; Sadiq, M.S.; Sujin, B. Butterfly optimization algorithm for optimum shape design of automobile suspension components. Mater. Test. 2020, 62, 365–370. [Google Scholar]
  36. Alweshah, M.; Al Khalaileh, S.; Gupta, B.B.; Almomani, A.; Hammouri, A.I.; Al-Betar, M.A. The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput. Appl. 2022, 34, 11267–11281. [Google Scholar] [CrossRef]
  37. Yue, Y.; Cao, L.; Lu, D.; Hu, Z.; Xu, M.; Wang, S.; Li, B.; Ding, H. Review and empirical analysis of sparrow search algorithm. Artif. Intell. Rev. 2023, 3, 1–53. [Google Scholar] [CrossRef]
  38. Lu, D.; Yue, Y.; Hu, Z.; Xu, M.; Tong, Y.; Ma, H. Effective detection of Alzheimer’s disease by optimizing fuzzy K-nearest neighbors based on salp swarm algorithm. Comput. Biol. Med. 2023, 159, 106930. [Google Scholar] [CrossRef] [PubMed]
  39. Zhang, Y.; Lv, Y.; Zhou, Y. Research on Economic Optimal Dispatching of Microgrid Based on an Improved Bacteria Foraging Optimization. Biomimetics 2023, 8, 150. [Google Scholar] [CrossRef] [PubMed]
  40. Liang, J.; Yan, C.; Zhang, Y.; Zhang, T.; Zheng, X.; Li, H. Rapid discrimination of Salvia miltiorrhiza according to their geographical regions by laser induced breakdown spectroscopy (LIBS) and particle swarm optimization-kernel extreme learning machine (PSO-KELM). Chemom. Intell. Lab. Syst. 2020, 197, 103930. [Google Scholar] [CrossRef]
  41. Baliarsingh, S.K.; Vipsita, S.; Muhammad, K.; Dash, B.; Bakshi, S. Analysis of high-dimensional genomic data employing a novel bio-inspired algorithm. Appl. Soft Comput. 2019, 77, 520–532. [Google Scholar] [CrossRef]
Figure 1. Flow chart of the PSOBOA-KELM algorithm.
Figure 1. Flow chart of the PSOBOA-KELM algorithm.
Biomimetics 08 00306 g001
Figure 2. Comparison of the function iteration calculation.
Figure 2. Comparison of the function iteration calculation.
Biomimetics 08 00306 g002aBiomimetics 08 00306 g002b
Figure 3. Comparison of the evaluation index parameters of the four algorithms.
Figure 3. Comparison of the evaluation index parameters of the four algorithms.
Biomimetics 08 00306 g003
Figure 4. Comparison of the simulation time consumption of the four algorithms.
Figure 4. Comparison of the simulation time consumption of the four algorithms.
Biomimetics 08 00306 g004
Figure 5. Comparison of the evaluation index parameters of the four algorithms.
Figure 5. Comparison of the evaluation index parameters of the four algorithms.
Biomimetics 08 00306 g005
Figure 6. Comparison of the simulation time consumption of the four algorithms.
Figure 6. Comparison of the simulation time consumption of the four algorithms.
Biomimetics 08 00306 g006
Table 1. Test functions.
Table 1. Test functions.
FunctionEquationDimensionBoundsOptimum
F1 i = 1 d ( j = 1 i x j ) 2 30[−100, 100]0
F2 max { | x i | , 1 i d } 30[−100, 100]0
F3 i = 1 d | x i sin ( x i ) + 0.1 x i | 30[−10, 100]0
F4 20 exp ( 0.2 1 d i = 1 d x 2 ) exp ( 1 d i = 1 d cos ( 2 π x i ) ) + 20 + exp ( 1 ) 30[−5.12, 5.12]0
F5 i = 1 4 c i exp ( j = 1 3 a i j ( x j p i j ) 2 ) 30[1, 3]−3.86
F6 i = 1 7 [ ( X a i ) ( X a i ) T + c i ] 1 30[0, 10]−10.4
F7 i = 1 10 [ ( X a i ) ( X a i ) T + c i ] 1 30[0, 10]−10.5
F8 i = 1 5 [ ( X a i ) ( X a i ) T + c i ] 1 30[0, 10]−10.1
Table 2. Detailed description of the dataset.
Table 2. Detailed description of the dataset.
NumberData SetNumber of
Categories
Sample SizeNumber of Features
1BreastEW25699
2CongressEW243516
3Hepatitis2138529
4JPNdata414818
5Parkinson219723
6SpectEW226722
7wdbc256930
Table 3. Parameter settings of the swarm intelligence optimization algorithm.
Table 3. Parameter settings of the swarm intelligence optimization algorithm.
AlgorithmParameters
PSOVmax = 4, wMax = 1, wMin = 0.9, c1 = 1.5, c2 = 2.
BBA [ f min , f max ] = [ 0 , 2 ] ; A = 0.5 ; r = 0.5 ; α = 0.9 ; γ = 0.05
BOAp = 0.8, power_exponent = 0.1, sensory_modality = 0.01.
PSOBOAp = 0.8, power_exponent = 0.1, sensory_modality = 0.01. wMax = 0.2, wMin = 0.9.
Table 4. Experimental results of four datasets: BreastEW, CongressEW, Hepatitis, and JPNdata.
Table 4. Experimental results of four datasets: BreastEW, CongressEW, Hepatitis, and JPNdata.
MetricsAlgorithmBreastEWCongressEWHepatitisJPNdata
AccuracyPSO-KELM0.955830.931820.806250.80625
BBA-KELM0.955830.909090.870830.74167
BOA-KELM0.96460.954550.870830.79286
PSOBOA-KELM0.964910.965640.878680.83958
precisionPSO-KELM0.9466610.50.76389
BBA-KELM0.94520.9610.77778
BOA-KELM0.9466610.8750.77778
PSOBOA-KELM0.9598110.78889
F-measurePSO-KELM0.965170.94340.571430.82353
BBA-KELM0.966110.923080.50.76842
BOA-KELM0.97260.961540.666670.80065
PSOBOA-KELM0.972970.970970.708330.84034
SensitivityPSO-KELM0.986110.92450.666670.875
BBA-KELM10.925930.333330.86607
BOA-KELM10.925930.583330.875
PSOBOA-KELM10.943730.583330.9375
SpecificityPSO-KELM0.9090910.881410.73214
BBA-KELM0.904760.9411810.71429
BOA-KELM0.90476110.75
PSOBOA-KELM0.92857110.75
MCCPSO-KELM0.906220.869230.452270.64569
BBA-KELM0.907310.815970.5360.57071
BOA-KELM0.925470.91020.618990.58872
PSOBOA-KELM0.925820.931580.663870.68104
Table 5. Experimental results of three datasets: Parkinson, SpectEW, and Wdbc.
Table 5. Experimental results of three datasets: Parkinson, SpectEW, and Wdbc.
MetricsAlgorithmParkinsonSpectEWWdbc
AccuracyPSO-KELM0.846050.814560.95659
BBA-KELM0.821050.79670.9469
BOA-KELM0.8750.814810.96491
PSOBOA-KELM0.90.846150.97383
precisionPSO-KELM0.899040.583331
BBA-KELM0.8333301
BOA-KELM0.874510.633331
PSOBOA-KELM0.928570.8751
F-measurePSO-KELM0.896060.365080.93841
BBA-KELM0.8777800.92308
BOA-KELM0.92330.392860.95
PSOBOA-KELM0.93750.333330.964
SensitivityPSO-KELM0.930950.266670.8842
BBA-KELM0.9285700.85714
BOA-KELM10.283330.90476
PSOBOA-KELM10.20.93074
SpecificityPSO-KELM0.6750.953461
BBA-KELM0.411
BOA-KELM0.611
PSOBOA-KELM0.7511
MCCPSO-KELM0.599940.303640.90873
BBA-KELM0.4615500.88901
BOA-KELM0.652480.400050.92582
PSOBOA-KELM0.728090.409880.94514
Table 6. Experimental results of the four datasets of Australian, Breastcancer, Dermatology, and HeartEW.
Table 6. Experimental results of the four datasets of Australian, Breastcancer, Dermatology, and HeartEW.
MetricsAlgorithmAustralianBreastcancerDermatologyHeartEW
AccuracyPSO-KELM0.833330.857490.775330.85185
BBA-KELM0.955830.909090.870830.74074
BOA-KELM0.927020.93750.916670.85185
PSOBOA-KELM0.944290.985710.972970.92593
precisionPSO-KELM0.740740.733330.785710.70314
BBA-KELM0.851850.833330.846150.77778
BOA-KELM0.925930.838710.8750.88889
PSOBOA-KELM0.941180.978260.944440.92857
F-measurePSO-KELM0.666670.777780.632460.65327
BBA-KELM0.857490.833330.775330.77778
BOA-KELM0.866670.916670.928570.89655
PSOBOA-KELM0.923080.989010.972220.9375
SensitivityPSO-KELM0.73370.777780.703140.8325
BBA-KELM0.827460.833330.857490.87607
BOA-KELM0.882350.962960.916670.90909
PSOBOA-KELM0.923080.978260.944440.93427
SpecificityPSO-KELM0.790570.775330.833330.82353
BBA-KELM0.8750.857490.851850.92857
BOA-KELM0.927020.916670.925930.92702
PSOBOA-KELM0.967740.958420.941180.93667
MCCPSO-KELM0.785710.869230.703140.64569
BBA-KELM0.907310.903230.866670.7532
BOA-KELM0.882350.933330.888890.84615
PSOBOA-KELM0.908890.969090.953660.85749
Table 7. Experimental results tested in the Diabetes, Glass, Heart, and Vote data sets.
Table 7. Experimental results tested in the Diabetes, Glass, Heart, and Vote data sets.
MetricsAlgorithmDiabetesGlassHeartVote
AccuracyPSO-KELM0.815710.76190.666670.68182
BBA-KELM0.888890.857140.772730.72127
BOA-KELM0.916670.887270.819050.80952
PSOBOA-KELM0.944440.972220.878680.88958
PrecisionPSO-KELM0.652170.695650.714290.72853
BBA-KELM0.727270.783810.857140.83636
BOA-KELM0.862730.896670.88520.87273
PSOBOA-KELM0.92190.951820.941820.91952
F-measurePSO-KELM0.719050.754290.739130.80952
BBA-KELM0.779050.809530.792730.86667
BOA-KELM0.88950.878180.895450.89143
PSOBOA-KELM0.945450.966670.974290.97381
SensitivityPSO-KELM0.681820.775330.857490.83333
BBA-KELM0.779020.833330.879020.90235
BOA-KELM0.916670.925930.93750.92702
PSOBOA-KELM0.957490.973730.961890.98774
SpecificityPSO-KELM0.666670.775330.593330.78571
BBA-KELM0.733330.790570.758620.85743
BOA-KELM0.882350.916670.908190.91751
PSOBOA-KELM0.962960.943230.97750.96774
MCCPSO-KELM0.775330.813630.703140.78235
BBA-KELM0.85230.882350.857490.88889
BOA-KELM0.906670.913210.812790.92702
PSOBOA-KELM0.933330.967740.972960.98774
Table 8. Comparison of the Sinc function fitting results.
Table 8. Comparison of the Sinc function fitting results.
PSO-KELMBBA-KELMBOA-KELMPSOBOA-KELM
IterationRMSEMAERSDRMSEMAERSDRMSEMAERSDRMSEMAERSD
50.0270.0240.9150.0210.0160.9290.0170.0150.9450.0140.0120.963
100.0250.0230.9280.0190.0140.9350.0150.0130.9510.0120.0080.969
150.0210.0210.9350.0180.0130.9420.0130.0120.9570.0110.0070.973
200.0190.0190.9410.0160.0120.9490.0120.0110.9630.010.0060.98
250.0180.0170.9450.0150.0110.9530.0110.010.9670.010.0060.984
300.0150.0150.9490.0130.010.9580.010.0090.9710.0090.0050.988
350.0130.0140.9530.0110.0090.9620.0090.0080.9750.0090.0050.99
400.0120.0130.9580.010.0090.9640.0090.0080.9780.0070.0050.991
450.0120.0110.9610.0090.0090.9680.0080.0070.9810.0060.0040.992
500.0120.0110.9610.0090.0090.9710.0070.0060.9830.0060.0040.993
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yue, Y.; Cao, L.; Chen, H.; Chen, Y.; Su, Z. Towards an Optimal KELM Using the PSO-BOA Optimization Strategy with Applications in Data Classification. Biomimetics 2023, 8, 306. https://doi.org/10.3390/biomimetics8030306

AMA Style

Yue Y, Cao L, Chen H, Chen Y, Su Z. Towards an Optimal KELM Using the PSO-BOA Optimization Strategy with Applications in Data Classification. Biomimetics. 2023; 8(3):306. https://doi.org/10.3390/biomimetics8030306

Chicago/Turabian Style

Yue, Yinggao, Li Cao, Haishao Chen, Yaodan Chen, and Zhonggen Su. 2023. "Towards an Optimal KELM Using the PSO-BOA Optimization Strategy with Applications in Data Classification" Biomimetics 8, no. 3: 306. https://doi.org/10.3390/biomimetics8030306

Article Metrics

Back to TopTop