Next Article in Journal
Relationships between Plasma Concentrations of Testosterone and Dihydrotestosterone and Geriatric Depression Scale Scores in Men and Women Aged 60–65 Years—A Multivariate Approach with the Use of Quade’s Test
Previous Article in Journal
Knowledge Mapping and Institutional Prospects on Circular Carbon Economy Based on Scientometric Analysis
Previous Article in Special Issue
Triaging Medical Referrals Based on Clinical Prioritisation Criteria Using Machine Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Technology-Oriented Pathway for Auxiliary Diagnosis in the Digital Health Age: A Self-Adaptive Disease Prediction Model

1
School of Business and Management, Jilin University, Changchun 130012, China
2
Information Resource Research Center, Jilin University, Changchun 130012, China
*
Authors to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2022, 19(19), 12509; https://doi.org/10.3390/ijerph191912509
Submission received: 11 August 2022 / Revised: 23 September 2022 / Accepted: 27 September 2022 / Published: 30 September 2022

Abstract

:
The advent of the digital age has accelerated the transformation and upgrading of the traditional medical diagnosis pattern. With the rise of the concept of digital health, the emerging information technologies, such as machine learning (ML) and data mining (DM), have been extensively applied in the medical and health field, where the construction of disease prediction models is an especially effective method to realize auxiliary medical diagnosis. However, the existing related studies mostly focus on the prediction analysis for a certain disease, using models with which it might be challenging to predict other diseases effectively. To address the issues existing in the aforementioned studies, this paper constructs four novel strategies to achieve a self-adaptive disease prediction process, i.e., the hunger-state foraging strategy of producers (PHFS), the parallel strategy for exploration and exploitation (EEPS), the perturbation–exploration strategy (PES), and the parameter self-adaptive strategy (PSAS), and eventually proposes a self-adaptive disease prediction model with applied universality, strong generalization ability, and strong robustness, i.e., multi-strategies optimization-based kernel extreme learning machine (MsO-KELM). Meanwhile, this paper selects six different real-world disease datasets as the experimental samples, which include the Breast Cancer dataset (cancer), the Parkinson dataset (Parkinson’s disease), the Autistic Spectrum Disorder Screening Data for Children dataset (Autism Spectrum Disorder), the Heart Disease dataset (heart disease), the Cleveland dataset (heart disease), and the Bupa dataset (liver disease). In terms of the prediction accuracy, the proposed MsO-KELM can obtain ACC values in analyzing these six diseases of 94.124%, 84.167%, 91.079%, 72.222%, 70.184%, and 70.476%, respectively. These ACC values have all been increased by nearly 2–7% compared with those obtained by the other models mentioned in this paper. This study deepens the connection between information technology and medical health by exploring the self-adaptive disease prediction model, which is an intuitive representation of digital health and could provide a scientific and reliable diagnostic basis for medical workers.

1. Introduction

With the rapid development of the economy and technology, public demands for improving healthcare are getting stronger and how to utilize information technology to achieve auxiliary diagnosis has received increasing social attention [1,2,3,4]. Moreover, the spread of the concept of digital health [5,6,7] promotes deep integration between information technology and healthcare, where a large number of machine learning (ML) models and data mining (DM) methods have been introduced into the traditional medical diagnosis pattern. To date, there are various existing studies adopting the technologies of ML and DM to predict diseases, such as predicting stable MCI patients [8], forecasting nuanced yet significant MT errors of clinical symptoms [9], survival risk prediction for esophageal cancer [10], conducting breast cancer diagnosis [11], preconception prediction for gestational diabetes mellitus [12], predicting Alzheimer’s disease [13], and heart disease prediction [14,15].
Disease prediction analysis based on ML and DM is a research trend in medical informatics. These research findings could provide a scientific and reliable diagnostic basis for medical workers and provide an effective technical support for the early intervention of related diseases. For example, Derevitskii et al. construct a hybrid predictive modelling for Thyrotoxic atrial fibrillation, which could be used as part of a decision support system for medical staff who work with thyrotoxicosis patients [16]. Similarly, Muhammad et al. develop a machine leaning predictive model for coronary artery disease (CAD), which could be used to develop an expert system for diagnosis of CAD patients in Nigeria [17]. Nevertheless, these aforementioned studies still contain a critical limitation, i.e., the existing proposed models mainly focus on a certain disease and mostly neglect expanding the application’s universality for predicting various other diseases, which means that these models may not obtain accurate prediction results in analyzing those other diseases. Furthermore, a model with better performance when predicting different diseases would be more significant for the medical workers and the auxiliary diagnosis process. Therefore, constructing a disease prediction model with self-adaptive ability, strong generalization ability, and strong robustness is the motivation to explore the technology-oriented pathway for auxiliary diagnosis in digital health age.
There may be large differences among the disease data from the real world, such as different attribute dimensions and different inner structures. In fact, an ideal disease prediction model which meets current medical needs should combine prediction accuracy with broad applicability. To this end, we take the kernel extreme learning machine (KELM) as the base model, which has advantages in generalization ability and robustness, and we introduce an improved swarm intelligence optimization algorithm to optimize the KELM, i.e., sparrow search algorithm (SSA) with the enhanced global searching ability (EGSSA). Finally, we design a novel disease prediction model, i.e., multi-strategies optimization-based kernel extreme learning machine (MsO-KELM). The main contributions and innovations of the MsO-KELM are highlighted as follows:
(1) To effectively predict various diseases, we utilize the EGSSA to optimize the base model by designing four novel strategies, i.e., the hunger-state foraging strategy of producers (PHFS), the parallel strategy for exploration and exploitation (EEPS), the perturbation–exploration strategy (PES), and the parameter self-adaptive strategy (PSAS). These strategies can enhance the prediction accuracy of the model and allow the model to be applied to various diseases.
(2) To verify the prediction performance of the proposed MsO-KELM, we adopt six different disease datasets as the experimental samples, consisting of the Breast Cancer dataset (cancer), the Parkinson dataset (Parkinson’s disease), the Autistic Spectrum Disorder Screening Data for Children dataset (Autism Spectrum Disorder), the Heart Disease dataset (heart disease), the Cleveland dataset (heart disease), and the Bupa dataset (liver disease), and we evaluate the prediction performance by four different evaluation metrics, i.e., the ACC, the sensitivity, the specificity, and the MCC. Notably, the ACC is the most significant metric to evaluate the prediction accuracy.
(3) To elaborate the details of the MsO-KELM, we conduct two-stage experiments in this paper. The first experiment is mainly to prove the better optimization performance of the EGSSA, which is the basis for achieving the self-adaptive characteristic of the MsO-KELM. The second experiment is mainly to compare the MsO-KELM with other state-of-the-art prediction models.
The self-adaptive prediction process of the MsO-KELM in analyzing the different diseases is shown in Figure 1.

2. Preparation

2.1. Disease Data Description

In this paper, we select six different real-world disease datasets as the experiment samples (these data are available at https://archive.ics.uci.edu/mL/index.php, accessed on 6 June 2022), i.e., the Breast Cancer dataset (cancer), the Parkinson dataset (Parkinson’s disease), the Autistic Spectrum Disorder Screening Data for Children dataset (Autism Spectrum Disorder), the Heart Disease dataset (heart disease), the Cleveland dataset (heart disease), and the Bupa dataset (liver disease). The reasons for selecting these six disease datasets are shown as follows:
  • They are the common diseases in the real world;
  • These disease data are extensively utilized by numerous investigators;
  • These disease data have different internal structures and different diagnosis indicators.
The characteristics of these six disease datasets are shown in Table 1.

2.2. The Base Disease Classifier—Kernel Extreme Learning Machine

The KELM is an excellent classifier which has advantages in generalization ability and learning speed [18,19,20,21,22]. Since its emergence, the KELM has been extensively studied by numerous investigators for problems such as hyperspectral image classification [23], data classification in enterprise cloud data [24], time-varying distributed parameter systems [25], and intrusion detection [26]. Notably, there are two significant parameters in the original KELM, i.e., the k value (kernel parameter) and the c value (regularization coefficient), which easily fall into local optimum in the original searching process [27,28] and can influence the final prediction accuracy. The main calculation processes of KELM are shown as follows:
Objective _ function _ ELM = h i d ( x ) × H _ o u t p T × ( I _ m a t r i x C o e f f i c i e n t _ r + H _ o u t p × H _ o u t p T ) 1 × L
ker nel _ matrix = H _ outp × H _ outp T = hid ( x i ) × hid ( x j ) = k ( x i ,   x j )
Objective _ function _ KELM = [ k ( x ,   x 1 ) k ( x ,   x n ) ] T × ( I _ m a t r i x C o e f f i c i e n t _ r + ker n e l _ m a t r i x ) 1 × L
where the Objective_function_ELM and Objective_function_KELM indicate the learning objective function of ELM and that of KELM, respectively. The hid(x) and H_outp represent the feature mapping matrix of hidden layer [18], I_matrix indicates the identity matrix, Cofficient_r indicates the regularization coefficient, L indicates the expectation matrix, and k(xi, xj) indicates the kernel function. In this paper, we adopt the Gaussian kernel function, and the kernel parameter k indicates the kernel width.

2.3. The Base Optimization Tool—Sparrow Search Algorithm

In order to improve the prediction performance of the base-classifier, we take SSA as a base-optimizer and conduct the improvements on SSA to design a more effective optimizer, i.e., the EGSSA. As a recent meta-heuristic algorithm [29], SSA has been applied in many real-world problems [30,31,32,33,34] because of the advantages in convergence speed and exploitation ability. There are two significant population roles in the optimization process of SSA, i.e., producers and scroungers, in which the producers could be regarded as the leader with a higher fitness [30]. The location update processes of these two roles are shown as follows:
posi i ,   d t + 1 = { p o s i i ,   d t exp ( i α i t e r max ) ,   R   <   S T p o s i i , ʠ d t + R andom M a t r i x ,   R     S T
posi i ,   d t + 1 = { R a n d o m exp ( p o s i w o r s t p o s i i ,   d t i 2 ) ,   i   >   n / 2 p o s i a t + 1 + | p o s i i ,   d t p o s i a t + 1 | V + M a t r i x ,   i     n / 2
posi i ,   d t + 1 = { p o s i best t + β | posi i ,   d t p o s i best t |   f i t n e s s i   >   f i t n e s s g posi i ,   d t + ψ   ( | posi i ,   d t p o s i worst t | ( f i t n e s s i f i t n e s s w ) + ξ )   f i t n e s s i   =   f i t n e s s g
Equation (4) describes the location update process of producer, the posi i ,   d t + 1 indicates the current location of the ith sparrow individual in the dth dimensional space when the populations are carrying out the tth iteration [30], t and i indicate the current iterations and current sparrow individual, respectively. The α, R, and Random are the random parameters set manually, the ST is a warning threshold, the values of which could be set in (0.5, 1) [30]. The Matrix indicates a row vector where each element value is set to 1 and the dimensions are d [30]. Equation (5) describes the location update process of scrounger, the p o s i a t + 1 indicates the optimal location searched by producer, the posiworst indicates the worst location in the current iteration [30], the n is the sparrow population size, and the V indicates a row vector where each element value is set to 1 or −1 randomly and the dimensions are d. Equation (6) describes the location update process of detection sparrow, the p o s i b e s t t indicates the optimal location when the populations are carrying out the tth iteration, and the β, ψ, and ξ indicate adjustment parameters [30].

2.4. The Introduction of Evaluation Metrics

To evaluate the disease prediction performance of the proposed MsO-KELM, we adopt the four evaluation metrics in this paper, i.e., classification accuracy (ACC, the value range is from 0 to 1.0) [18], sensitivity (the value range is from 0 to 1.0) [18], specificity (the value range is from 0 to 1.0) [18], and Mathews correlation coefficient (MCC, the value range is from −1.0 to 1.0) [18]. The ACC emphasizes the number of samples which are correctly predicted (the most significant evaluation metric to measure the classification performance of a model), the sensitivity shows the ability to correctly predict a positive sample among all positive samples, the specificity indicates the ability to correctly predict a negative sample among all negative samples, and the MCC mainly shows the reliability of a model (the closer the MCC value is to 1, the more accurate and effective the model is). The calculation processes of these four evaluation metrics are shown as follows:
A C C = T P + T N T P + T N + F P + F N
s e n s i t i v i t y = T P T P + F N
s p e c i f i c i t y = T N T N + F P
M C C = T P × T N F P × F N ( T N + F P ) × ( T N + F N ) × ( T P + F N ) × ( T P + F P )
where the TP indicates the number of samples with the positive prediction result and the positive label, the TN indicates that the number of samples with the negative prediction result and the negative label, the FP indicates that the number of samples with the positive prediction result and the negative label, and the FN indicates that the number of samples with the negative prediction result and the positive label.

3. The Proposed Methodology

The core idea of the proposed methodology is to construct a self-adaptive disease prediction model with high accuracy, strong generalization ability, and strong robustness. Therefore, we select an excellent base-classifier (KELM) and design an enhanced meta-heuristic algorithm (EGSSA) as the optimizer to finally construct the self-adaptive disease prediction model, i.e., the MsO-KELM. Specifically, there are four novel strategies in the MsO-KELM, i.e., the hunger-state foraging strategy of producers (PHFS), the parallel strategy for exploration and exploitation (EEPS), the perturbation–exploration strategy (PES), and the parameter self-adaptive strategy (PSAS), where the EGSSA consists of the PHFS, the EEPS, and the PES. In addition, the PSAS will act on a parameter acquisition mechanism which is formed by combining the EGSSA with the KELM. In this section, the technological details of the MsO-KELM will be discussed as follows:

3.1. Foraging Strategy of Producers in Hunger-State (PHFS)

In the original SSA, the producers could be regarded as the leader roles in the sparrow populations, which are responsible for searching for food-rich positions and providing foraging directions for all scroungers, and the scroungers could follow the producers to achieve the foraging process. In that case, if the producers could expand the searching range to find a safer and more adequate position, it would provide more possibilities for scroungers to improve the foraging rate and finally enhance the global convergence performance.
However, the existing location update mechanism mainly focuses on the exploitation ability (local searching ability) of the original SSA, which may cause the producers trapping into the local optimum situation. To address this issue and enhance the exploration ability (global searching ability), we introduce the hunger games search algorithm (HGS) [35,36] into the original SSA to construct the PHFS for expanding the searching range of optimal position. Specifically, the PHFS is a hybrid strategy, which retains the advantage in exploitation ability of the original SSA while combining the exploration approach of the HGS algorithm with the location update mechanism of the producers.
In the original SSA, the location update process of producers is shown as Equation (4), and the key calculation function affecting the convergence efficiency is shown in Equation (11):
function ( i ) = exp ( i α i t e r max )
where the α and itermax are the parameters set manually. It is clearly shown that Equation (11) has a descending trend and will eventually converge to 0, which means the producers can easily repeat the searching behavior at a certain position with the number of individuals increasing, and eventually fall into a local optimum. By contrast, the HGS has a significant advantage in exploration ability, and the hungry feature function affecting the convergence efficiency is shown in Equation (12):
function ( j ) = 1 exp ( | p a r _ m a n u a l j | )
where the par_maunal is a parameter set manually. It is clearly shown that Equation (12) is not a single ascending or descending trend, but the function trend is affected by the parameter value. When the inputting value is less than the par_maunal value, the function could have an ascending trend with the inputting value increasing. Moreover, when the inputting value is larger than the par_maunal value, the function could have a descending trend with the inputting value increasing. Therefore, introducing the hungry feature of HGS into the searching behavior of producers could expand the searching range of optimal position, and ultimately enhance the exploration ability of SSA.
The searching processes based on Equations (11) and (12) are shown in Figure 2.
According to Figure 2, we could find that the original location update method can easily search in a local space (like the red region in Figure 2a), but the hungry roles of HGS could search in a relatively global region (like Figure 2b). Therefore, the PHFS could be described as follows:
posi _ new i ,   d t + 1 = ( 1 exp ( | p o s i _ c u r r e n t b e s t p o s i _ c u r r e n t i ,   d t | ) ) rand ο
where the meaning of posi _ new i ,   d t + 1 is similar to that of posi i ,   d t + 1 in Equation (4), the p o s i _ c u r r e n t i , d t indicates the location of last iteration, the rand is a random number, and the o is an adjustment parameter which is set to 2 in this paper.

3.2. Parallel Strategy for Exploration and Exploitation (EEPS)

As shown in Equation (4), we could find that the location update process of producers consists of two different stages. The PHFS acts on stage 1 (R < ST), and the EEPS described below is going to act on stage 2 (R ≥ ST). According to the principle of SSA, the producers would move to a safer place when perceiving danger coming. In the new location update process, the producers would search globally with a normally distributed random manner and eventually converge to the optimal position. Nevertheless, there are still some limitations in this moving process, i.e., (i) the existing searching range could continue to be expanded; and (ii) the single global searching process could affect the convergence accuracy. Therefore, we design the EEPS to balance the exploration process and exploitation process in this paper. In fact, the balancing effect of EEPS is able to dynamically adjust the position searching approach of the producers, i.e., by promoting the producers exploring the whole space with a global searching approach for expanding the searching range of the potential optimal position in the early stage, while exploiting the current area with a local searching approach when a certain area is close to the optimal position.
Similarly, inspired by the literature [37,38] related to HGS, we introduce the pattern of food-approaching into the location update process of producers in stage 2, and construct the balance factor shown in Equation (14):
b a l a n c e _ f a c t o r i t e r = ( 2 p a r a 1 ) × ( δ × ( 1 i t e r i t e r _ max ) )
where the meaning of para is a random number in the range of (0, 1), the δ indicates a control parameter (it is set to 2), the iter indicates the current number of iterations, and the iter_max indicates the maximum number of iterations.
Subsequently, the EEPS, which combines the balance factor with the original location update mechanism, is shown in Equation (15):
p o s i i ,   d t + 1 = ( ( 2 p a r a 1 ) × ( δ × ( 1 i t e r i t e r _ max ) ) ) ( p o s i best t p o s i i ,   d t ) + Random M a t r i x   R     S T
where the meaning of the parameters are similar to Equations (4), (6) and (14).
The comparison results of the original searching process and the searching process based on after-EEPS are shown in Figure 3a,b, respectively.
In Figure 3, the points in the same color are the positions of all producers in one iteration, while different colors indicate different iterations. Figure 3a shows that the original location update mechanism of producers in stage 2 could achieve a global searching process to some extent, but it lacks the local exploitation behavior and the global searching range is not large enough, which can finally affect the convergence accuracy. As a comparison, Figure 3b shows that the EEPS not only expands the global searching range, but retains the exploitation ability, which can be clearly seen in the color changing process of points.

3.3. Perturbation–Exploration Strategy (PES)

Notably, the PHFS and EEPS only enhance the exploration ability of producer populations, which means that there remains the possibility of trapping into local optimum for the whole sparrow populations. Therefore, we introduce the Cauchy distribution operator [39,40,41] to construct the PES to further expand the searching range and avoid the local optimum situation in late iterations.
In the PES, we adopt the Cauchy distribution operator to obtain a variant of the current optimal individual. Moreover, we compare the fitness of the current optimal individual with that of the variant to save the better solution. In this paper, the probability density function of Cauchy distribution operator is shown as follows:
c a u c h y ( k ) = 1 π a a + k 2
where the a is a parameter (in this paper, the a is equal to 1), and the k indicates a variable, the value range of which is from negative infinity to positive infinity. Based on the Cauchy distribution operator, we could obtain the perturbation–exploration process as follows:
{ r = tan ( ( random 0 . 5 ) × π ) posi b e s t n e w = p o s i b e s t t + r × p o s i b e s t t
where the r is a Cauchy distribution random variable generating function.

3.4. Parameter Self-Adaptive Strategy (PSAS)

As we all know, obtaining the appropriate parameter values (k and c) is most significant for the original KELM. However, the existing method of parameter taking is to adopt grid searching, which not only increases the computational cost but also affects the final results of parameter obtaining. To achieve a self-adaptive process of these two parameters, we propose the PSAS via combining the EGSSA with the KELM.
Specifically, there are four significant stages in the PSAS, and the technical details are as follows: (i) achieving the location initialization process of sparrow populations (this paper adopts the random generation method of the original SSA); (ii) the core parameters (k and c) could be automatically obtained by adopting the EGSSA; (iii) to eliminate the randomness of these two obtained parameters, this paper utilizes 10-fold cross-validation [18] to re-obtain the optimal parameter values; and (iv) the re-obtained optimal parameter values are introduced into the original KELM, and finally the MsO-KELM is formed to perform the prediction performance on six different disease datasets based on 10-fold cross-validation.
In summary, the novel MsO-KELM model utilizes the four major strategies, i.e., PHFS, EEPS, PES, and PSAS, to achieve improved prediction performance, and the specific details of the MsO-KELM are shown in Algorithm 1.
Algorithm 1: MsO-KELM model
Ijerph 19 12509 i001

4. Results

In this section, we mainly analyze the results of two experiments, i.e., the performance analysis of the EGSSA and the disease prediction evaluation of the MsO-KELM. To ensure the reasonableness of these two experiments, we set the population size and the maximum iterations of each experiment group to (30, 500) and (10, 50), respectively. Importantly, each algorithm in experiment 1 runs 30 times independently on each function to minimize the effects of algorithmic randomness [30]. Moreover, all experiments are conducted in the same running environment, i.e., Intel Core i7, 2.40 GHz, 8 GB RAM, MATLAB2021a, and the IBM SPSS Statistics 20.

4.1. Experiment 1: The Performance Analysis of EGSSA

In this paper, we select nine different swarm intelligence algorithms to conduct the comparison analysis, which contain seven classical algorithms, i.e., SSA [29], PSO [42], GWO [43], HHO [44], LSA [45], WOA [46], and FPA [47], and two classical swarm intelligence variants, i.e., SCACSSA [48] and HHOHGSO [49]. The parameters which exist in both EGSSA and SSA are set to the same values, and the specific parameter settings in this experiment are shown in Table 2.

4.1.1. Performance Analysis Based on 23 Classical Benchmark Functions

The 23 classical benchmark functions are frequently adopted to evaluate the performance of optimization algorithms, which could be divided into three categories, i.e., the unimodal functions, the multimodal functions, and the fixed-dimension multimodal functions [30]. The specific characteristics of these 23 benchmark functions are shown in Table 3, and the experimental results of the aforementioned algorithms based on the 23 functions are shown in Table 4.
In this experiment, considering that all the swarm intelligence optimization algorithms should perform multiple iterations, we utilize the mean value (avg) and the standard deviation value (std) to measure the performance [30], where the avg value is the key metric. The avg value being closer to the optimal value means the algorithm has a better performance in solving the current function. In addition, the std value could show the stability of this algorithm in solving the current function. Notably, when there are the same avg values between two algorithms, the std value could be the second evaluation metric.
According to the results in Table 4, we find that the EGSSA could obtain the best results in solving the benchmark functions F1, F2, F3, F4, F5, F6, F9, F10, F11, F12, F13, F16, F17, F18, F19, F20, F21, F22, and F23 among all 10 compared algorithms. Furthermore, in solving the unimodal functions F1-F4, the EGSSA not only obtains the optimal values, but also has the most stable performance. In solving the multimodal functions F9, F11, and the fixed-dimension multimodal functions F16-F19, the EGSSA can obtain the optimal values, while solving the multimodal functions F10, F12, F13, and the fixed-dimension multimodal functions F20–F23, the EGSSA can obtain the best results among all 10 compared algorithms and show more stable performance than others.
Apart from the comparison results table, we could verify the validity of EGSSA based on the convergence curve figure. In Figure 4, the F5 and F6 are unimodal functions, the F12 and F13 are multimodal functions, and the F22 and F23 are the fixed-dimension multimodal functions. According to Figure 4, it indicates that the EGSSA can obtain the fastest convergence among the 10 compared algorithms in solving the benchmark functions F6, F12, F13, F22, and F23. When solving the function F5, the convergence speed of EGSSA is similar to that of HHO, SSA, and SCASCSSA during the first 100 iterations, which is significantly superior to others, and the EGSSA can obtain the fastest convergence among the 10 compared algorithms after 100 iterations. In summary, the EGSSA enhances the overall convergence performance compared with the original SSA, other classical swarm intelligence optimization algorithms, and other variants. Therefore, the EGSSA could provide good support for the construction of the subsequent disease prediction model.

4.1.2. Statistical Test

To show the statistical significance of the proposed EGSSA, we introduce the Wilcoxon rank-sum test into this section [30]. The better one of any two compared algorithms could be identified by comparing the obtained significance level value (p-value) with 0.05. The p-value being less than 0.05 indicates that the former (the algorithm to be verified) has more significant advantages than the latter (the algorithm being compared); otherwise, there is no significant difference between the latter and the former. The results of the statistical test are shown in Table 5.
According to Table 5, we find that the EGSSA could obtain 16 better statistical test results among 23 benchmark functions compared with the original SSA; meanwhile, it could also obtain 7 statistical test results which are equal to those of the original SSA. For the comparison results between the EGSSA and the other six classical swarm intelligence algorithms, it is clearly shown that the p-values of EGSSA versus PSO, EGSSA versus GWO, EGSSA versus HHO, EGSSA versus LSA, EGSSA versus WOA, and EGSSA versus FPA are much less than 0.05, and the EGSSA could obtain 16–19 better benchmark function results than the aforementioned compared algorithms. When comparing the EGSSA with the HHO, although the p-value is larger than 0.05, the number of benchmark functions which could obtain better results is still significantly higher than with the HHO. Similarly, for the comparison results between the EGSSA and the SCACSSA, we could find that the EGSSA has significant advantages compared with SCACSSA, and the former could obtain 19 better solving results among 23 benchmark functions while it could also obtain four statistical test results which are equal to those of the SCACSSA. When comparing the EGSSA with the HHOHGSO, although the p-value is larger than 0.05, the number of benchmark functions for which better results could be obtained is still significantly higher than the HHOHGSO.

4.2. Experiment 2: The Prediction Analysis in Solving Different Disease Datasets

This section is mainly to verify the prediction performance of the proposed MsO-KELM in solving the aforementioned six disease datasets. Furthermore, we select another five different optimization algorithms-based KELM variants to be the compared group, i.e., GWO-KELM, HHO-KELM, FPA-KELM, WOA-KELM, and SSA-KELM. The prediction results of different algorithms on four different evaluation metrics are shown in Table 6, Table 7, Table 8, Table 9, Table 10 and Table 11 and Figure 5, Figure 6 and Figure 7.

4.2.1. Data Pre-Processing and Parameter Settings

Considering that the number of samples with missing values is about 1–3% of the sample volume of the relevant dataset, we therefore remove these samples with missing values. Meanwhile, we normalize all attributes to the interval form −1 to 1. To obtain a fair comparison result for these six different KELM variants, we set the same population size and maximum iteration numbers for all compared algorithms, to 10 and 50, respectively. In addition, to effectively estimate the generalization ability of the proposed model, we utilize 10-fold cross-validation to divide the original data samples into 10 groups, where each group can be used separately as a testing set, and the other 9 groups can be used as a training set, and, in that case, the 10 models can be obtained. These 10 models are evaluated in the 10 testing sets, respectively, and the final cross-validation result is obtained by summing and averaging the results of the 10 models. For the other special parameters, we can set the values based on Table 2.

4.2.2. Experiment Results Analysis

The prediction results in solving Breast Cancer via 10-fold cross-validation method are shown in Table 6. Based on Table 6 and Figure 5a, it is clearly shown that the proposed MsO-KELM could obtain the best average values of ACC, sensitivity, specificity, and MCC among all six compared algorithms, at 94.124%, 94.434%, 94.539%, and 87.099%, respectively. As a comparison, the GWO-KELM obtains the second best average values of these four evaluation metrics at 93.695%, 92.628%, 94.212%, and 86.068%, respectively. In addition, the HHO-KELM, the WOA-KELM, and the SSA-KELM all obtain the worst ACC values, specificity values, and MCC values at 92.237%, 93.944%, and 82.433%, respectively, but the FPA-KELM obtains the worst sensitivity value at 88.539%. According to these evaluation metric results, we could find that the proposed MsO-KELM enhances the overall prediction performance and outperforms the other optimization algorithm-based KELM models in predicting breast cancer; therefore, the MsO-KELM could be adopted in the early detection of breast cancer.
The prediction results in solving Parkinson via 10-fold cross-validation method are shown in Table 7. Based on Table 7 and Figure 5b, it is clearly shown that the proposed MsO-KELM could obtain the best average values of ACC, sensitivity, specificity, and MCC among all six compared algorithms at 84.167%, 89.596%, 67.571%, and 57.140%, respectively. As a comparison, the FPA-KELM could obtain the second best ACC value at 82.833% and the second best MCC value at 54.365%; the HHO-KELM and the WOA-KELM could obtain the second best sensitivity value at 88.423% and the second best specificity value at 67.500%, respectively. In addition, the SSA-KELM obtained the worst ACC value, specificity value, and MCC value at 80.333%, 61.500%, and 47.749%, respectively, but the GWO-KELM obtained the worst sensitivity value at 85.991%. According to these evaluation metric results, we could find that the proposed MsO-KELM has better prediction performance than other compared models for Parkinson’s disease.
The prediction results in solving Autistic Spectrum Disorder Screening Data for Children via 10-fold cross-validation method are shown in Table 8. Based on Table 8 and Figure 6a, it is clearly shown that the proposed MsO-KELM could obtain the best average values of ACC and MCC among all six compared algorithms at 91.079% and 82.899%, respectively. As a comparison, the WOA-KELM obtained the second best ACC values at 89.433% and the FPA-KELM obtained the second best MCC values at 80.578%. In addition, for the sensitivity value and specificity value, the MsO-KELM could obtain the results at 98.424% and 83.379%, respectively, which are close to the best values for these two metrics. According to these evaluation metric results, we could find that the proposed MsO-KELM has the best prediction accuracy among these compared models in the diagnosis of autism spectrum disorders.
The prediction results in solving Heart Disease via 10-fold cross-validation method are shown in Table 9. Based on Table 9 and Figure 6b, it is clearly shown that the proposed MsO-KELM could obtain the best average values of ACC, sensitivity, and MCC among all six compared algorithms at 72.222%, 68.853%, and 44.557%, respectively. As a comparison, the FPA-KELM obtained the second best ACC value, sensitivity value, and MCC values at 70.741%, 63.198%, and 41.245%, respectively. In addition, the SSA-KELM obtained the worst ACC value, sensitivity value, and MCC value at 66.296%, 42.951%, and 32.280%, respectively. For the sensitivity value, the MsO-KELM could obtain a result which is close to that of the other models. According to these evaluation metric results, we could find that the proposed MsO-KELM has more advantages and could obtain a better classification performance than the other compared models.
The prediction results in solving Cleveland via 10-fold cross-validation method are shown in Table 10. Based on Table 10 and Figure 7a, it is clearly shown that the proposed MsO-KELM could obtain the best average values of ACC and MCC among all six compared algorithms at 70.184% and 40.608%, respectively. As a comparison, the WOA-KELM obtained the second best average values of ACC and MCC at 68.839% and 38.364%, respectively. In addition, the SSA-KELM could obtain the worst ACC value, specificity value, and MCC value at 64.517%, 45.252%, and 30.108%, respectively, but the GWO-KELM and the HHO-KELM obtained the worst sensitivity value at 68.668%. According to these evaluation metric results, we could find that the proposed MsO-KELM is superior to the others and could be adopted to diagnose the Cleveland heart disease.
The prediction results in solving Bupa via 10-fold cross-validation method are shown in Table 11. Based on Table 11 and Figure 7b, it is clearly shown that the proposed MsO-KELM could obtain the best average values of ACC, specificity, and MCC among all six compared algorithms at 70.476%, 59.423%, and 36.706%, respectively. As a comparison, the WOA-KELM could obtain the second best ACC value at 68.095%, the WOA-KELM and the HHO-KELM could obtain the second best specificity value at 55.407%, and the FPA-KELM could obtain the second best MCC value at 33.780%. For the sensitivity value, the MsO-KELM could obtain a result which is close to that of the other models. According to these evaluation metric results, we could find that the proposed MsO-KELM has more advantages in predicting liver disease than the other compared models.

5. Discussion

Combining information technology with medical information to realize the auxiliary diagnosis is a research trend in the digital health age; therefore, researchers have conducted a large number of prediction studies on common diseases by utilizing some machine learning models. However, the existing studies have weak generalization performance, which limits their further application for other diseases. In other words, these findings may obtain better prediction results in solving a certain disease but may obtain worse prediction results in solving other, different diseases. The reason why these findings have weak generalization performance is that the researchers mainly emphasize result-orientation for a certain disease, i.e., they focus on training with single disease data to obtain an effective prediction model which is suitable for the current disease. In fact, a prediction model with better generalization performance could help medical workers to diagnose various diseases. Therefore, exploring a prediction model with strong generalization ability has important theoretical and practical significance in the current digital health age.

5.1. Theoretical Significance

This paper is an attempt to explore a technology-oriented pathway for auxiliary diagnosis. On the one hand, our investigation adopts different disease data as the experimental objects, and aims to expand the application scope of the final findings by mining the internal characteristics of these different disease data, which could provide a novel research idea for researchers to eliminate the application limitations of the existing studies, and could provide effective theoretical guidance for researchers to realize comprehensive auxiliary diagnosis in facing various diseases.
On the other hand, utilizing an enhanced meta-heuristic algorithm to optimize the operating mechanism and inner structure of the original classifier is an efficient method to improve the performance of a model. The proposed MsO-KELM not only enhances the generalization ability and rapid learning ability of the original KELM classifier, but also realizes the self-adaptive process in predicting different diseases by introducing the EGSSA optimizer. The results of the evaluation metrics show that the MsO-KELM has significant advantages among all compared models in predicting the six diseases. Specifically, the MsO-KELM can obtain the best ACC value when predicting each disease, such that the ACC value is 94.124% in predicting breast cancer, the ACC value is 91.079% in predicting Autistic Spectrum Disorder, and the ACC value is 84.167% in predicting Parkinson’s disease. These ACC values could reflect the effectiveness of the MsO-KELM in disease prediction. Compared with some specific disease models [11,50], there may be a slight decrease in prediction accuracy of the MsO-KELM. However, according to the No Free Lunch (NFL) theory [51], although the MsO-KELM has slightly decreased accuracy in predicting a certain disease, it could predict more different diseases, which enhances the medical application value of the model.

5.2. Practical Significance

On the one hand, for the six different real-world diseases selected in this study, our investigation could assist doctors to diagnose or screen patients with related diseases, guide patients to prevent the diseases in a targeted manner, reduce the risk level of these diseases, and finally improve the survival quality of the patients.
On the other hand, there may be a consistent one-to-one match between the existing prediction models and their prediction targets, which means they will increase the economic cost and time cost for medical departments when analyzing different diseases. Therefore, a universal prediction model with strong generalization ability and robustness could predict more different diseases so as to improve the overall work efficiency of the medical workers.

5.3. Limitations

This study completes the construction of a self-adaptive prediction model in solving different diseases, but some limitations related to the MsO-KELM should be noted.
(i) In terms of model, although this paper presents four entirely novel optimization strategies to further optimize the prediction performance of the model, the performance improvement is accompanied by an increase in computational complexity, which is inevitable and could be comprehended according to the NFL theory. Therefore, we will further enhance the prediction performance of the MsO-KELM by redesigning a novel parameter optimization mechanism and optimizing the fundamental structure of the prediction model to reduce the computational complexity in the future studies.
(ii) In terms of data, on the one hand, the disease data analyzed in this paper are publicly available datasets, where each disease dataset has a limited sample size and some datasets even have specific geographical characteristics; on the other hand, the number of disease categories covered by the selected datasets are not enough. Therefore, we will increase the category of disease data and expand the sample size of disease data. In addition, we will also strive for collecting global disease data to eliminate the influence of the geographical factors.

6. Conclusions

To effectively predict more diseases and provide more comprehensive auxiliary diagnosis for medical workers, this paper proposes a novel disease prediction model, i.e., the MsO-KELM. There are two experiments conducted in this paper to elaborate the details of the MsO-KELM. The first experiment shows an optimization process for the EGSSA, which aims to construct a self-adaptive characteristic for the subsequent MsO-KELM model. The second experiment proves that the MsO-KELM has high accuracy, strong generalization ability, and strong robustness, which highlights the better disease prediction performance of the MsO-KELM. In the future, we will conduct in-depth research into optimizing the prediction model and expanding the disease data sample to achieve a further breakthrough in medical informatics.

Author Contributions

Conceptualization, Z.H. and J.M.; methodology, Z.H.; software, Z.H. and W.S.; validation, W.S.; writing—original draft preparation, Z.H.; writing—review and editing, Z.H.; supervision, J.M.; project administration, J.M.; funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Major Project of the National Social Science Foundation of China (grant number: 20&ZD125), the Natural Science Foundation of Jilin Province (grant number: 20210101480JC), and the Innovation Team Project of Philosophy and Social Science Research on "Chinese modernization road" and "new form of human civilization" of Jilin University (grant number: 2022CXTD20).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The medical diagnostic datasets adopted in this paper are from the UCI machine learning repository (https://archive.ics.uci.edu/mL/index.php, accessed on 6 June 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yuan, X.M.; Chen, J.H.; Zhang, K.; Wu, Y.; Yang, T.T. A stable AI-based binary and multiple class heart disease prediction model for IoMT. IEEE Trans. Ind. Inform. 2022, 18, 2032–2040. [Google Scholar] [CrossRef]
  2. Oh, S.M.; Stefani, K.M.; Kim, H.C. Development and application of chronic disease risk prediction models. Yonsei Med. J. 2014, 55, 853–860. [Google Scholar] [CrossRef] [PubMed]
  3. Chen, H.L.; Du, M.; Zhang, Y.Y.; Yang, C. Research on disease prediction method based on R-Lookahead-LSTM. Comput. Intell. Neurosci. 2022, 2022, 8431912. [Google Scholar] [CrossRef] [PubMed]
  4. Lin, X.; Wang, X.; Wang, Y.; Du, X.; Jin, L.; Wan, M.; Ge, H.; Yang, X. Optimized neural network based on genetic algorithm to construct hand-foot-and-mouth disease prediction and early-warning model. Int. J. Environ. Res. Public Health 2021, 18, 2959. [Google Scholar] [CrossRef]
  5. Godinho, M.A.; Martins, H.; Al-Shorbaji, N.; Quintana, Y.; Liaw, S.T. “Digital Health Diplomacy” in Global Digital Health? A call for critique and discourse. J. Am. Med. Inf. Assoc. 2022, 29, 1019–1024. [Google Scholar] [CrossRef]
  6. Nielsen, P.; Sahay, S. A critical review of the role of technology and context in digital health research. Dig Health 2022, 8, 20552076221109554. [Google Scholar] [CrossRef]
  7. Fadahunsi, K.P.; O’Connor, S.; Akinlua, J.T.; Wark, P.A.; Gallagher, J.; Carroll, C.; Car, J.; Majeed, A.; O’Donoghue, J. Information quality frameworks for digital health technologies: Systematic review. J. Med. Int. Res. 2021, 23, e23479. [Google Scholar] [CrossRef]
  8. Lin, R.H.; Wang, C.C.; Tung, C.W. A machine learning classifier for predicting stable MCI patients using gene biomarkers. Int. J. Environ. Res. Public Health 2022, 19, 4839. [Google Scholar] [CrossRef]
  9. Ji, M.; Xie, W.X.; Huang, R.L.; Qian, X.B. Forecasting erroneous neural machine translation of disease symptoms: Development of bayesian probabilistic classifiers for cross-lingual health translation. Int. J. Environ. Res. Public Health 2021, 18, 9873. [Google Scholar] [CrossRef]
  10. Wang, Y.F.; Wang, H.H.; Li, S.Y.; Wang, L.D. Survival risk prediction of esophageal cancer based on the kohonen network clustering algorithm and kernel extreme learning machine. Mathematics 2022, 10, 1367. [Google Scholar] [CrossRef]
  11. Rasool, A.; Bunterngchit, C.; Tiejian, L.; Islam, M.R.; Qu, Q.; Jiang, Q. Improved machine learning-based predictive models for breast cancer diagnosis. Int. J. Environ. Res. Public Health 2022, 19, 3211. [Google Scholar] [CrossRef] [PubMed]
  12. Kumar, M.; Ang, L.T.; Png, H.; Ng, M.; Tan, K.; Loy, S.L.; Tan, K.H.; Chan, J.K.Y.; Godfrey, K.M.; Chan, S.Y.; et al. Automated machine learning (AutoML)-derived preconception predictive risk model to guide early intervention for gestational diabetes mellitus. Int. J. Environ. Res. Public Health 2022, 19, 6792. [Google Scholar] [CrossRef] [PubMed]
  13. Alkenani, A.H.; Li, Y.F.; Xu, Y.; Zhang, Q. Predicting Alzheimer’s disease from spoken and written language using fusion-based stacked generalization. J. Biom. Inf. 2021, 118, 103803. [Google Scholar] [CrossRef] [PubMed]
  14. Nagavelli, U.; Samanta, D.; Chakraborty, P. Machine learning technology-based heart disease detection models. J. Healthc. Eng. 2022, 2022, 7351061. [Google Scholar] [CrossRef]
  15. Chlif, M.; Ammar, M.M.; Ben Said, N.; Sergey, L.; Ahmaidi, S.; Alassery, F.; Hamam, H. Mechanism of dyspnea during exercise in children with corrected congenital heart disease. Int. J. Environ. Res. Public Health 2022, 19, 99. [Google Scholar] [CrossRef]
  16. Derevitskii, I.V.; Savitskaya, D.A.; Babenko, A.Y.; Kovalchuk, S.V. Hybrid predictive modelling: Thyrotoxic atrial fibrillation case. J. Comput. Sci. 2021, 51, 101365. [Google Scholar] [CrossRef]
  17. Muhammad, L.J.; Al-Shourbaji, I.; Haruna, A.A.; Mohammed, I.A.; Ahmad, A.; Jibrin, M.B. Machine learning predictive models for coronary artery disease. SN Comput. Sci. 2021, 2, 350. [Google Scholar] [CrossRef]
  18. Xia, J.F.; Yang, D.Q.; Zhou, H.; Chen, Y.Y.; Zhang, H.L.; Liu, T.; Heidari, A.A.; Chen, H.L.; Pan, Z.F. Evolving kernel extreme learning machine for medical diagnosis via a disperse foraging sine cosine algorithm. Comput. Biol. Med. 2022, 141, 105137. [Google Scholar] [CrossRef]
  19. Zhang, N.; Ding, S.F. Unsupervised and semi-supervised extreme learning machine with wavelet kernel for high dimensional data. Mem. Comput. 2017, 9, 129–139. [Google Scholar] [CrossRef]
  20. Mohanty, F.; Rup, S.; Dash, B. Automated diagnosis of breast cancer using parameter optimized kernel extreme learning machine. Biomed. Signal Process. Control 2020, 62, 102108. [Google Scholar] [CrossRef]
  21. Ergul, U.; Bilgin, G. MCK-ELM: Multiple composite kernel extreme learning machine for hyperspectral images. Neur. Comput. Appl. 2020, 32, 6809–6819. [Google Scholar] [CrossRef]
  22. Chen, Y.L.; Zhang, X.J.; Karimian, H.; Xiao, G.; Huang, J.S. A novel framework for prediction of dam deformation based on extreme learning machine and Levy flight bat algorithm. J. Hydr. 2021, 23, 935–949. [Google Scholar] [CrossRef]
  23. Lv, F.; Han, M. Hyperspectral image classification based on multiple reduced kernel extreme learning machine. Int. J. Mach. Learn Cyb. 2019, 10, 3397–3405. [Google Scholar] [CrossRef]
  24. Liu, A.; Zhao, D.N.; Li, T.J. A data classification method based on particle swarm optimisation and kernel function extreme learning machine. Enterp. Inf. Syst. 2021, 1–16. [Google Scholar] [CrossRef]
  25. Zhu, C.J.; Yang, H.D.; Fan, Y.J.; Fan, B.; Xu, K.K. Online spatiotemporal modeling for time-varying distributed parameter systems using Kernel-based Multilayer Extreme Learning Machine. Nonlinear Dyn. 2022, 107, 761–780. [Google Scholar] [CrossRef]
  26. Shen, Y.P.; Zheng, K.F.; Wu, C.H.; Yang, Y.X. A nature-inspired multiple kernel extreme learning machine model for intrusion detection. KSII. Trans. Int. Inf. Syst. 2020, 14, 702–723. [Google Scholar]
  27. Khan, M.A.; Kadry, S.; Zhang, Y.D.; Akram, T.; Sharif, M.; Rehman, A.; Saba, T. Prediction of COVID-19-Pneumonia based on selected deep features and one class kernel extreme learning machine. Comput. Electr. Eng. 2021, 90, 106960. [Google Scholar] [CrossRef]
  28. Afzal, A.L.; Nair, N.K.; Asharaf, S. Deep kernel learning in extreme learning machines. Pattern Anal. Appl. 2021, 24, 11–19. [Google Scholar] [CrossRef]
  29. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng. 2020, 8, 22–34. [Google Scholar] [CrossRef]
  30. Ma, J.; Hao, Z.Y.; Sun, W.J. Enhancing sparrow search algorithm via multi-strategies for continuous optimization problems. Inf. Proc. Manag. 2022, 59, 102854. [Google Scholar] [CrossRef]
  31. Wu, C.Y.; Fu, X.S.; Pei, J.K.; Dong, Z.G. A novel sparrow search algorithm for the traveling salesman problem. IEEE Access 2021, 9, 153456–153471. [Google Scholar] [CrossRef]
  32. Dong, J.; Dou, Z.H.; Si, S.Q.; Wang, Z.C.; Liu, L.X. Optimization of capacity configuration of wind-solar-diesel-storage using improved sparrow search algorithm. J. Electr. Eng. Technol. 2022, 17, 1–14. [Google Scholar] [CrossRef]
  33. Gao, B.W.; Shen, W.; Guan, H.; Zheng, L.T.; Zhang, W. Research on multistrategy improved evolutionary sparrow search algorithm and its application. IEEE Access. 2022, 10, 62520–62534. [Google Scholar] [CrossRef]
  34. Li, G.Y.; Shu, C.; Liang, Z.W.; Peng, B.H.; Cheng, L.F. A modified sparrow search algorithm with application in 3d route planning for UAV. Sensors 2021, 21, 1224. [Google Scholar] [CrossRef]
  35. Yang, Y.T.; Chen, H.L.; Heidari, A.A.; Gandomi, A.H. Hunger games search: Visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Exp. Syst. Appl. 2021, 177, 114864. [Google Scholar] [CrossRef]
  36. Adel, H.; Dahou, A.; Mabrouk, A.; Abd Elaziz, M.; Kayed, M.; El-Henawy, I.M.; Alshathri, S.; Ali, A.A. Improving crisis events detection using distilBERT with hunger games search algorithm. Mathematics 2022, 10, 447. [Google Scholar] [CrossRef]
  37. Fahim, S.R.; Hasanien, H.M.; Turky, R.A.; Alkuhayli, A.; Al-Shamma’a, A.A.; Noman, A.M.; Tostado-Veliz, M.; Jurado, F. Parameter identification of proton exchange membrane fuel cell based on hunger games search algorithm. Energies 2021, 14, 5022. [Google Scholar] [CrossRef]
  38. Xu, B.Y.; Heidari, A.A.; Kuang, F.J.; Zhang, S.Y.; Chen, H.L.; Cai, Z.N. Quantum nelder-mead hunger games search for optimizing photovoltaic solar cells. Int. J. Energy Res. 2022, 46, 12417–12466. [Google Scholar] [CrossRef]
  39. Miao, F.H.; Yao, L.; Zhao, X.J. Symbiotic organisms search algorithm using random walk and adaptive Cauchy mutation on the feature selection of sleep staging. Exp. Syst. Appl. 2021, 176, 114887. [Google Scholar] [CrossRef]
  40. Zhao, S.W.; Wang, P.J.; Heidari, A.A.; Zhao, X.H.; Ma, C.; Chen, H.L. An enhanced Cauchy mutation grasshopper optimization with trigonometric substitution: Engineering design and feature selection. Eng. Comput. 2021, 1–34. [Google Scholar] [CrossRef]
  41. Liu, L.; Zhao, D.; Yu, F.H.; Heidari, A.A.; Li, C.Y.; Ouyang, J.S.; Chen, H.L.; Mafarja, M.; Turabieh, H.; Pan, J.Y. Ant colony optimization with Cauchy and greedy Levy mutations for multilevel COVID 19 X-ray image segmentation. Comput. Biol. 2021, 136, 104609. [Google Scholar] [CrossRef]
  42. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  43. Mirjalili, S.; Mirjalili, S.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  44. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H.L. Harris hawks optimization: Algorithm and applications. Future Gen. Comput. Syst.-Int. J. Esc. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  45. Shareef, H.; Ibrahim, A.; Mutlag, A. Lightning search algorithm. Appl. Soft Comput. 2015, 36, 315–333. [Google Scholar] [CrossRef]
  46. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  47. Yang, X.; Karamanoglu, M.; He, X. Flower pollination algorithm: A novel approach for multiobjective optimization. Eng. Opt. 2014, 46, 1222–1237. [Google Scholar] [CrossRef]
  48. Zhang, J.N.; Xia, K.W.; He, Z.P.; Yin, Z.X.; Wang, S.J. Semi-supervised ensemble classifier with improved sparrow search algorithm and its application in pulmonary nodule detection. Math. Probl. Eng. 2021, 2021, 6622935. [Google Scholar] [CrossRef]
  49. Xie, W.; Xing, C.; Wang, J.S.; Guo, S.S.; Guo, M.W.; Zhu, L.F. Hybrid henry gas solubility optimization algorithm based on the harris hawk optimization. IEEE Access. 2020, 8, 144665–144692. [Google Scholar] [CrossRef]
  50. El-Hasnony, I.M.; Barakat, S.I.; Mostafa, R.R. Optimized ANFIS model using hybrid meta-heuristic algorithms for Parkinson’s disease prediction in IoT environment. IEEE Access. 2020, 8, 119252–119270. [Google Scholar] [CrossRef]
  51. Sterkenburg, T.F.; Grunwald, P.D. The no-free-lunch theorems of supervised learning. Synthese 2021, 199, 9979–10015. [Google Scholar] [CrossRef]
Figure 1. The self-adaptive process of the MsO-KELM in predicting different diseases.
Figure 1. The self-adaptive process of the MsO-KELM in predicting different diseases.
Ijerph 19 12509 g001
Figure 2. The searching result figures based on Equations (11) and (12). (a) searching processes based on Equation (11); (b) searching processes based on Equation (12).
Figure 2. The searching result figures based on Equations (11) and (12). (a) searching processes based on Equation (11); (b) searching processes based on Equation (12).
Ijerph 19 12509 g002
Figure 3. The comparison result figures of searching processes. (a) the original searching process; (b) the searching process based on after-EEPS.
Figure 3. The comparison result figures of searching processes. (a) the original searching process; (b) the searching process based on after-EEPS.
Ijerph 19 12509 g003
Figure 4. The convergence curves of EGSSA and other algorithms. (a) the convergence curves on F5; (b) the convergence curves on F6; (c) the convergence curves on F12; (d) the convergence curves on F13; (e) the convergence curves on F22; (f) the convergence curves on F23.
Figure 4. The convergence curves of EGSSA and other algorithms. (a) the convergence curves on F5; (b) the convergence curves on F6; (c) the convergence curves on F12; (d) the convergence curves on F13; (e) the convergence curves on F22; (f) the convergence curves on F23.
Ijerph 19 12509 g004
Figure 5. The figure of evaluation metric results on Breast Cancer and Parkinson. (a) the evaluation criterion result on Breast Cancer; (b) the evaluation criterion result on Parkinson.
Figure 5. The figure of evaluation metric results on Breast Cancer and Parkinson. (a) the evaluation criterion result on Breast Cancer; (b) the evaluation criterion result on Parkinson.
Ijerph 19 12509 g005
Figure 6. The figure of evaluation metric results on Autistic and Heart Disease. (a) the evaluation criterion result on Autistic; (b) the evaluation criterion result on Heart Disease.
Figure 6. The figure of evaluation metric results on Autistic and Heart Disease. (a) the evaluation criterion result on Autistic; (b) the evaluation criterion result on Heart Disease.
Ijerph 19 12509 g006
Figure 7. The figure of evaluation metric results on Cleveland and Bupa. (a) the evaluation criterion result on Cleveland; (b) the evaluation criterion result on Bupa.
Figure 7. The figure of evaluation metric results on Cleveland and Bupa. (a) the evaluation criterion result on Cleveland; (b) the evaluation criterion result on Bupa.
Ijerph 19 12509 g007
Table 1. The characteristics of the six disease datasets.
Table 1. The characteristics of the six disease datasets.
DatasetsData VolumeAttributesMissing ValuesPositive VolumeNegative Volume
Breast cancer699916458241
Heart disease270130150120
Parkinson19523014748
Autistic Spectrum Disorder Screening Data292214141151
Cleveland303134139164
Bupa34560145200
Table 2. The parameter values of different experiments.
Table 2. The parameter values of different experiments.
AlgorithmParameter ValuesReference
SSAST = 0.8, PD = 20%, SD = 10% [29]
PSOc1 = 2, c2 = 2, Vmax = 10 [42]
GWOa = [2, 0] [43]
HHONo input parameters required [44]
LSActime = 10 [45]
WOAa = [2 0], a2 = [−2 −1], b = 1 [46]
FPAp = 0.5 [47]
SCACSSAST = 0.8, PD = 20%, SD = 10%, a = 2 [48]
HHOHGSOα = 2, β = 2, K = 1, M1 = 0.1, M2 = 0.2 [49]
Table 3. Characteristics of the 23 classical benchmark functions [30].
Table 3. Characteristics of the 23 classical benchmark functions [30].
FunctionFunction EquationuationDimRangeOptimal
UnimodalF1 f 1 ( x ) = i = 1 n x i 2 30 [−100, 100]0
F2 f 2 ( x ) = i = 1 n | x i | + i = 1 n | x i | 30 [−10, 10]0
F3 f 3 ( x ) = i = 1 n ( j 1 i x j ) 2 30 [−100, 100]0
F4 f 4 ( x ) = m a x i { | x i | , 1 i n } 30 [−100, 100]0
F5 f 5 ( x ) = i = 1 n 1 [ 100 ( x i + 1 x i 2 ) 2 + ( x i 1 ) 2 ] 30 [−30, 30]0
F6 f 6 ( x ) = i = 1 n ( [ x i + 0.5 ] ) 2 30 [−100, 100]0
F7 f 7 ( x ) = i = 1 n i x i 4 + r a n d o m [ 0 , 1 ) 30 [−1.28, 1.28]0
MultimodalF8 f 8 ( x ) = i = 1 n x i s i n ( | x i | ) 30 [−500, 500]−418.9829 × 5
F9 f 9 ( x ) = i = 1 n [ x i 2 10 c o s ( 2 π x i ) + 10 ] 30 [−5.12, 5.12]0
F10 f 10 ( x ) = 20 e x p ( 0.2 1 n i = 1 n x i ) e x p ( 1 n i = 1 n c o s ( 2 π x i ) ) + 20 + e 30 [−32, 32]0
F11 f 11 ( x ) = 1 4000 i = 1 n x i 2 i = 1 n c o s ( x i i ) + 1 30 [−600, 600]0
F12 f 12 ( x ) = π n { 10 s i n ( π y 1 ) + i = 1 n 1 ( y i 1 ) 2 [ 1 + 10 s i n 2 ( π y i + 1 ) ] + ( y n 1 ) 2 } + i = 1 n μ ( x i , 10 , 100 , 4 ) } 30 [−50, 50]0
y i = 1 + x i + 1 4
μ ( x i , a , k , m ) = { k ( x i a ) m                                           x i > a 0                                       a < x i < a k ( x i a ) m                 x i < a
F13 f 13 ( x ) = 0.1 { s i n 2 ( 3 π x 1 ) + i = 1 n ( x i 1 ) 2 [ 1 + s i n 2 ( 3 π x i + 1 ) ] + ( x n 1 ) 2 [ 1 + s i n 2 ( 2 π x n ) ] } + i = 1 n μ ( x i , 5 , 100 , 4 ) 30 [−50, 50]0
Fixed-dimension multimodalF14 f 14 ( x ) = ( 1 500 + j = 1 25 1 j + i = 1 2 ( x i a i j ) 6 ) 1 2 [−65, 65]1
F15 f 15 ( x ) = i = 1 11 [ a i x 1 ( b i 2 + b i x 2 ) b i 2 + b i x 3 + x 4 ] 2 4 [−5, 5]0.00030
F16 f 16 ( x ) = 4 x 1 2 2.1 x i 4 + 1 3 x 1 6 + x 1 x 2 4 x 2 2 + 4 x 2 4 2 [−5, 5]−1.0316
F17 f 17 ( x ) = ( x 2 5.1 4 π 2 x 1 2 + 5 π x 1 6 ) 2 + 10 ( 1 1 8 π ) c o s x 1 + 10 2 [−5, 5]0.398
F18 f 18 ( x ) = [ 1 + ( x 1 + x 2 + 1 ) 2 ( 19 14 x 1 + 3 x 1 2 14 x 2 + 6 x 1 x 2 + 3 x 2 2 ) ] × [ 30 + ( 2 x 1 3 x 2 ) 2 × ( 18 32 x 1 + 12 x 1 2 + 48 x 2 36 x 1 x 2 + 27 x 2 2 ) ] 2 [−2, 2]3
F19 f 19 ( x ) = i = 1 4 c i e x p ( j = 1 3 a i j ( x i p i j ) 2 ) 3 [1, 3]−3.86
F20 f 20 ( x ) = i = 1 4 c i e x p ( j = 1 6 a i j ( x i p i j ) 2 ) 6 [0, 1]−3.32
F21 f 21 ( x ) = i = 1 5 [ ( X a i ) ( X a i ) T + c i ] 1 4 [0, 10]−10.1532
F22 f 22 ( x ) = i = 1 7 [ ( X a i ) ( X a i ) T + c i ] 1 4 [0, 10]−10.4028
F23 f 23 ( x ) = i = 1 10 [ ( X a i ) ( X a i ) T + c i ] 1 4 [0, 10]−10.5363
Table 4. The results based on the 23 classical benchmark functions (retaining two decimal places).
Table 4. The results based on the 23 classical benchmark functions (retaining two decimal places).
EGSSASSAPSOGWOHHOLSAWOAFPASCACSSAHHOHGSO
F1avg0.00 × 10001.30 × 10−496.25 × 10−011.19 × 10−271.61 × 10−931.11 × 10−037.15 × 10−724.60 × 10−016.17 × 10−165.82 × 10−268
std0.00 × 10007.11 × 10−493.36 × 10−011.64 × 10−278.22 × 10−935.93 × 10−033.90 × 10−711.39 × 10−013.35 × 10−150.00 × 1000
F2avg0.00 × 10005.36 × 10−273.26 × 10011.28 × 10−162.34 × 10−482.48 × 10−019.69 × 10−512.80 × 10001.53 × 10−079.93 × 10−160
std0.00 × 10002.90 × 10−265.50 × 10011.32 × 10−161.25 × 10−473.80 × 10−014.84 × 10−503.64 × 10−016.56 × 10−075.44 × 10−159
F3avg0.00 × 10002.61 × 10−293.37 × 10021.43 × 10−057.32 × 10−641.37 × 10024.99 × 10043.28 × 10−011.76 × 10−082.17 × 10−309
std0.00 × 10001.18 × 10−289.30 × 10013.34 × 10−054.01 × 10−638.69 × 10011.66 × 10041.02 × 10−015.08 × 10−080.00 × 1000
F4avg0.00 × 10003.66 × 10−262.77 × 10007.27 × 10−079.66 × 10−499.62 × 10005.04 × 10013.67 × 10−019.26 × 10−265.81 × 10−147
std0.00 × 10001.99 × 10−254.30 × 10−011.06 × 10−064.89 × 10−484.25 × 10002.86 × 10014.42 × 10−024.86 × 10−253.18 × 10−146
F5avg2.56 × 10−024.61 × 10−016.55 × 10022.73 × 10011.01 × 10−021.21 × 10022.80 × 10017.96 × 10011.03 × 10002.73 × 1001
std9.08 × 10−027.42 × 10−014.83 × 10028.13 × 10−019.64 × 10−031.84 × 10024.42 × 10−011.81 × 10011.56 × 10006.72 × 10−01
F6avg9.24 × 10−063.05 × 10−026.67 × 10−016.92 × 10−012.90 × 10−047.36 × 10−053.94 × 10−011.23 × 10002.35 × 10−021.13 × 10−03
std1.38 × 10−051.64 × 10−023.37 × 10−013.65 × 10−016.00 × 10−043.71 × 10−042.32 × 10−013.99 × 10−011.14 × 10−021.27 × 10−03
F7avg1.11 × 10−047.70 × 10−042.52 × 10002.25 × 10−031.90 × 10−043.05 × 10−023.59 × 10−035.52 × 10−016.41 × 10−029.76 × 10−05
std1.46 × 10−046.79 × 10−042.80 × 10001.12 × 10−032.11 × 10−048.45 × 10−035.59 × 10−032.76 × 10−015.67 × 10−029.71 × 10−05
F8avg−8.10 × 1003−7.79 × 1003−6.63 × 1003−5.89 × 1003−1.26 × 1004−7.57 × 1003−1.02 × 1004−4.26 × 1001−6.04 × 1003−1.13 × 1004
std2.23 × 10033.03 × 10037.63 × 10021.07 × 10031.40 × 10007.48 × 10021.70 × 10032.65 × 10007.52 × 10021.19 × 1003
F9avg0.00 × 10000.00 × 10001.43 × 10021.92 × 10000.00 × 10006.92 × 10010.00 × 10004.22 × 10013.46 × 10−100.00 × 1000
std0.00 × 10000.00 × 10003.74 × 10013.27 × 10000.00 × 10001.60 × 10010.00 × 10001.93 × 10011.53 × 10−090.00 × 1000
F10avg8.88 × 10−168.88 × 10−161.85 × 10001.04 × 10−138.88 × 10−162.90 × 10003.85 × 10−151.38 × 10002.35 × 10−078.88 × 10−16
std0.00 × 10000.00 × 10007.06 × 10−011.83 × 10−140.00 × 10008.34 × 10−012.30 × 10−152.31 × 10−011.28 × 10−060.00 × 1000
F11avg0.00 × 10000.00 × 10006.22 × 10−024.49 × 10−030.00 × 10007.22 × 10−035.86 × 10−031.66 × 10−023.20 × 10−110.00 × 1000
std0.00 × 10000.00 × 10004.06 × 10−028.49 × 10−030.00 × 10001.09 × 10−023.21 × 10−025.72 × 10−031.76 × 10−100.00 × 1000
F12avg1.43 × 10−071.00 × 10−021.17 × 10003.97 × 10−021.01 × 10−056.76 × 10−012.70 × 10−026.34 × 10−022.15 × 10−031.00 × 10−04
std1.18 × 10−073.16 × 10−022.30 × 10001.49 × 10−021.27 × 10−051.41 × 10001.61 × 10−022.52 × 10−021.27 × 10−038.86 × 10−05
F13avg6.36 × 10−052.53 × 10−014.69 × 10−017.10 × 10−018.67 × 10−056.36 × 10−025.62 × 10−011.07 × 10001.41 × 10−011.88 × 10−02
std1.74 × 10−041.31 × 10−013.07 × 10−012.93 × 10−011.02 × 10−041.32 × 10−013.28 × 10−013.07 × 10−011.06 × 10−011.72 × 10−02
F14avg9.95 × 10001.09 × 10013.33 × 10004.52 × 10001.26 × 10001.36 × 10003.00 × 10001.27 × 10011.27 × 10011.03 × 1000
std4.04 × 10003.83 × 10002.81 × 10004.22 × 10009.32 × 10−011.02 × 10003.06 × 10001.34 × 10−149.61 × 10−111.81 × 10−01
F15avg3.58 × 10−044.79 × 10−048.90 × 10−043.05 × 10−033.81 × 10−045.94 × 10−041.17 × 10−033.08 × 10−044.85 × 10−044.08 × 10−04
std8.86 × 10−051.46 × 10−041.33 × 10−046.91 × 10−032.13 × 10−043.25 × 10−042.43 × 10−031.19 × 10−062.36 × 10−042.40 × 10−04
F16avg−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000−1.03 × 1000
std3.00 × 10−104.74 × 10−165.30 × 10−161.45 × 10−083.95 × 10−106.58 × 10−169.97 × 10−101.53 × 10−092.47 × 10−037.54 × 10−13
F17avg3.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−013.98 × 10−017.78 × 10003.98 × 10−013.98 × 10−01
std2.09 × 10−048.68 × 10−090.00 × 10006.52 × 10−072.87 × 10−060.00 × 10001.26 × 10−052.71 × 10−151.23 × 10−062.63 × 10−10
F18avg3.00 × 10003.00 × 10003.00 × 10003.00 × 10003.00 × 10003.00 × 10003.00 × 10003.00 × 10003.00 × 10003.00 × 1000
std2.85 × 10−059.65 × 10−154.41 × 10−154.39 × 10−053.54 × 10−072.56 × 10−159.64 × 10−052.14 × 10−106.60 × 10−041.03 × 10−11
F19avg−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000−3.86 × 1000
std1.36 × 10−042.46 × 10−052.36 × 10−151.91 × 10−035.19 × 10−032.67 × 10−157.91 × 10−036.33 × 10−092.77 × 10−051.09 × 10−05
F20avg−3.29 × 1000−3.26 × 1000−3.27 × 1000−3.27 × 1000−3.07 × 1000−3.27 × 1000−3.20 × 1000−3.28 × 1000−3.25 × 1000−3.25 × 1000
std5.33 × 10−026.84 × 10−026.03 × 10−027.31 × 10−021.26 × 10−015.92 × 10−021.56 × 10−013.26 × 10−026.29 × 10−028.31 × 10−02
F21avg−9.85 × 1000−8.24 × 1000−7.14 × 1000−9.31 × 1000−5.21 × 1000−7.72 × 1000−8.65 × 1000−5.06 × 1000−8.79 × 1000−6.58 × 1000
std5.45 × 10−012.46 × 10003.38 × 10001.92 × 10008.73 × 10−013.13 × 10002.85 × 10002.94 × 10−072.29 × 10002.38 × 1000
F22avg−1.02 × 1001−9.16 × 1000−6.19 × 1000−1.02 × 1001−5.23 × 1000−8.11 × 1000−6.97 × 1000−5.09 × 1000−8.81 × 1000−6.68 × 1000
std4.27 × 10−012.29 × 10003.39 × 10009.70 × 10−018.01 × 10−013.34 × 10003.36 × 10002.36 × 10−072.48 × 10002.48 × 1000
F23avg−1.02 × 1001−8.73 × 1000−8.59 × 1000−1.01 × 1001−5.48 × 1000−7.27 × 1000−6.77 × 1000−5.13 × 1000−9.64 × 1000−6.39 × 1000
std7.11 × 10−012.59 × 10003.34 × 10001.75 × 10001.34 × 10003.85 × 10003.23 × 10003.42 × 10−072.05 × 10002.33 × 1000
Table 5. The results of the statistical test.
Table 5. The results of the statistical test.
AlgorithmBetterEquationualWorstW+W−p-Value
EGSSA versus SSA167013600.000438
EGSSA versus PSO1841176140.001116
EGSSA versus GWO1751155160.002472
EGSSA versus HHO137395410.162673
EGSSA versus LSA1841176140.001116
EGSSA versus WOA1652140310.017621
EGSSA versus FPA193120910.000103
EGSSA versus SCACSSA194019000.000132
EGSSA versus HHOHGSO128388320.111769
Table 6. The prediction results on disease dataset Breast Cancer.
Table 6. The prediction results on disease dataset Breast Cancer.
Breast Cancer
IndicatorAlgorithmsMeanStd1#2#3#4#5#6#7#8#9#10#
ACCGWO-KELM0.936950.038470.867650.955880.955880.970590.926470.867650.926470.985290.955880.95775
HHO-KELM0.922370.046570.838240.926470.941180.970590.955880.852940.882350.985290.941180.92958
FPA-KELM0.925310.045800.852940.926470.941180.985290.955880.852940.882350.985290.941180.92958
WOA-KELM0.922370.046570.838240.926470.941180.970590.955880.852940.882350.985290.941180.92958
SSA-KELM0.922370.046570.838240.926470.941180.970590.955880.852940.882350.985290.941180.92958
MsO-KELM0.941240.037850.852940.955880.955880.970590.926470.897060.941180.955880.970590.98592
SensitivityGWO-KELM0.926280.065860.794120.925930.923081.000000.969701.000000.894741.000000.909090.84615
HHO-KELM0.887510.112780.735290.925930.923080.973681.000001.000000.684211.000000.863640.76923
FPA-KELM0.885390.121820.764710.925930.923081.000001.000001.000000.684211.000000.863640.69231
WOA-KELM0.887510.112780.735290.925930.923080.973681.000001.000000.684211.000000.863640.76923
SSA-KELM0.887510.112780.735290.925930.923080.973681.000001.000000.684211.000000.863640.76923
MsO-KELM0.944340.068140.764710.962960.923080.973680.96971.000000.894741.000000.954551.00000
SpecificityGWO-KELM0.942120.048380.941180.975610.976190.933330.885710.826920.938780.982460.978260.98276
HHO-KELM0.939440.048470.941180.926830.952380.966670.914290.807690.959180.982460.978260.96552
FPA-KELM0.941170.049650.941180.926830.952380.966670.914290.807690.959180.982460.978260.98276
WOA-KELM0.939440.048470.941180.926830.952380.966670.914290.807690.959180.982460.978260.96552
SSA-KELM0.939440.048470.941180.926830.952380.966670.914290.807690.959180.982460.978260.96552
MsO-KELM0.945390.037530.941180.951220.976190.966670.885710.865380.959180.947370.978260.98276
MCCGWO-KELM0.860680.072950.743380.907710.906350.941630.856520.727480.820830.948990.898520.85542
HHO-KELM0.824330.097360.691280.847790.875460.940350.915480.705010.696260.948990.864400.75824
FPA-KELM0.829160.098670.717140.847790.875460.970510.915480.705010.696260.948990.864400.75053
WOA-KELM0.824330.097360.691280.847790.875460.940350.915480.705010.696260.948990.864400.75824
SSA-KELM0.824330.097360.691280.847790.875460.940350.915480.705010.696260.948990.864400.75824
MsO-KELM0.870990.0719060.717140.908860.906350.940350.856520.775890.853920.862760.932810.95528
Table 7. The prediction results on disease dataset Parkinson.
Table 7. The prediction results on disease dataset Parkinson.
Parkinson
IndicatorAlgorithmsMeanStd1#2#3#4#5#6#7#8#9#10#
ACCGWO-KELM0.810000.066330.750000.750000.800000.750000.900000.850000.800000.950000.750000.80000
HHO-KELM0.826670.082060.800000.650000.800000.950000.850000.800000.900000.750000.900000.86667
FPA-KELM0.828330.061030.750000.800000.900000.750000.800000.800000.800000.850000.900000.93333
WOA-KELM0.816670.051640.850000.850000.800000.800000.850000.750000.850000.850000.700000.86667
SSA-KELM0.803330.080210.900000.850000.900000.650000.850000.700000.800000.850000.800000.73333
MsO-KELM0.841670.070420.850000.950000.850000.900000.750000.850000.700000.800000.900000.86667
SensitivityGWO-KELM0.859910.093670.750000.750000.846150.764711.000000.923080.785711.000000.933330.84615
HHO-KELM0.884230.102330.882350.800000.764711.000000.933330.928570.933330.666670.933331.00000
FPA-KELM0.880710.092500.769230.812500.928570.705880.875000.928570.846150.941181.000001.00000
WOA-KELM0.868090.106970.937500.823531.000000.692310.928570.812501.000000.882350.687500.91667
SSA-KELM0.865210.107200.937501.000000.937500.666670.937500.733330.937500.928570.823530.75000
MsO-KELM0.895960.0684950.875000.933330.846150.888890.800000.875000.800001.000000.941181.00000
SpecificityGWO-KELM0.652860.176660.750000.750000.714290.666670.600000.714290.833330.800000.200000.50000
HHO-KELM0.644170.277190.333330.200001.000000.875000.600000.500000.800001.000000.800000.33333
FPA-KELM0.651190.185050.714290.750000.833331.000000.500000.500000.714290.333330.666670.50000
WOA-KELM0.675000.187030.500001.000000.428571.000000.666670.500000.571430.666670.750000.66667
SSA-KELM0.615000.140920.750000.700000.750000.600000.500000.600000.250000.666670.666670.66667
MsO-KELM0.675710.246850.750001.000000.857141.000000.600000.750000.600000.200000.666670.33333
MCCGWO-KELM0.505790.199720.419310.419310.560440.336120.727610.663390.579070.866030.192450.29417
HHO-KELM0.533380.246180.215690.000000.572480.898720.577350.490990.733330.577350.733330.53452
FPA-KELM0.543650.142320.470760.490990.761900.514500.375000.490990.560440.326730.763760.68139
WOA-KELM0.539870.125790.490100.641690.572480.663750.629940.288680.681390.490100.357220.58333
SSA-KELM0.477490.184070.687500.733800.687500.235700.490100.302610.250000.629940.404230.35355
MsO-KELM0.571400.146770.577350.881920.684740.666670.377960.577350.408250.397360.607840.53452
Table 8. The prediction results on disease dataset Autistic Spectrum Disorder Screening Data for Children.
Table 8. The prediction results on disease dataset Autistic Spectrum Disorder Screening Data for Children.
Autistic Spectrum Disorder Screening Data for Children
IndicatorAlgorithmsMeanStd1#2#3#4#5#6#7#8#9#10#
ACCGWO-KELM0.886650.044230.896550.931030.896550.827590.862070.896550.931030.793100.896550.93548
HHO-KELM0.886430.049250.896550.896550.827590.931030.862070.862070.793100.896550.931030.96774
FPA-KELM0.893770.062620.862070.931030.827590.793100.931031.000000.965520.896550.827590.90323
WOA-KELM0.894220.050880.965520.896550.862070.896550.931030.896550.896550.965520.793100.83871
SSA-KELM0.849170.058380.862070.931030.793100.965520.827590.793100.793100.862070.793100.87097
MsO-KELM0.910790.056210.896550.965520.896550.931030.965520.931030.931030.758620.896550.93548
SensitivityGWO-KELM0.986060.028071.000000.923081.000001.000001.000001.000001.000000.937501.000001.00000
HHO-KELM0.917860.084720.937500.928570.800001.000000.789470.923080.800001.000001.000001.00000
FPA-KELM0.986190.027641.000001.000001.000001.000000.928571.000001.000001.000000.933331.00000
WOA-KELM0.912460.088290.937500.928570.909090.933331.000000.866670.857141.000000.692311.00000
SSA-KELM0.840790.096570.833331.000000.722221.000000.875000.714290.750000.812500.818180.88235
MsO-KELM0.984240.031980.909091.000001.000001.000001.000001.000001.000001.000001.000000.93333
SpecificityGWO-KELM0.786950.097980.823530.937500.785710.642860.714290.769230.866670.615380.857140.85714
HHO-KELM0.869320.059860.846150.866670.857140.875001.000000.812500.785710.833330.866670.95000
FPA-KELM0.807710.109830.777780.846150.705880.647060.933331.000000.933330.750000.714290.76923
WOA-KELM0.878510.078691.000000.866670.833330.857140.875000.928570.933330.928570.875000.68750
SSA-KELM0.863220.053920.882350.894740.909090.928570.769230.866670.823530.923080.777780.85714
MsO-KELM0.833790.111790.888890.933330.833330.888890.933330.833330.714290.562500.812500.93750
MCCGWO-KELM0.786200.082910.811680.860580.809170.694370.750940.804840.870820.594330.789540.87574
HHO-KELM0.779640.098570.791300.795240.657140.870820.750940.732070.585710.809170.870820.93318
FPA-KELM0.805780.107250.755230.867260.705880.656790.861901.000000.933330.798440.666960.81200
WOA-KELM0.793980.100180.933030.795240.724360.794270.870820.795240.794270.933030.581460.71807
SSA-KELM0.699170.116680.715690.863490.613010.933030.650520.589430.573530.732070.581460.73950
MsO-KELM0.828990.089100.786160.933330.809170.867260.933330.863490.809170.604670.812500.87083
Table 9. The prediction results on disease dataset Heart Disease.
Table 9. The prediction results on disease dataset Heart Disease.
Heart Disease
IndicatorAlgorithmsMeanStd1#2#3#4#5#6#7#8#9#10#
ACCGWO-KELM0.696300.087330.666670.592590.814810.703700.703700.592590.666670.814810.814810.59259
HHO-KELM0.692590.087720.666670.592590.851850.666670.703700.592590.666670.814810.814810.59259
FPA-KELM0.707410.071150.703700.666670.777780.703700.740740.666670.666670.814810.777780.55556
WOA-KELM0.703700.097990.666670.629630.814810.740740.703700.592590.666670.777780.888890.55556
SSA-KELM0.662960.073050.592590.592590.777780.666670.666670.592590.629630.666670.814810.62963
MsO-KELM0.722220.078130.703700.703700.851850.777780.740740.666670.703700.814810.703700.55556
SensitivityGWO-KELM0.576900.156820.600000.428570.900000.571430.571430.250000.636360.636360.636360.53846
HHO-KELM0.569760.158530.600000.428570.900000.500000.571430.250000.636360.636360.636360.53846
FPA-KELM0.631980.136360.700000.571430.800000.642860.642860.333330.545450.727270.818180.53846
WOA-KELM0.609370.166830.600000.500000.900000.642860.571430.250000.636360.636360.818180.53846
SSA-KELM0.429510.073050.400000.357140.700000.500000.500000.166670.545450.272730.545450.30769
MsO-KELM0.688530.117800.800000.642860.800000.785710.642860.416670.636360.727270.818180.61538
SpecificityGWO-KELM0.800420.097430.705880.769230.764710.846150.846150.866670.687500.937500.937500.64286
HHO-KELM0.800420.097430.705880.769230.823530.846150.846150.866670.687500.937500.937500.64286
FPA-KELM0.773500.093680.705880.769230.764710.769230.846150.933330.750000.875000.750000.57143
WOA-KELM0.787020.103680.705880.769230.764710.846150.846150.866670.687500.875000.937500.57143
SSA-KELM0.855480.095340.705880.846150.823530.846150.846150.933330.687500.937501.000000.92857
MsO-KELM0.753070.120570.647060.769230.882350.769230.846150.866670.750000.875000.625000.50000
MCCGWO-KELM0.390370.178110.300620.209660.642420.432070.432070.149240.320240.617510.617510.18232
HHO-KELM0.383850.177660.300620.209660.703140.366900.432070.149240.320240.617510.617510.18232
FPA-KELM0.412450.141830.394450.346420.548800.414370.497280.341120.300620.612810.558740.10989
WOA-KELM0.403220.200240.300620.278570.642420.497280.432070.149240.320240.533000.768900.10989
SSA-KELM0.322800.154460.108470.231790.523530.366900.366900.158110.232950.290770.644660.30390
MsO-KELM0.445570.150810.432070.414370.682350.554950.497280.321270.386360.612810.438230.11602
Table 10. The prediction results on disease dataset Cleveland.
Table 10. The prediction results on disease dataset Cleveland.
Cleveland
IndicatorAlgorithmsMeanStd1#2#3#4#5#6#7#8#9#10#
ACCGWO-KELM0.678620.098060.700000.500000.766670.700000.733330.600000.766670.833330.600000.58621
HHO-KELM0.678620.098060.700000.500000.766670.700000.733330.600000.766670.833330.600000.58621
FPA-KELM0.685060.093720.600000.633330.700000.633330.866670.666670.766670.766670.700000.51724
WOA-KELM0.688390.092410.600000.633330.700000.666670.866670.666670.766670.766670.700000.51724
SSA-KELM0.645170.084810.600000.633330.600000.633330.833330.633330.70000.733330.533330.55172
MsO-KELM0.701840.087410.700000.600000.700000.700000.866670.666670.766670.800000.666670.55172
SensitivityGWO-KELM0.674210.124660.727270.500000.928570.666670.692310.571430.785710.727270.642860.50000
HHO-KELM0.674210.124660.727270.500000.928570.666670.692310.571430.785710.727270.642860.50000
FPA-KELM0.592560.154310.454550.571430.714290.400000.692310.642860.857140.545450.714290.33333
WOA-KELM0.599230.147120.454550.571430.714290.466670.692310.642860.857140.545450.714290.33333
SSA-KELM0.452520.140780.272730.428570.500000.333330.615380.571430.714290.454550.357140.27778
MsO-KELM0.635190.111340.545450.571430.714290.533330.692310.642860.857140.636360.714290.44444
SpecificityGWO-KELM0.686680.107280.684210.500000.625000.733330.764710.625000.750000.894740.562500.72727
HHO-KELM0.686680.107280.684210.500000.625000.733330.764710.625000.750000.894740.562500.72727
FPA-KELM0.770130.110230.684210.687500.687500.866671.000000.687500.687500.894740.687500.81818
WOA-KELM0.770130.110230.684210.687500.687500.866671.000000.687500.687500.894740.687500.81818
SSA-KELM0.818000.124260.789470.812500.687500.933331.000000.687500.687500.894740.687501.00000
MsO-KELM0.759060.118860.789470.625000.687500.866671.000000.687500.687500.894740.625000.72727
MCCGWO-KELM0.362450.190570.397470.000000.573090.400890.457010.196430.534520.635850.205360.22391
HHO-KELM0.362450.190570.397470.000000.573090.400890.457010.196430.534520.635850.205360.22391
FPA-KELM0.377420.173920.138760.260680.400890.301510.748630.330360.548330.479690.400890.16449
WOA-KELM0.383640.172190.138760.260680.400890.363700.748630.330360.548330.479690.400890.16449
SSA-KELM0.301080.175250.070870.262450.190940.333330.689580.260680.400890.397960.047250.35681
MsO-KELM0.406080.165620.342380.196430.400890.424260.748630.330360.548330.558490.339290.17172
Table 11. The prediction results on disease dataset Bupa.
Table 11. The prediction results on disease dataset Bupa.
Liver Disorders
IndicatorAlgorithmsMeanStd1#2#3#4#5#6#7#8#9#10#
ACCGWO-KELM0.655710.125270.342860.742860.685710.714290.542860.628570.828570.685710.685710.70000
HHO-KELM0.678100.045830.657140.742860.742860.714290.714290.628570.657140.600000.657140.66667
FPA-KELM0.664290.104050.428570.742860.685710.714290.542860.657140.828570.657140.685710.70000
WOA-KELM0.680950.043490.657140.742860.742860.714290.714290.657140.657140.600000.657140.66667
SSA-KELM0.667140.072860.542860.771430.685710.714290.628570.657140.742860.542860.685710.70000
MsO-KELM0.704760.052160.685710.714290.771430.771430.657140.657140.771430.628570.657140.73333
SensitivityGWO-KELM0.849230.095080.800000.727270.944440.724140.941180.909090.923080.944440.703700.87500
HHO-KELM0.801010.082830.800000.727270.833330.724141.000000.818180.769230.722220.740740.87500
FPA-KELM0.843680.090890.800000.727270.944440.724140.941180.909090.923080.888890.703700.87500
WOA-KELM0.810100.088980.800000.727270.833330.724141.000000.909090.769230.722220.740740.87500
SSA-KELM0.805460.101560.800000.757580.777780.724141.000000.909090.884620.611110.777780.81250
MsO-KELM0.821330.079470.800000.696970.888890.793100.941180.909090.846150.722220.740740.87500
SpecificityGWO-KELM0.510410.218300.266671.000000.411760.666670.166670.500000.555560.411760.625000.50000
HHO-KELM0.554070.184970.633331.000000.647060.666670.444440.541670.333330.470590.375000.42857
FPA-KELM0.524580.208960.366671.000000.411760.666670.166670.541670.555560.411760.625000.50000
WOA-KELM0.554070.184970.633331.000000.647060.666670.444440.541670.333330.470590.375000.42857
SSA-KELM0.532470.193780.500001.000000.588240.666670.277780.541670.333330.470590.375000.57143
MsO-KELM0.594230.166480.666671.000000.647060.666670.388890.541670.555560.529410.375000.57143
MCCGWO-KELM0.335460.130570.053380.363640.423970.310300.169410.393040.522980.423970.285660.40825
HHO-KELM0.308910.134650.306410.363640.490100.310300.528990.337570.102560.199440.107580.34247
FPA-KELM0.337800.115490.122870.363640.423970.310300.169410.427140.522980.343810.285660.40825
WOA-KELM0.317860.139160.306410.363640.490100.310300.528990.427140.102560.199440.107580.34247
SSA-KELM0.298710.113770.210730.389250.373410.310300.396750.427140.252750.082510.146780.39747
MsO-KELM0.367060.115440.333330.340820.554370.383570.392860.427140.401710.256720.107580.47246
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hao, Z.; Ma, J.; Sun, W. The Technology-Oriented Pathway for Auxiliary Diagnosis in the Digital Health Age: A Self-Adaptive Disease Prediction Model. Int. J. Environ. Res. Public Health 2022, 19, 12509. https://doi.org/10.3390/ijerph191912509

AMA Style

Hao Z, Ma J, Sun W. The Technology-Oriented Pathway for Auxiliary Diagnosis in the Digital Health Age: A Self-Adaptive Disease Prediction Model. International Journal of Environmental Research and Public Health. 2022; 19(19):12509. https://doi.org/10.3390/ijerph191912509

Chicago/Turabian Style

Hao, Zhiyuan, Jie Ma, and Wenjing Sun. 2022. "The Technology-Oriented Pathway for Auxiliary Diagnosis in the Digital Health Age: A Self-Adaptive Disease Prediction Model" International Journal of Environmental Research and Public Health 19, no. 19: 12509. https://doi.org/10.3390/ijerph191912509

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop