Next Article in Journal
Artificial Intelligence, Social Media and Supply Chain Management: The Way Forward
Previous Article in Journal
Intelligent Reflective Surfaces for Wireless Networks: An Overview of Applications, Approached Issues, and Open Problems
Previous Article in Special Issue
D-GENE-Based Discovery of Frequent Occupational Diseases among Female Home-Based Workers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Heart Disease Prediction Using Particle Swarm Optimization Based Stacked Sparse Autoencoder

Department of Electrical and Electronic Engineering Science, University of Johannesburg, Johannesburg 2006, South Africa
*
Author to whom correspondence should be addressed.
Electronics 2021, 10(19), 2347; https://doi.org/10.3390/electronics10192347
Submission received: 26 July 2021 / Revised: 16 August 2021 / Accepted: 25 August 2021 / Published: 25 September 2021
(This article belongs to the Special Issue Applied Data Mining)

Abstract

:
Heart disease is the leading cause of death globally. The most common type of heart disease is coronary heart disease, which occurs when there is a build-up of plaque inside the arteries that supply blood to the heart, making blood circulation difficult. The prediction of heart disease is a challenge in clinical machine learning. Early detection of people at risk of the disease is vital in preventing its progression. This paper proposes a deep learning approach to achieve improved prediction of heart disease. An enhanced stacked sparse autoencoder network (SSAE) is developed to achieve efficient feature learning. The network consists of multiple sparse autoencoders and a softmax classifier. Additionally, in deep learning models, the algorithm’s parameters need to be optimized appropriately to obtain efficient performance. Hence, we propose a particle swarm optimization (PSO) based technique to tune the parameters of the stacked sparse autoencoder. The optimization by the PSO improves the feature learning and classification performance of the SSAE. Meanwhile, the multilayer architecture of autoencoders usually leads to internal covariate shift, a problem that affects the generalization ability of the network; hence, batch normalization is introduced to prevent this problem. The experimental results show that the proposed method effectively predicts heart disease by obtaining a classification accuracy of 0.973 and 0.961 on the Framingham and Cleveland heart disease datasets, respectively, thereby outperforming other machine learning methods and similar studies.

1. Introduction

Cardiovascular disease (CVD) is a life-threatening condition. It is the leading cause of death globally, with 30% of all global deaths attributed to it, amounting to 17 million deaths globally [1]. Additionally, deaths related to CVDs are estimated to rise to 22 million in 2030 if urgent measures are not taken [2]. Statistics from the American Heart Association (AHA) show that about 50% of American adults suffer from cardiovascular diseases [3]. Meanwhile, it is difficult to detect heart disease due to numerous contributory risk factors, including high blood pressure, diabetes, high cholesterol, arrhythmia, etc. [4]. However, these risk factors can be reduced through appropriate lifestyle changes.
Common symptoms of heart disease include shortness of breath, swollen feet, and body weakness [5]. Early detection is difficult but can significantly improve patient survival rates. Therefore, enhanced detection through machine learning (ML) based predictive models has been recently supported by clinicians to minimize the death rate and enhance the clinical decision-making process. The use of machine learning in clinical decision making can aid clinicians to detect heart disease risk and provide necessary treatments and recommendations to manage the risk [2]. To achieve this, electronic health records have been used to train ML models and learn the hidden relationships in the data. A few publicly available heart disease datasets and numerous predictive models have been developed over the years [6,7,8,9,10]. Meanwhile, researchers and scientists still have difficulty obtaining high prediction performance and identifying the most relevant heart disease risk factors [11].
The growing prevalence and high heart disease mortality rate necessitate rapid and accurate diagnosis of the disease [12]. Meanwhile, several ML research studies have focused on developing data preprocessing and transformation techniques to improve classification performance [13,14,15]. A typical example of such a data transformation technique is feature learning [16,17,18]. Recent advances in deep neural networks (DNN) have provided ML researchers with an opportunity to improve performance in different predictive models. Furthermore, the effectiveness of DNN can be utilized in learning the hidden correlations between the various features (i.e., risk factors) in heart disease datasets that would lead to improved prediction performance.
A set of salient features can be obtained by applying feature learning techniques, and the resulting low-dimensional feature representation can simplify the classification task. Feature learning or representation learning ensures the model automatically discovers the required representation for effective feature detection and classification of the input data [19]. Recently, feature learning techniques have received much attention, especially autoencoders, which have achieved state-of-the-art performance in learning a good representation of data [20]. An autoencoder (AE) can efficiently process and extract hidden representations from heart disease data for proper classification.
The network learns a hidden correlation between the input features and reconstructs it at the output. Furthermore, a sparsity constraint can be imposed on the hidden units that enable the network to better represent the input [19,21,22], thereby allowing the supervised learning algorithm to perform better classification. Meanwhile, determining the optimal parameters to train the feature learning algorithm is essential in achieving excellent deep feature learning [23,24]. Therefore, this paper presents an efficient heart disease prediction approach using the particle swarm optimization technique to optimize the parameters of a stacked sparse autoencoder. Optimizing the parameters of the SSAE ensures the algorithm achieves efficient feature learning.
We also aim to prevent internal covariate shift [25], a problem associated with multilayer autoencoders such as the proposed SSAE, by introducing batch normalization to the network. Finally, the effectiveness of the proposed approach is demonstrated by comparing it with other well-known ML methods, including logistic regression (LR), k-nearest neighbor (KNN), decision tree, linear discriminant analysis (LDA), support vector machine (SVM), extreme gradient boosting (XGBoost), adaptive boosting (AdaBoost), random forest, softmax classifier, and some methods in recent literature.
The rest of the paper is organized as follows. Section 2 briefly reviews some related works. The dataset used in the experiment, the concepts incorporated in the proposed approach, together with the proposed framework, are described in Section 3. Section 4 presents and discuss the simulation results, and Section 5 concludes the paper.

2. Related Works

This section briefly presents some recently proposed related works, including heart disease prediction models and relevant autoencoder studies. Predicting heart disease using machine learning can effectively reduce the number of deaths attributed to the disease. Recently, researchers have employed ML to detect heart disease and identify the most relevant risk factors [26,27]. For example, Mohan et al. [4] proposed an enhanced heart disease prediction model to identify the essential features in the data that would improve the prediction accuracy. The authors employed a hybrid random forest that achieved a classification accuracy of 88.7%.
Haq et al. [28] employed a feature selection approach and an improved logistic regression algorithm to predict heart disease, and an accuracy of 89% was obtained. Furthermore, Samuel et al. [29] proposed a novel method for heart disease prediction using a fuzzy analytic hierarchy process (Fuzzy AHP) and an artificial neural network (ANN), and their technique obtained an accuracy of 91%. Meanwhile, heart disease prediction remains a problem for researchers and scientists because most of the predictive models developed so far have only achieved moderate performance [1].
Autoencoder neural networks have shown remarkable performance in unsupervised feature learning. Several research works have studied the sparse autoencoder (SAE) [20]. Meanwhile, different variants of autoencoders can learn diverse characteristics of the input data. Yang et al. [30] proposed a representation learning approach by serially connecting marginalized autoencoder (MDA) and stacked robust autoencoder with graph regularisation (SRAG) to utilize the learning capabilities of the different autoencoders. The MDA was used to capture domain invariant features, while the SRAG improved the quality of feature representation. Enhanced feature representation was obtained by serially connecting the two autoencoders.
Furthermore, Du et al. [31] developed a sparse autoencoder (SAE) combined with an improved multiple kernel learning strategy to extract high-level discriminative features. The SAE learned hidden representation from facial images and classified the data efficiently. The approach achieved comparable performance with existing facial expression recognition methods. Tai et al. [32] stacked several denoising sparse autoencoders for high-resolution range profile recognition. The training process consists of layer-by-layer pre-training. The training strategy ensured the model converged faster and obtained excellent performance.
In another study, Xiong and Lu [33] proposed a multistage approach to predict Parkinson’s disease. The study combined the adaptive grey wolf optimization algorithm and sparse autoencoder. The former was used to identify the predictor candidate subset, and the latter to extract the latent representation of the candidate features for efficient supervised classification. The approach outperformed the benchmark models and showed the importance of representation learning in complex disease diagnosis. Furthermore, Mienye et al. [34] proposed a model to predict heart disease by combining SAE and ANN. The SAE was enhanced to obtain a good feature representation, and the ANN was employed to classify the learned features. The approach achieved a good prediction performance.
Furthermore, Ebiaredoh-Mienye et al. [35] developed an approach to predict diseases using sparse autoencoder and softmax classifier. Unlike other sparse autoencoders that achieve sparsity by penalizing the neural network weights, their technique penalizes the hidden layer activations. When tested on three disease datasets, the method showed improvement over other benchmarked algorithms. Meanwhile, our research aims to obtain an efficient feature representation of the heart disease data by optimizing the parameters of the stacked sparse autoencoder using the particle swarm optimization technique.

3. Materials and Methods

In this section, a description of the heart disease datasets employed in this research is presented. Additionally, this section provides a detailed background of the autoencoder neural network. Lastly, Section 3.3 presents the proposed heart disease prediction methodology, including how the PSO is integrated into the autoencoder architecture with a step-by-step description of the method applied.

3.1. Datasets

This research uses two publicly available heart study datasets, including the Cleveland heart disease dataset [36] and the Framingham heart dataset [37]. The Cleveland dataset consists of 303 samples and 14 attributes, and there are 165 sick patients and 138 healthy patients in the dataset. Meanwhile, the Framingham dataset contains 4238 samples and 16 features, with 3594 healthy patients and 644 sick patients. The attributes of both datasets include risk factors from physical examination of the patients, behavioral risk factors, and medical risk factors. Table 1 and Table 2 show the features of the Cleveland and Framingham datasets, respectively.
The datasets are split into 70% for training and 30% for testing and validation. Both datasets contain missing values, and we utilize mean imputation to handle the missing data. Imputation is a method used to replace missing data with a substitute value based on other available details in order to preserve most of the information in the dataset. Data imputation also ensures proper data are used for machine learning since missing data can be problematic for many machine learning algorithms [38]. Furthermore, we utilize MinMaxScaler to transform the features into the range [0, 1] because the proposed autoencoder involves using a sigmoid activation function that outputs values between 0 and 1.

3.2. Autoencoder

Autoencoders are unsupervised artificial neural networks used for representation learning. The autoencoder architecture is designed to impose a bottleneck in the network that forces a compressed knowledge representation of the input. Therefore, the correlation between the input features is learned and reconstructed. Autoencoders have an encoder-decoder structure [39]. The encoder maps the original input x to the hidden layer, which is considered a latent space representation. The decoder then reconstructs this latent representation into   x ^ [20]. The encoding and decoding processes are defined in Equations (1) and (2), respectively.
h = σ W x + b ,
x ^ = σ W h + b ,
where x = ( x 1 ,   x 2 ,   x n ) represents the input data vector, h = ( h 1 ,   h 2 ,   h n ) denotes the low-dimensional vector obtained from the hidden layer, and x ^ = (   x ^ 1 ,     x ^ 2 ,     x ^ n ) is the reconstructed input. W   and W are weight matrices, b and b are bias vectors, and σ represents the sigmoid activation function, i.e., σ = 1 1 + e x . We employ the mean squared error function as the reconstruction error function between h and x ^ :
E = 1 N i = 1 N | | x ^ i x i | | 2 ,
Overfitting is a common challenge that occurs when training autoencoder networks. An efficient way to solve this problem is by applying a weight penalty to the cost function [34]:
E = 1 N i = 1 N 1 2 | | x ^ i x i | | 2 + λ 2 W 2 + W 2 ,  
where λ is the weight attenuation coefficient. Additionally, a sparse penalty term is introduced in the autoencoder hidden layer to achieve better feature learning under sparse constraints and avoid a situation where the autoencoder copies the input data to the output [40]. Assuming p j ^ denotes the average activation of the hidden layer neurons, it is defined as p j ^ = 1 N i = 1 N h j x i , and ρ is the sparsity proportion, often a small positive value near 0. To achieve sparsity, we limit ρ j ^ = ρ [41], and the Kullback-Leibler (KL) divergence is introduced to the loss function as a regularization term:
K L ( ρ ^ | | ρ ) = j = 1 K ρ log ρ ρ j ^ + 1 ρ log 1 ρ 1 ρ j ^ ,
where K is the number of hidden neurons. Hence, the loss function of the sparse autoencoder now contains three parts: the mean squared error, the weight attenuation, and the sparsity regularization parts:
E = 1 N i = 1 N 1 2 | | x ^ i x i | | 2 + λ 2 W 2 + W 2 + β K L ( ρ ^ | | ρ ) ,
where β is the sparsity regularisation parameter. Furthermore, we stack multiple sparse autoencoders to achieve enhanced feature learning. The architecture entails connecting the encoding layer to the input layer of the next sparse autoencoder, thereby ensuring the network achieves better representation learning.

3.3. Proposed Methodology

In this study, a stacked sparse autoencoder network is proposed. In the SSAE network, the hidden layer of the preceding sparse autoencoder serves as input to the next sparse autoencoder. Additionally, it is essential to note that after training the various sparse autoencoders, their decoder layers are deleted since the learned features reside in the hidden layer. The final hidden layer is then connected to the softmax classifier, which performs the classification. Therefore, the proposed SSAE network consists of the trained sparse autoencoders and softmax classifier. Backpropagation is applied to finetune the parameters of the entire network with the training instances and their labels. The finetuning stage considers the various layers of the network as one model. Assuming y 1 ,   y 2 ,   , y m represents the target variables of the training data, the cost function of the entire network is defined as:
E = 1 m i = 1 m j = 1 N 1 y i = j   log e θ j T x i l = 1 N e θ l T x i ,
where 1 · represents the indicator function, i.e., 1 y i = j = 1   i f   y = 1 , and 1 y i = j = 0   i f   y j , N denotes the number classes, and θ i denotes the weight matrix linking the i t h output unit.
Furthermore, we utilize the PSO to optimize the SSAE’s parameters, i.e., optimal weights and bias values. The choice of weights and bias is crucial in training robust neural networks. Usually, the determination of optimal network configuration is achieved by trial and error by searching a space of potential hyperparameter combinations [42]. The conventional manual parameter settings and grid search method requires domain knowledge and is error-prone, consumes time, and is computationally expensive [43]. Meanwhile, genetic algorithm (GA) have been used to optimize some neural networks, including autoencoders [44,45]. However, PSO is a less complex algorithm compared to GA, and it converges faster. Hence, in this research, we employ the PSO to optimize the SSAE.
The PSO is a heuristic search method inspired by biological populations’ swarming or cooperative behavior [46]. It is an intelligent evolutionary optimization method that iteratively optimizes a problem to enhance a candidate solution regarding a specified quality measure. Every particle is considered a potential solution to the optimization problem and consists of speed, position, and a fitness value computed by the optimization function [47]. Therefore, the introduction of PSO in the SSAE network improves the convergence speed and ensures the autoencoder learns the most representative features.
In the PSO algorithm, there are m particles in the swarm. Every particle p i t is randomly initialized to separate positions x i t in the search space, with speed v i t . After that, the SSAE would be configured for training by applying the value of x i t . Then, the final SAE’s reconstruction error (6) is utilized as the PSO’s fitness value. The PSO aims to search the space for the minimum fitness value. Assuming l i t represents the fitness value. At every iteration, l i t is computed using Equation (6), and if it is the minimum fitness value that exists in the swarm, then the position x i t is set as the best position g . Additionally, if l i t is the minimum fitness value estimated by particle p i t , then x i t is used as the best position b i t of the particle. After that, the speed of all the particles is updated based on their location, current speed, and global swarm information. The update of the speed and position of the particles is achieved using Equations (8) and (9), respectively [48]:
v t + 1 = ω v t + c 1 r 1 b i x t + c 2 r 2 g x t ,
x t + 1 = x t + v t + 1 ,
where v t + 1 represent the speed of the particle at time t + 1 , ω represents the inertia weight, v t is the current speed, c 1 and c 2 are acceleration factors, and their values are 1.5 and 2, respectively, r 1 and r 2 are random numbers between 0 and 1. After the given number of iterations, g contains the best solution, which is utilized to initialize the parameters of the SSAE. Lastly, the algorithm ends the optimization when the evaluation function converges, and then the optimal solution is given as the initial weights and bias; else, the iterations continue. Algorithm 1 summarises the steps followed to achieve the optimization of the SSAE by the PSO.
Algorithm 1. Proposed methodology.
Set the parameters of the PSO, i.e., c 1 , c 2 , ω , the maximum number of iterations T m a x , and the number of particles m
  • for each particle p i t
  • perform SSAE training using the current particle
  • compute   E using (6) and set fitness value l i t = E
  • update the best position and speed of each particle based on the fitness function, i.e., using (8) and (9)
  • if the evaluation function converges, then return the optimal solution   g as the SSAE weights and bias
  • else return to step 1 and continue the iteration
Meanwhile, when training deep neural networks such as the SSAE proposed in this paper, internal covariate shift slows down the training process and affects the model’s performance. Batch normalization was proposed in 2015 by Ioffe and Szegedy [49] and has been used in recent research works [50,51] to make neural networks faster and more stable via normalizing the layers’ inputs through re-scaling and re-centering. Internal covariate shift is the change in the distribution of the neural network activations resulting from the parameter updates during training. It occurs when there is a change in the neural network’s input distribution. Usually, when there is a change in input distribution, the hidden layers tend to adapt to this new distribution, which slows down the neural network training. When the training is slowed down, it takes longer for the network to converge to a global minimum. Therefore, we apply batch normalization to improve the SSAE’s speed and stability by normalizing layer inputs to prevent this problem. Batch normalization also acts as a regularizer and reduces the need to apply the dropout regularization technique. Similar to the dropout technique, research has shown that batch normalization also prevents overfitting in deep neural networks [52]. Lastly, Figure 1 shows a flowchart of the proposed approach to visually demonstrate all the steps taken to arrive at the PSO optimized SSAE model.
In implementing the proposed heart disease prediction method shown in Figure 1, the data are first preprocessed as discussed in Section 3.1. The preprocessed data are split into training and testing sets. The proposed model is pre-trained and finetuned to obtain optimal performance. The pre-training is employed for the SSAE training, whereas the backpropagation is utilized to finetune the model. As shown in Figure 1, several steps are taken to arrive at the final optimal SSAE model.
  • Step 1: The encoder maps the input data into a low dimensional space, while the decoder reconstructs the data. In order to obtain optimal reconstruction, we employed backpropagation to minimize the reconstruction error. During this step, the bias and weight matrices of the encoder and decoder are obtained using backpropagation.
  • Step 2: In this step, the sparse autoencoders are stacked to form the SSAE. It is achieved by stacking new hidden and output layers into the first autoencoder, where the hidden layer of the first autoencoder serves as input to the new autoencoder. In this study, only three autoencoders are stacked in order not to make the model too complex.
  • Step 3: The final hidden layer is then connected to the softmax classifier to complete the SSAE network. Backpropagation is applied to finetune the parameters of the entire network, including the weights and biases.
  • Step 4: The fourth step involves using PSO to search for the optimal initial weights and bias values of the SSAE. In the PSO implementation, we initialized parameters such as the number of particles, the maximum number of iterations, the acceleration factors, etc., as stated in Algorithm 1. During the backpropagation finetuning in step 3, the PSO is employed to find the optimal SSAE parameters, which leads to enhanced performance.

4. Results and Discussion

This section presents and discusses the results obtained from the experiments. In the experiment, accuracy, precision, sensitivity, and F-measure were used as the performance evaluation metrics, and their formulas are as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N   ,
P r e c i s i o n = T P T P + F P   ,
S e n s i t i v i t y = T P T P + F N     ,
F   m e a s u r e = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l ,
where T P represents the number of true positives and T N denotes the number of true negatives. In contrast, F P and F N represent the number of false-positive and false negatives, respectively. Other performance evaluation metrics used in this work are the receiver operating characteristic (ROC) curve and the area under the ROC curve (AUC). All the simulations were carried out using the Python programming language and the scikit-learn machine learning library. The experimental results are shown in Table 3 and Table 4, where the proposed method is compared with some well-known algorithms. The algorithms include KNN [53], LR [54], LDA [55], SVM [56], decision tree [57], XGBoost [58], random forest [59], softmax [60], and AdaBoost [61]. It is necessary to note that since our problem is that of binary classification, hence, the softmax classifier mentioned above has K = 2 classes. This softmax algorithm is a type of logistic regression that is optimized by SGD. However, it differs from the logistic regression mentioned above as the latter uses the limited-memory Broyden-Fletcher-Goldfarb-Shanno (L-BFGS) optimization algorithm. Furthermore, we employed the well-known GridSearch tuning technique for hyperparameter tuning of the baseline models to obtain the best performance.
Table 3 shows the performance of the various algorithms when trained with the Framingham dataset. The proposed approach achieved an accuracy of 0.973, a precision of 0.948, a sensitivity of 1.000, and an F-measure of 0.973. From Table 4, the proposed SSAE-PSO method obtained an accuracy of 0.961, precision of 0.930, a sensitivity of 0.988, and an F-measure of 0.958. In both instances, the proposed SSAE-PSO outperformed the other algorithms. However, it is not sufficient to conclude that the proposed method is better using just two datasets. Hence, to further validate the robustness of the proposed method, we performed experiments using two different disease datasets, i.e., the cervical cancer risk factors [62] and chronic kidney disease [63] datasets, and the results are shown in Table 5 and Table 6, respectively.
The experimental results in Table 5 and Table 6 demonstrate that our approach also obtained superior performance in both instances. Worthy of note is that the proposed method performed better than the ensemble learning algorithms that have obtained state-of-the-art performance in diverse applications. The improved performance in the proposed method can be attributed to the enhanced feature learning in the SSAE, which was made possible by the parameter optimization performed by the PSO.
Furthermore, test sensitivity is an essential metric in medical diagnosis. It measures the proportion of positive or sick patients in the dataset that are correctly predicted [64]. In the Framingham and CKD datasets, the proposed approach obtained sensitivity values of 100%, which implies all the positive samples were correctly predicted in the test set. Additionally, the proposed method achieved a sensitivity of 98.8% and 97.8% for the Cleveland and cervical cancer datasets, respectively. Additionally, the high sensitivity shows the proposed method obtained only a few false negatives. The ROC plots and AUC values of the classifiers in the various test instances are shown below to indicate the superior performance of the proposed approach. We compare the proposed method with the ensemble learning algorithms, which are known to achieve high performance.
Figure 2, Figure 3, Figure 4 and Figure 5 visibly demonstrate the proposed method’s superior performance over some powerful ensemble algorithms in terms of ROC curves and AUC values. This shows the PSO’s capability to optimize the SSAE parameters, which resulted in improved feature learning and classification. Furthermore, the proposed method is benchmarked with some recently developed heart disease prediction models that utilized the Cleveland dataset. The methods include an enhanced hybrid random forest [4], an improved logistic regression model with feature selection (FS) [28], an intelligent heart disease prediction approach using a feature selection method and Naïve Bayes model [65], a system based on Fuzzy AHP and ANN [29], an XGBoost with enhanced data resampling technique [2], a method based on mutual information (MI) and deep neural network (DNN) [66], an improved decision tree ensemble [10], a sparse autoencoder combined with ANN [34], a hybrid GA with fuzzy classifier [67], and a fuzzy ensemble system with GA and modified dynamic PSO [68]. The comparison is shown in Table 7. Meanwhile, the Cleveland dataset was chosen because it is a more popular dataset for heart disease prediction and has been widely utilized in numerous studies.
The classification accuracy of 96.1% makes the proposed approach better than some state-of-the-art methods tabulated in Table 7. An essential contribution of this work is that the proposed method learns efficiently using few samples, which is vital because data collection is a complex and expensive task. Hence, a technique that can learn from a few instances is more desirable. Furthermore, the experimental results obtained from this research have shown that the proposed PSO based stacked sparse autoencoder yields efficient representation learning of the input data, which resulted in improved classification performance.
Meanwhile, heart disease detection is considered among the most vital studies in clinical data analysis [70], and its prediction is difficult due to the many contributory risk factors. Hence, researchers and scientists have studied different ML methods to accurately predict this disease. This study’s optimized feature learning method ensures that the input features’ salient relationships are learned to achieve efficient and reliable prediction. Thus, the proposed approach could assist clinicians in the early detection of disease risk when deployed as a clinical decision support system, especially in developing countries with limited healthcare infrastructure. Since there is a high possibility of curing diseases when they are detected early, clinicians can use tools like this for early detection and adequately proffer the needed treatment and recommend lifestyle changes where necessary.

5. Conclusions

Heart disease prediction is a critical challenge in clinical machine learning. This research proposed an effective heart disease prediction method using a stacked sparse autoencoder with a softmax layer. In the proposed approach, the hidden layer of the last sparse autoencoder was connected to the softmax classifier, which made up the SSAE network. A notable contribution in this study is the integration of the PSO to optimize the parameters of the SSAE, which enhanced the feature learning and improved the classification performance. The approach introduced in this study ensures the problem of internal covariate shift associated with multilayer autoencoders such as the SSAE is prevented. When experimented on the Framingham and Cleveland heart datasets, the proposed method achieved an accuracy of 0.973 and 0.961, respectively, which showed significant improvement compared to other ML algorithms and recent studies. In the future, we will study the impact of feature learning on different classification algorithms and focus on stacking different variants of autoencoders and observing the effect on classification performance.

Author Contributions

Conceptualization, I.D.M. and Y.S.; methodology, I.D.M. and Y.S.; software, I.D.M.; validation, I.D.M. and Y.S.; formal analysis, I.D.M. and Y.S.; investigation, I.D.M. and Y.S.; resources, Y.S.; data curation, I.D.M.; writing—original draft preparation, I.D.M.; writing—review and editing, Y.S.; visualization, I.D.M.; supervision, Y.S.; project administration, Y.S.; funding acquisition, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the South African National Research Foundation under Grant 120106 and Grant 132797, and in part by the South African National Research Foundation Incentive under Grant 132159.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chicco, D.; Jurman, G. Machine learning can predict survival of patients with heart failure from serum creatinine and ejection fraction alone. BMC Med. Inform. Decis. Mak. 2020, 20, 16. [Google Scholar] [CrossRef]
  2. Fitriyani, N.L.; Syafrudin, M.; Alfian, G.; Rhee, J. HDPM: An Effective Heart Disease Prediction Model for a Clinical Decision Support System. IEEE Access 2020, 8, 133034–133050. [Google Scholar] [CrossRef]
  3. Benjamin, E.J.; Muntner, P.; Alonso, A.; Bittencourt, M.S.; Callaway, C.W.; Carson, A.P.; Chamberlain, A.M.; Chang, A.R.; Cheng, S.; Das, S.R.; et al. Heart Disease and Stroke Statistics-2019 Update: A Report From the American Heart Association. Circulation 2019, 139, e56–e528. [Google Scholar] [CrossRef]
  4. Mohan, S.; Thirumalai, C.; Srivastava, G. Effective Heart Disease Prediction Using Hybrid Machine Learning Techniques. IEEE Access 2019, 7, 81542–81554. [Google Scholar] [CrossRef]
  5. Li, J.P.; Haq, A.U.; Din, S.U.; Khan, J.; Khan, A.; Saboor, A. Heart Disease Identification Method Using Machine Learning Classification in E-Healthcare. IEEE Access 2020, 8, 107562–107582. [Google Scholar] [CrossRef]
  6. Mdhaffar, A.; Bouassida Rodriguez, I.; Charfi, K.; Abid, L.; Freisleben, B. CEP4HFP: Complex Event Processing for Heart Failure Prediction. IEEE Trans. NanoBioscience 2017, 16, 708–717. [Google Scholar] [CrossRef]
  7. Jin, B.; Che, C.; Liu, Z.; Zhang, S.; Yin, X.; Wei, X. Predicting the Risk of Heart Failure With EHR Sequential Data Modeling. IEEE Access 2018, 6, 9256–9261. [Google Scholar] [CrossRef]
  8. Ali, L.; Rahman, A.; Khan, A.; Zhou, M.; Javeed, A.; Khan, J.A. An Automated Diagnostic System for Heart Disease Prediction Based on x2 Statistical Model and Optimally Configured Deep Neural Network. IEEE Access 2019, 7, 34938–34945. [Google Scholar] [CrossRef]
  9. Alaa, A.M.; Bolton, T.; Angelantonio, E.D.; Rudd, J.H.F.; van der Schaar, M. Cardiovascular disease risk prediction using automated machine learning: A prospective study of 423,604 UK Biobank participants. PLoS ONE 2019, 14, e0213653. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Mienye, I.D.; Sun, Y.; Wang, Z. An improved ensemble learning approach for the prediction of heart disease risk. Inform. Med. Unlocked 2020, 20, 100402. [Google Scholar] [CrossRef]
  11. Buchan, T.A.; Ross, H.J.; McDonald, M.; Billia, F.; Delgado, D.; Duero Posada, J.G.; Luk, A.; Guyatt, G.H.; Alba, A.C. Physician Prediction versus Model Predicted Prognosis in Ambulatory Patients with Heart Failure. J. Heart Lung Transplant. 2019, 38, S381. [Google Scholar] [CrossRef]
  12. Oh, S.L.; Jahmunah, V.; Ooi, C.P.; Tan, R.-S.; Ciaccio, E.J.; Yamakawa, T.; Tanabe, M.; Kobayashi, M.; Acharya, U.R. Classification of heart sound signals using a novel deep WaveNet model. Comput. Methods Programs Biomed. 2020, 105604. [Google Scholar] [CrossRef] [PubMed]
  13. Bengio, Y.; Courville, A.; Vincent, P. Representation Learning: A Review and New Perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [PubMed]
  14. Kasongo, S.M.; Sun, Y. A Deep Learning Method With Filter Based Feature Engineering for Wireless Intrusion Detection System. IEEE Access 2019, 7, 38597–38607. [Google Scholar] [CrossRef]
  15. Kasongo, S.M.; Sun, Y. Performance Analysis of Intrusion Detection Systems Using a Feature Selection Method on the UNSW-NB15 Dataset. J. Big Data 2020, 7, 105. [Google Scholar] [CrossRef]
  16. Reddy, G.T.; Reddy, M.P.K.; Lakshmanna, K.; Kaluri, R.; Rajput, D.S.; Srivastava, G.; Baker, T. Analysis of Dimensionality Reduction Techniques on Big Data. IEEE Access 2020, 8, 54776–54788. [Google Scholar] [CrossRef]
  17. Wickramasinghe, C.S.; Marino, D.L.; Manic, M. ResNet Autoencoders for Unsupervised Feature Learning From High-Dimensional Data: Deep Models Resistant to Performance Degradation. IEEE Access 2021, 9, 40511–40520. [Google Scholar] [CrossRef]
  18. Zhang, C.; Cheng, X.; Liu, J.; He, J.; Liu, G. Deep Sparse Autoencoder for Feature Extraction and Diagnosis of Locomotive Adhesion Status. J. Control. Sci. Eng. 2018, 2018, 8676387. [Google Scholar] [CrossRef]
  19. Ng, A. Sparse Autoencoder. 2011. Available online: https://web.stanford.edu/class/cs294a/sparseAutoencoder.pdf (accessed on 6 June 2020).
  20. Liu, J.; Li, C.; Yang, W. Supervised Learning via Unsupervised Sparse Autoencoder. IEEE Access 2018, 6, 73802–73814. [Google Scholar] [CrossRef]
  21. Mienye, I.D.; Ainah, P.K.; Emmanuel, I.D.; Esenogho, E. Sparse noise minimization in image classification using Genetic Algorithm and DenseNet. In Proceedings of the 2021 Conference on Information Communications Technology and Society (ICTAS), Durban, South Africa, 10–11 March 2021; pp. 103–108. [Google Scholar] [CrossRef]
  22. Mienye, I.D.; Sun, Y.; Wang, Z. Improved Predictive Sparse Decomposition Method with Densenet for Prediction of Lung Cancer. Int. J. Comput. 2020, 533–541. [Google Scholar] [CrossRef]
  23. Lin, C.-J.; Jeng, S.-Y. Optimization of Deep Learning Network Parameters Using Uniform Experimental Design for Breast Cancer Histopathological Image Classification. Diagnostics 2020, 10, 662. [Google Scholar] [CrossRef]
  24. Kaur, S.; Aggarwal, H.; Rani, R. Hyper-parameter optimization of deep learning model for prediction of Parkinson’s disease. Mach. Vis. Appl. 2020, 31, 32. [Google Scholar] [CrossRef]
  25. Bickel, S.; Brückner, M.; Scheffer, T. Discriminative Learning Under Covariate Shift. J. Mach. Learn. Res. 2009, 10, 2137–2155. [Google Scholar]
  26. Pasha, S.J.; Mohamed, E.S. Novel Feature Reduction (NFR) Model With Machine Learning and Data Mining Algorithms for Effective Disease Risk Prediction. IEEE Access 2020, 8, 184087–184108. [Google Scholar] [CrossRef]
  27. Ali, S.A.; Raza, B.; Malik, A.K.; Shahid, A.R.; Faheem, M.; Alquhayz, H.; Kumar, Y.J. An Optimally Configured and Improved Deep Belief Network (OCI-DBN) Approach for Heart Disease Prediction Based on Ruzzo–Tompa and Stacked Genetic Algorithm. IEEE Access 2020, 8, 65947–65958. [Google Scholar] [CrossRef]
  28. Haq, A.U.; Li, J.P.; Memon, M.H.; Nazir, S.; Sun, R. A Hybrid Intelligent System Framework for the Prediction of Heart Disease Using Machine Learning Algorithms. Mob. Inf. Syst. 2018, 2018, 3860146. [Google Scholar] [CrossRef]
  29. Samuel, O.W.; Asogbon, G.M.; Sangaiah, A.K.; Fang, P.; Li, G. An integrated decision support system based on ANN and Fuzzy_AHP for heart failure risk prediction. Expert Syst. Appl. 2017, 68, 163–172. [Google Scholar] [CrossRef]
  30. Yang, S.; Zhang, Y.; Zhu, Y.; Li, P.; Hu, X. Representation learning via serial autoencoders for domain adaptation. Neurocomputing 2019, 351, 1–9. [Google Scholar] [CrossRef]
  31. Du, L.; Wu, Y.; Hu, H.; Wang, W. Self-adaptive weighted synthesised local directional pattern integrating with sparse autoencoder for expression recognition based on improved multiple kernel learning strategy. IET Comput. Vis. 2020, 14, 73–83. [Google Scholar] [CrossRef]
  32. Tai, G.; Wang, Y.; Li, Y.; Hong, W. Radar HRRP target recognition based on stacked denosing sparse autoencoder. J. Eng. 2019, 2019, 7945–7949. [Google Scholar] [CrossRef]
  33. Xiong, Y.; Lu, Y. Deep Feature Extraction From the Vocal Vectors Using Sparse Autoencoders for Parkinson’s Classification. IEEE Access 2020, 8, 27821–27830. [Google Scholar] [CrossRef]
  34. Mienye, I.D.; Sun, Y.; Wang, Z. Improved sparse autoencoder based artificial neural network approach for prediction of heart disease. Inform. Med. Unlocked 2020, 18, 100307. [Google Scholar] [CrossRef]
  35. Ebiaredoh-Mienye, S.A.; Esenogho, E.; Swart, T.G. Integrating Enhanced Sparse Autoencoder-Based Artificial Neural Network Technique and Softmax Regression for Medical Diagnosis. Electronics 2020, 9, 1963. [Google Scholar] [CrossRef]
  36. UCI Machine Learning Repository: Heart Disease Data Set. Available online: http://archive.ics.uci.edu/ml/datasets/Heart+Disease (accessed on 9 April 2020).
  37. Framingham Heart Study Dataset. Available online: https://kaggle.com/amanajmera1/framingham-heart-study-dataset (accessed on 24 January 2020).
  38. Khan, S.I.; Hoque, A.S.M.L. SICE: An improved missing data imputation technique. J. Big Data 2020, 7, 37. [Google Scholar] [CrossRef]
  39. Pathirage, C.S.N.; Li, J.; Li, L.; Hao, H.; Liu, W.; Wang, R. Development and application of a deep learning–based sparse autoencoder framework for structural damage identification. Struct. Health Monit. 2019, 18, 103–122. [Google Scholar] [CrossRef] [Green Version]
  40. Yan, B.; Han, G. Effective Feature Extraction via Stacked Sparse Autoencoder to Improve Intrusion Detection System. IEEE Access 2018, 6, 41238–41248. [Google Scholar] [CrossRef]
  41. Li, G.; Han, D.; Wang, C.; Hu, W.; Calhoun, V.D.; Wang, Y.-P. Application of deep canonically correlated sparse autoencoder for the classification of schizophrenia. Comput. Methods Programs Biomed. 2020, 183, 105073. [Google Scholar] [CrossRef] [PubMed]
  42. Doaud, M.; Mayo, M. Using Swarm Optimization To Enhance Autoencoders Images. arXiv 2018, arXiv:1807.03346. [Google Scholar]
  43. Fernandes Junior, F.E.; Yen, G.G. Particle swarm optimization of deep neural networks architectures for image classification. Swarm Evol. Comput. 2019, 49, 62–74. [Google Scholar] [CrossRef]
  44. Chiroma, H.; Noor, A.S.M.; Abdulkareem, S.; Abubakar, A.I.; Hermawan, A.; Qin, H.; Hamza, M.F.; Herawan, T. Neural Networks Optimization through Genetic Algorithm Searches: A Review. Appl. Math. Inf. Sci. 2017, 11, 1543–1564. [Google Scholar] [CrossRef]
  45. Feng, X.; Zhao, J.; Kita, E. Genetic Algorithm-based Optimization of Deep Neural Network Ensemble. Rev. Socionetwork Strat 2021, 15, 27–47. [Google Scholar] [CrossRef]
  46. Yang, X.-S. Chapter 8-Particle Swarm Optimization. In Nature-Inspired Optimization Algorithms, 2nd ed.; Yang, X.-S., Ed.; Academic Press: Cambridge, MA, USA, 2021; pp. 111–121. [Google Scholar]
  47. Kennedy, J. Particle Swarm Optimization. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2010; pp. 760–766. [Google Scholar]
  48. Qolomany, B.; Maabreh, M.; Al-Fuqaha, A.; Gupta, A.; Benhaddou, D. Parameters optimization of deep learning models using Particle swarm optimization. In Proceedings of the 2017 13th International Wireless Communications and Mobile Computing Conference (IWCMC), Valencia, Spain, 26–30 June 2017; pp. 1285–1290. [Google Scholar] [CrossRef] [Green Version]
  49. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  50. Hiriyannaiah, S.; Srinivas, A.M.D.; Shetty, G.K.; Siddesh, G.M.; Srinivasa, K.G. Chapter 4-A computationally intelligent agent for detecting fake news using generative adversarial networks. In Hybrid Computational Intelligence; Bhattacharyya, S., Snášel, V., Gupta, D., Khanna, A., Eds.; Academic Press: Cambridge, MA, USA, 2020; pp. 69–96. [Google Scholar]
  51. Theodoridis, S. Chapter 18-Neural Networks and Deep Learning. In Machine Learning, 2nd ed.; Theodoridis, S., Ed.; Academic Press: Cambridge, MA, USA, 2020; pp. 901–1038. [Google Scholar] [CrossRef]
  52. Garbin, C.; Zhu, X.; Marques, O. Dropout vs. batch normalization: An empirical study of their impact to deep learning. Multimed Tools Appl. 2020, 79, 12777–12815. [Google Scholar] [CrossRef]
  53. Altman, N.S. An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression. Am. Stat. 1992, 46, 175–185. [Google Scholar] [CrossRef] [Green Version]
  54. Cramer, J.S. The Origins of Logistic Regression; Social Science Research Network: Rochester, NY, USA, 2002. [Google Scholar] [CrossRef] [Green Version]
  55. Hastie, T.; Tibshirani, R.; Friedman, J. Linear Methods for Classification. In The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Hastie, T., Tibshirani, R., Friedman, J., Eds.; Springer: New York, NY, USA, 2009; pp. 101–137. [Google Scholar] [CrossRef]
  56. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  57. Krzywinski, M.; Altman, N. Classification and regression trees. Nat. Methods 2017, 14, 757–758. [Google Scholar] [CrossRef]
  58. Breiman, L. Arcing the Edge; Technical Report 486; Statistics Department, University of California: Los Angeles, CA, USA, 1997. [Google Scholar]
  59. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  60. Bridle, J.S. Probabilistic Interpretation of Feedforward Classification Network Outputs, with Relationships to Statistical Pattern Recognition. In Neurocomputing; Springer: Berlin/Heidelberg, Germany, 1990; pp. 227–236. [Google Scholar] [CrossRef]
  61. Schapire, R.E. A brief introduction to boosting. In Proceedings of the 16th Internation Joint Conference on Artificial Intelligence, IJCAI, Stockholm, Sweden, 31 July–6 August 1999; Volume 2, pp. 1401–1406. [Google Scholar]
  62. UCI Machine Learning Repository: Cervical Cancer (Risk Factors) Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/Cervical+cancer+%28Risk+Factors%29 (accessed on 15 April 2021).
  63. UCI Machine Learning Repository: Chronic_Kidney_Disease Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/Chronic_Kidney_Disease (accessed on 20 July 2021).
  64. Mienye, I.D.; Sun, Y. Performance analysis of cost-sensitive learning methods with application to imbalanced medical data. Inform. Med. Unlocked 2021, 25, 100690. [Google Scholar] [CrossRef]
  65. Repaka, A.N.; Ravikanti, S.D.; Franklin, R.G. Design And Implementing Heart Disease Prediction Using Naives Bayesian. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019; pp. 292–297. [Google Scholar] [CrossRef]
  66. Ali, L.; Bukhari, S.A.C. An Approach Based on Mutually Informed Neural Networks to Optimize the Generalization Capabilities of Decision Support Systems Developed for Heart Failure Prediction. IRBM 2020. [Google Scholar] [CrossRef]
  67. Reddy, G.T.; Reddy, M.P.K.; Lakshmanna, K.; Rajput, D.S.; Kaluri, R.; Srivastava, G. Hybrid genetic algorithm and a fuzzy logic classifier for heart disease diagnosis. Evol. Intel. 2020, 13, 185–196. [Google Scholar] [CrossRef]
  68. Paul, A.K.; Shill, P.C.; Rabin, M.R.I.; Murase, K. Adaptive weighted fuzzy rule-based system for the risk level assessment of heart disease. Appl. Intell. 2018, 48, 1739–1756. [Google Scholar] [CrossRef]
  69. Ali, L.; Niamat, A.; Khan, J.A.; Golilarz, N.A.; Xingzhong, X.; Noor, A.; Nour, R.; Bukhari, S.A.C. An Optimized Stacked Support Vector Machines Based Expert System for the Effective Prediction of Heart Failure. IEEE Access 2019, 7, 54007–54014. [Google Scholar] [CrossRef]
  70. Kim, J.O.R.; Jeong, Y.-S.; Kim, J.H.; Lee, J.-W.; Park, D.; Kim, H.-S. Machine Learning-Based Cardiovascular Disease Prediction Model: A Cohort Study on the Korean National Health Insurance Service Health Screening Database. Diagnostics 2021, 11, 943. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart of the proposed approach.
Figure 1. Flowchart of the proposed approach.
Electronics 10 02347 g001
Figure 2. ROC curve of the classifiers using the Framingham dataset.
Figure 2. ROC curve of the classifiers using the Framingham dataset.
Electronics 10 02347 g002
Figure 3. ROC curve of the classifiers using the Cleveland dataset.
Figure 3. ROC curve of the classifiers using the Cleveland dataset.
Electronics 10 02347 g003
Figure 4. ROC curve of the classifiers using the cervical cancer dataset.
Figure 4. ROC curve of the classifiers using the cervical cancer dataset.
Electronics 10 02347 g004
Figure 5. ROC curve of the classifiers using the CKD dataset.
Figure 5. ROC curve of the classifiers using the CKD dataset.
Electronics 10 02347 g005
Table 1. Features in the Cleveland datasets.
Table 1. Features in the Cleveland datasets.
S/NFeatureDescription
1AgeThe individual’s age, in years
2SexThe sex of the patient (1 = male, 0 = female)
3Chest pain typeThe type of chest pain experienced by the patient (1 = typical angina, 2 = atypical angina, 3 = non-anginal pain, 4 = asymptomatic)
4Resting Blood PressureResting blood pressure value of the patient in mmHg
5Serum CholesterolSerum cholesterol in mg/dL
6Fasting Blood SugarFasting blood sugar > 120 mg/dL (1 = true, 0 = false)
7Resting ECGECG result (0 = normal, 1 = ST-T abnormality, 2 = left ventricular hypertrophy)
8MaxHeart rateMaximum heart rate achieved by the patient
9Exercise induced anginaExercise induced angina (1 = true, 0 = false)
10ST depressionST depression induced by exercise relative to rest
11SlopeThe slope of the peak exercise ST segment (1 = upsloping, 2 = flat, 3 = downsloping)
12Number of vesselsNumber of major vessels (0–3) colored by fluoroscopy
13ThalassemiaType of thalassemia disorder (3 = normal, 6 = fixed disorder, 7 = reversible disorder)
14Target variableThe status of heart disease diagnosis (0 = absence, 1, 2, 3, 4 = present)
Table 2. Features in the Framingham dataset.
Table 2. Features in the Framingham dataset.
S/NFeatureDescription
1SexPatient’s gender
2AgePatient’s age
3EducationEducational level (1 = high school, 2 = GED certificate, 3 = vocational training, 4 = college degree)
4currentSmokerWhether the individual smokes or not
5cigsPerDayThe average number of cigarettes the patient smokes per day
6BPMedsWhether the individual is on BP medication or not
7prevalentStrokeWhether the individual previously had a stroke or not
8PrevalentHypWhether the individual is hypertensive
9DiabetesWhether the individual is diabetic
10totCholTotal cholesterol level
11sysBPSystolic blood pressure
12diaBPDiastolic blood pressure
13BMIBody mass index
14heartrateThe patient’s heart rate
15GlucoseGlucose level
16Target variable (TenYearCHD)Whether or not the patient has a ten-year risk of coronary heart disease
Table 3. Assessment of the performance of the algorithms using the Framingham dataset.
Table 3. Assessment of the performance of the algorithms using the Framingham dataset.
AlgorithmAccuracyPrecisionSensitivityF-Measure
KNN0.7830.8010.7890.800
LR0.8380.8390.8410.839
LDA0.8250.8060.8280.817
SVM0.8050.8370.8230.830
Decision tree0.7490.7350.7450.739
Softmax classifier0.7940.8030.7860.794
XGBoost0.9180.9380.9720.955
Random forest0.8830.9110.9030.907
AdaBoost0.8950.9550.8890.920
Proposed SSAE + PSO0.9730.9481.0000.973
Table 4. Assessment of the performance of the algorithms using the Cleveland dataset.
Table 4. Assessment of the performance of the algorithms using the Cleveland dataset.
AlgorithmAccuracyPrecisionSensitivityF-Measure
KNN0.6240.6080.5940.601
LR0.7830.7900.7810.785
LDA0.7810.8040.7920.798
SVM0.7960.8000.7890.794
Decision tree0.7100.6990.7080.703
Softmax classifier0.7380.7150.7000.708
XGBoost0.8750.8640.9400.900
Random forest0.8680.9140.8870.900
AdaBoost0.8770.9320.8650.897
Proposed SSAE + PSO0.9610.9300.9880.958
Table 5. Assessment of the performance of the algorithms using the cervical cancer dataset.
Table 5. Assessment of the performance of the algorithms using the cervical cancer dataset.
AlgorithmAccuracyPrecisionSensitivityF-Measure
KNN0.9560.9130.8300.870
LR0.9400.9420.9780.959
LDA0.9420.8760.9040.890
SVM0.9330.9180.9200.919
Decision tree0.8920.9100.9020.906
Softmax classifier0.9380.8410.9240.881
XGBoost0.9660.9030.9170.910
Random forest0.9640.8550.9120.883
AdaBoost0.9550.8600.9120.885
Proposed SSAE + PSO0.9880.9840.9780.981
Table 6. Assessment of the performance of the algorithms using the CKD dataset.
Table 6. Assessment of the performance of the algorithms using the CKD dataset.
AlgorithmAccuracyPrecisionSensitivity F-Measure
KNN0.9250.9100.8940.903
LR0.9270.9140.8800.897
LDA0.9000.8960.8970.896
SVM0.8960.9070.9000.904
Decision tree0.9100.9210.9210.921
Softmax classifier0.9300.9240.9170.920
XGBoost0.9400.9400.9320.936
Random Forest0.9350.9460.9300.938
AdaBoost0.9470.9470.9730.960
Proposed SSAE + PSO0.9820.9741.0000.987
Table 7. Performance comparison with other heart disease studies.
Table 7. Performance comparison with other heart disease studies.
Author(s)MethodAccuracyPrecisionSensitivityF-Measure
Mohan et al. [4]Hybrid random forest0.8840.9010.9280.900
Haq et al. [28]Improved Logistic Regression Model + FS0.890-0.770-
Repaka et al. [65]Naïve Bayes + FS0.8977---
Samuel et al. [29]Fuzzy AHP + ANN0.910---
Ali et al. [69]Optimized SVM0.9220.8291.000-
Li et al. [5]FS + SVM0.923-0.98-
Fitriyani et al. [2]XGBoost + Resampling0.9840.9850.9830.983
Ali and Bukhari [66]MI + DNN0.933-0.902-
Mienye et al. [10]Ensemble learning0.9300.9600.9100.930
Mienye et al. [34]SAE + ANN0.9000.8900.9100.900
Reddy et al. [67]Hybrid GA + Fuzzy Classifier0.900-0.910-
Paul et al. [68]Ensemble Fuzzy system + GA + PSO0.923---
Our approachSSAE + PSO0.9610.9300.9880.958
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mienye, I.D.; Sun, Y. Improved Heart Disease Prediction Using Particle Swarm Optimization Based Stacked Sparse Autoencoder. Electronics 2021, 10, 2347. https://doi.org/10.3390/electronics10192347

AMA Style

Mienye ID, Sun Y. Improved Heart Disease Prediction Using Particle Swarm Optimization Based Stacked Sparse Autoencoder. Electronics. 2021; 10(19):2347. https://doi.org/10.3390/electronics10192347

Chicago/Turabian Style

Mienye, Ibomoiye Domor, and Yanxia Sun. 2021. "Improved Heart Disease Prediction Using Particle Swarm Optimization Based Stacked Sparse Autoencoder" Electronics 10, no. 19: 2347. https://doi.org/10.3390/electronics10192347

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop