Next Article in Journal
A Multimodal Deep Learning Approach to Predicting Systemic Diseases from Oral Conditions
Next Article in Special Issue
Challenges for Artificial Intelligence in Recognizing Mental Disorders
Previous Article in Journal
Correction Equation for Hemoglobin Values Obtained Using Point of Care Tests—A Step towards Realistic Anemia Burden Estimates
Previous Article in Special Issue
Asynclitism and Its Ultrasonographic Rediscovery in Labor Room to Date: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ensemble Model for Diagnostic Classification of Alzheimer’s Disease Based on Brain Anatomical Magnetic Resonance Imaging

by
Yusera Farooq Khan
1,
Baijnath Kaushik
1,
Chiranji Lal Chowdhary
2,* and
Gautam Srivastava
3,4,5
1
School of CSE, Shri Mata Vaishno Devi University, Katra 182320, India
2
School of Information Technology and Engineering, Vellore Institute of Technology, Vellore 632014, India
3
Department of Mathematics and Computer Science, Brandon University, Brandon, MB R7A 6A9, Canada
4
Research Centre for Interneural Computing, China Medical University, Taichung 40402, Taiwan
5
Department of Computer Science and Math, Lebanese American University, Beirut 1102, Lebanon
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(12), 3193; https://doi.org/10.3390/diagnostics12123193
Submission received: 27 October 2022 / Revised: 8 November 2022 / Accepted: 15 November 2022 / Published: 16 December 2022

Abstract

:
Alzheimer’s is one of the fast-growing diseases among people worldwide leading to brain atrophy. Neuroimaging reveals extensive information about the brain’s anatomy and enables the identification of diagnostic features. Artificial intelligence (AI) in neuroimaging has the potential to significantly enhance the treatment process for Alzheimer’s disease (AD). The objective of this study is two-fold: (1) to compare existing Machine Learning (ML) algorithms for the classification of AD. (2) To propose an effective ensemble-based model for the same and to perform its comparative analysis. In this study, data from the Alzheimer’s Diseases Neuroimaging Initiative (ADNI), an online repository, is utilized for experimentation consisting of 2125 neuroimages of Alzheimer’s disease (n = 975), mild cognitive impairment (n = 538) and cognitive normal (n = 612). For classification, the framework incorporates a Decision Tree (DT), Random Forest (RF), Naïve Bayes (NB), and K-Nearest Neighbor (K-NN) followed by some variations of Support Vector Machine (SVM), such as SVM (RBF kernel), SVM (Polynomial Kernel), and SVM (Sigmoid kernel), as well as Gradient Boost (GB), Extreme Gradient Boosting (XGB) and Multi-layer Perceptron Neural Network (MLP-NN). Afterwards, an Ensemble Based Generic Kernel is presented where Master-Slave architecture is combined to attain better performance. The proposed model is an ensemble of Extreme Gradient Boosting, Decision Tree and SVM_Polynomial kernel (XGB + DT + SVM). At last, the proposed method is evaluated using cross-validation using statistical techniques along with other ML models. The presented ensemble model (XGB + DT + SVM) outperformed existing state-of-the-art algorithms with an accuracy of 89.77%. The efficiency of all the models was optimized using Grid-based tuning, and the results obtained after such process showed significant improvement. XGB + DT + SVM with optimized parameters outperformed all other models with an efficiency of 95.75%. The implication of the proposed ensemble-based learning approach clearly shows the best results compared to other ML models. This experimental comparative analysis improved understanding of the above-defined methods and enhanced their scope and significance in the early detection of Alzheimer’s disease.

1. Introduction

Alzheimer’s disease (AD) can be treated with the use of healthcare informatics. The Alzheimer’s enigma has troubled both researchers and clinicians. A German doctor Alois Alzheimer over a century ago spotted anomalies in brain sections of a patient with dementia. Researchers tried to study these strange anomalies known as plaques and tangles in the brain. The brain with dementia is characterized by wiped memories and intellect getting destroyed, thereby resulting in memory failure [1]. Dementia that attacks memory, self-awareness and rational intellect is accompanied by plaques and tangles in the brain as a part of aging. As Alzheimer’s disease progresses, brain tissues shrink and die, which in turn causes the gradual loss of cognitive ability, difficulty in thinking, decision-making, and confusion, leading to ultimate brain atrophy [2]. Early signs are the building-up of plaques around the neurons. These plaques are made up of proteins which stop their actual functioning as soon as they undergo mutation [3].
The World Health Organization (WHO) has observed that by 2030, the worldwide economic influence of Alzheimer’s disease (AD) will exceed USD 2 trillion per year [4]. According to the Alzheimer’s Association (https://www.alz.org/, accessed on 14 November 2022), the disease has become the 6th foremost reason for death in the United States. More than 16 million individuals are given unpaid care and, predictably, 18.5 billion hours of care costs nearly USD 234 billion. As per the survey, one in three senior citizens died of AD or another dementia [5]. The death rate due to AD is more than the deaths caused by breast cancer and prostate cancer. In 2021, it was predicted to cost USD 355 billion, and by 2050, it might increase to USD 1.1 trillion.
Alzheimer’s & Related Disorders Society of India (ARDSI) [6] predicts that the number of people living with dementia would rise from 7.6 million in 2030 to 14.3 million by 2050, putting an enormous burden on families. Figure 1 represents the estimated Indian population with dementia from 2005 to the near future 2050 taken from ARDSI [6]. Certainly, it is a necessity to detect and predict initial changes in AD so that cost-effective treatments and therapies can be applied to address AD and other related disorders. Despite all advancements in the research to improve our understanding of neurodegeneration, no medication exists [7]. Therefore, it is very important to detect and diagnose this devastating condition as early as possible. AD is a progressive neurodegenerative disorder marked by the build-up of insoluble mutant proteins known as plaques and tangles [8]. Plaques are mysterious peptides called Amyloid-β (Aβ) [9]. These peptides are formed when two enzymes sequentially breakdown a protein called Amyloid precursor protein: β and γ secretase [10]. Additional fragments are formed as a result of this cleavage, which may contribute to the disease’s pathogenesis. Due to its tendency to fold incorrectly, Aβ tends to clump together and form solubilizing polymers [11]. Few of them convert into intricate fibrils known as plaques which accumulate in the brain. Because of plaque deposits, the synapse starts to fail, and irregular configurations of activity emerge. Ultimately, the brain becomes unable to process and store information accurately. An additional significant feature of AD is neurodegeneration, neuronal loss [12], and destruction, which initiates other proteins called Tau [13], which also seem to be a constituent of massive jumbles called tangles. In a brain cell, particles are passed via an axon on a chain of tracks called microtubules. These tubular structures are stabilized by Tau [14]. Although Tau develops in AD, the microtubules that hold it in place are reformed, enabling it to break off and migrate from the axon to the cell body. Like amyloid-β, they also stick to deposition called neurofibrillary tangles. Eventually, these processes kill neurons which is a gradual and irreversible process [15]. Neurodegeneration leads to brain atrophy differentiating the normal brain from the diseased brain.
The most important and familiar tool for understanding and studying neurodegeneration and neurodegenerative diseases is neuroimaging [16] which includes multiple imaging techniques such as computed tomography (CT), dopamine transporter scan (DatScan) magnetic resonance imaging (MRI), positron emission tomography (PET), and Single-photon emission computed tomography (SPECT) [17]. Neuroimaging gives a comprehensive understanding of the potential mechanisms underlying these neurodegenerative diseases such as AD, Parkinson’s disease (PD) [18], Attention-Deficit/Hyperactivity Disorder (ADHD) [19], and Autism Spectrum Disorder (ASD) [20] so that better clinical trials are developed and deployed to improve diagnosis and effective healthcare. These large-scale neuroimaging databases may be used to train and evaluate diagnostics using Artificial intelligence and Machine learning methods for neurodegenerative diseases (e.g., Alzheimer’s disease). Predicting Alzheimer’s disease with machine learning has been the subject of several research studies.
However, scrutinizing previous studies, some gaps are identified related to the diagnosis of AD which encourages us to work in the direction to overcome those drawbacks. Taking this into consideration, this paper presents an effective ensemble-based technique and provides an in-depth comparison of Machine learning methods to predict AD, as well as future research possibilities. Ensemble learning is the process of transforming numerous weak models into a single robust model. This can be achieved using a variety of ensemble strategies, including bagging, boosting, and stacking. Ensemble learning integrates the information to acquire new information and enhances the performance accuracy of predictions done adaptively through voting schemes. The fundamental goal of voting is to increase generalization by balancing individual model errors, particularly when the models perform well on a predictive modelling task. Moreover, in the Ensemble approach, if it is found that the same model performs well when using different hyperparameters, then combining multiple fits of the same machine-learning algorithm is beneficial. When compared to a single model, an ensemble model always provides more accurate results and has greater applicability.
This paper provides a comprehensive evaluation of learning approaches for Alzheimer’s prediction, as well as potential future developments.
The main contributions of this article can be understood through the points mentioned as under:
  • This study evaluated the performance of various machine learning techniques, i.e., 11 models which lack in the state-of-the-art literature.
  • To enhance the diagnostic performance of AD, this article presents an ensemble technique formed by combining SVM kernel master demonstration, along with extreme gradient boosting transitioning sequential model to concur the hidden difficult extractions in the dataset. The proposed architecture (XGB + DT + SVM) uses a Decision Tree to reduce invariability in model demonstration and performance. Overall, the proposed model not only decreases inheriting variance but enhances tough sample prediction using boosting.
  • This article performed a comparative analysis of algorithms before and after hyper parameter tuning accomplished through Grid search and the outcomes are validated using k-fold cross-validation.
  • To ensure the generalization of the considered models, this article evaluated the efficiency of the same using statistical test, i.e., Friedman’s rank test and t-test.
  • Various performance metrics such as receiver operating characteristics (ROC) curve, sensitivity, specificity and log loss are used to compare the proposed model with other ML models.
  • This article also performed a comparison of the proposed model with existing state-of-the-art studies to prove its efficiency in detecting AD.
  • The article also focuses on the future scope that should be addressed for reliable AD diagnosis.
The remaining part of the paper is organized as follows: Section 2 discusses related work. Section 3 discusses the existing approaches and dataset. Section 4 presents a description of the experimental analysis. Section 5 discusses the comparative performance evaluation of each method of the framework and Section 6 concludes the paper following the future work in Section 7.

2. Related Work

As with other neurodegenerative illnesses, Alzheimer’s disease causes persistent and irreversible cognitive deterioration, culminating in brain shrinkage. It is an expensive medical condition that places a substantial financial strain on families and care givers. As neurodegeneration is an irreversible process, it cannot be treated. For this reason, researchers are focusing on the early identification and prediction of neurodegeneration to give optimal treatment to those suffering from neurodegenerative disorders. Substantial research has been performed and continues to be conducted in this field to enhance the early identification and prediction of a better diagnostic assistance system.
Roman Casanova et al. [21] used a large-scale regularization approach based on logistic regression to classify MRI scans to determine a person’s level of thinking and reasoning. The authors used a regularization approach on a dataset of cognitively Normal Controls (NC) and disease patients (AD). Their results showed 85.7% sensitivity, 82.9% specificity, and 90% accuracy, respectively. In a study by Costafreda et al. [22], MRI data were used to identify people with mild mental impairment who were at risk of developing Alzheimer’s disease. The authors used SVM to classify mild cognitive impairment as Alzheimer’s disease, with 80% and 77% sensitivity, respectively. Battineni et al. [23] tested four machine learning (ML) models, including NB, ANN, K-NN, and SVM, and the receiver operating characteristic (ROC) curve metric was used to verify the model performance. Each model was evaluated in three separate experiments. Firstly, the models were trained on manually selected features, and ANN produced the greatest ROC accuracy (0.812). In the second experiment, automated feature selection was performed via wrapping approaches, with the NB achieving the greatest ROC of 0.942 and finally, the ensemble model showed an ROC of 0.991.
Padilla et al. [24] examined neuroimaging data from MRI, PET, and SPECT scans of healthy and sick subjects. As a result, they used the Fisher discriminant ratio (FDR) and non-negative matrix factorization (NMF) to identify and categorize the most significant features. Their method yielded a 90% accuracy rate. Xin Liu et al. [25] investigated structural-MRI by using locally linear embedding (LLE) to transform multivariate MRI data into linear space, focusing on regional brain volume and cortical thickness. In this study, LLE was found to be useful in identifying Alzheimer’s disease. Gray et al. [26] performed a multi-modal classification of MRI data, focusing on MRI volumes and voxel-based PET signal intensities. Heung-II Suk et al. [27] proposed a deep learning feature-based strategy. The authors concentrated on grey matter tissue volumes, which are low-level MRI characteristics. They also used PET signal intensities in their method. The dataset was obtained from ADNI, and the experimental results had a 95.9% accuracy rate. Juergen Dukart et al. [28] proposed the use of SVM on information derived from the combination of MRI and FDG PET for detecting frontotemporal lobar degeneration. Using the ADNI dataset to train the classifier, they achieved 91% classification accuracy. Combining MRI and FED-PET findings may better determine Alzheimer’s disease (AD).
Liu et al. [29] introduced a novel hierarchical technique combining multi-level classifiers resulting in ensemble classification. They continued to add features from various parts of their brains until they were all in one place. They used imaging parameters such as spatially normalized grey matter (GM) from each MRI scan to achieve an accuracy of 92.0%. Authors presented MK-Boost to classify ADNI data using a Whole-Brain Hierarchical Network (WBHN) based on an Automated Anatomical Labelling atlas (ALL). With an AUC of 0.907, its AD/MCI classification model was 89.53% accurate. Tingting et al. [30] proposed a multimodality-based classification of Alzheimer’s disease and mild cognitive impairment (MCI). They used a linear regression model with group sparsity regularization on weights from the ADNI dataset to improve classification performance and identify syndrome-associated biomarkers useful for illness analysis. Zu et al. [31] classified Alzheimer’s disease and mild cognitive impairment using a new multimodal technique. As a result, label-aligned multiple-task feature extraction and multimodal categorization have been added. They combined selected multimodality characteristics using a multi-kernel SVM. They achieved 95.05% classification accuracy using the ADNI MRI and FDG-PET datasets.
According to Sorensen et al. [32], variations in volumetric alterations contribute to early cognitive decline. T1-weighted MRI images were analyzed from the ADNI database. Liu et al. [33] used a blend of various kernels to link edge characteristics with node architecture while analyzing MRI data from the Alzheimer’s disease Neuroimaging Initiative (ADNI). Their investigation discovered that 3D texture was more effective for categorization findings, with an accuracy of 91%. Wang et al. [34] used two data sources, from the OASIS and the Washington University School of Medicine. They formed an ethical committee and signed consent from each subject before entering the study. The study included four native hospitals.
Table 1 depicts a summary of existing work using MR imaging in Alzheimer’s disease categorization using multiple techniques
The after section comprises the methodology that we used for the classification of AD and MCI using different categories of algorithms.

3. Materials and Methods

This section comprises the methodology for comparative analysis of classification algorithms for predicting individuals with Alzheimer’s disease and mild cognitive impairment. The given procedure consists of different phases: (i) collection of MRI data (ii) classification, and (iii) comparison of performance among these algorithms.
In Section 3.1, the framework of this study is addressed. In Section 3.2, a description of the ADNI dataset, data pre-processing and feature extraction is given. In Section 3.3, we discussed the different learning techniques to identify individuals with AD and MCI.

3.1. Framework

Figure 2 presents the framework of the study. With the advent of ML throughout the last couple of years, a variety of methods have been introduced to enhance AD classification using neuroimaging as a biomarker. In this study, we used the axial MR Neuroimaging modality, and the dataset is divided into Train-set (70%) and Test-set (30%). Classification algorithms are trained for three classes of data by providing a training-set as input.

3.2. Dataset Description

Neuroimaging data were obtained from the ADNI website (http://adni.loni.usc.edu/ (accessed on 7 September 2022)). The ADNI initiative was launched by Dr. Michael W. Weiner as its primary investigator, at understanding the evolution from usual ageing to insignificant memory problems in Alzheimer’s disease. Along with the neuroimaging initiative, they also process the genetic and clinical information to analyze and measure various genes for determining the extent of genetic risk participating in the increase in the Alzheimer’s disease enigma. In this study, we have taken 2127 MRI (T1 and T2 weighted) data from ADNI. The downloaded dataset is categorized into three classes: 612 samples for the AD class, 538 samples for MCI, and 975 samples for the CN class as shown in Table 2.

3.2.1. Pre-Processing Steps

The data taken from the ADNI database require some basic tasks before it is presented to the algorithms:
(a)
Data filtering is performed based on the acquisition plane and ADNI phase. We selected the axial acquisition plan of the MR (T1 and T2 weighted) images and ADNI phase 3 for three classes: AD, CN, and MCI as shown in Figure 3.
(b)
The images are reshaped to 512 × 512 to provide the best quality pixel.
(c)
The labels about the classes of the images are extracted and put into the vector storage. The classes undergo label encoding, i.e., 0 for AD, 1 for MCI and 2 for CN.
(d)
The loaded images are standardized and normalized using a standard scaler and min-max scaler as the normalization functions.
(e)
Once the preprocessing of the data is completed, the data are split into training and testing to perform validation on unseen data.
The MRI axial scans are fed into the model as input, which is then converted into NumPy arrays. A NumPy array is an n-dimensional array that is represented as a grid of values with non-negative index tuples. The rank of the array gives the number of dimensions, and the shape of the tuple gives the array size. Figure 4 given below presents the workflow of this study as discussed below.
  • Data: Brain axial T1 and T2 weighted MRI data taken from the ADNI database is labelled into three classes such that AD CN, and MCI along with the preprocessing and data loading pipelines.
  • Data training: using VGG-16 [43] as a feature extractor primarily using the illustrations provided by the Imagenet on the ADNI dataset. After Feature Extraction using VGG-16, the extracted blob is then processed and sent to conventional machine learning models and proposed ensemble architecture.
  • Parameter optimization: post-training hyperparameter tuning is achieved using GridSearch.
  • Data Evaluation: after rigorous training, the models are tested and compared for overfitting. If overfitting is found, then retraining is done with reduced hyper-planes and data sampling. Else, a comparative analysis of the result is evaluated.

3.2.2. Feature Extraction

Feature Extraction is an integral aspect of image-based machine learning as it allows for the inculcation of the feature-wise similarity graph or correlation table available to the deep neural architectures such as VGG-16 that is utilized in this paper. In this study, a freezing bottleneck of VGG-16 is implemented, with the initial convolutional layer extraction being transformed into dense arrays of 1000 vectors for use in the machine learning proposition. In comparison to local filtering or Gaussian noise extraction methods, this method accelerates convergence and improves accuracy by 31.27% in aggregate. VGG-16 uses the internal feature extractor using images of size 100 × 100 × 1 which is self-convoluted 100 × 100 × 3 using 3 filtered convolutions with padding maintained at the same value. Further, the kernel association blob that is predicted or processed by the VGG-16 architecture is 2 × 2 × 512, when linearized using flatten and dense layers it is converted to 4096 vectors. This blob is fed into traditional machine learning models and the suggested ensemble architecture, from which AD, CN, and MCI predictions are derived. The proposed VGG-16 model uses label-encoded vectors for dependent variables and cross-channel normalization. In VGG-16, modelling, the batch size is 32, and the loss function is Cross-entropy loss. The feature stimulator extracts image information using 1 × 1, 3 × 3, and 5 × 5 kernels, as well as the gallort weight initializer and pre-trained Imagenet data.

3.3. Classification Models

In this subsection, classical machine learning models are discussed in brief. The mathematical background of various classical learning models discussed includes Decision Tree (DT), Random Forest (FR), Naïve Bayes (NB), K-Nearest Neighbor (K-NN), SVM (RBF kernel), SVM (Polynomial Kernel), SVM (Sigmoid kernel), Gradient Boost (GB) and Extreme Gradient Boosting (XGB).

3.3.1. Naive Bayes (NB)

NB is a machine learning algorithm whose work is based on the standards of the Bayes theorem named after Thomas Bayes which is based on conditional probability [44,45]. It is used as a classifier in face recognition, weather prediction, news classification, and medical diagnosis. The posterior probability in this algorithm is calculated using Equation (1).
P   ( C | X ) = ( P ( X | C ) P ( C ) P ( X )
Equation (1) above is used to depict posterior probability P   ( C | X ) where C is a class variable and X is a dependent feature vector. The posterior probability is denoted by P   ( C | X ) , likelihood is denoted by ( P ( X | C ) class prior probability by P ( C ) and predictor prior probability by P ( X ) .

3.3.2. Decision Tree (DT)

DT classifier has the inherent capability of eliminating useless elements from the input data, thereby increasing the performance of the algorithm [46]. The classification technique is based on non-parametric supervised learning, which predicts the class of given data by learning features from data in the form of a tree structure. This classification technique is used for predicting the type of class to which our data belong. Thus, it supports diagnosis decisions by dividing the dataset into smaller sets to specify a sequence of decisions and consequences. Numerous metrics can be used to determine the optimal method for splitting the records. These metrics are described in terms of pre-splitting and post-splitting class distributions of the documents. p ( i | t ) represents the percentage of records from class 1 located at node t. The degree of impurity in the child nodes influences several measures used to determine the ideal split. The degree of impurity is proportional to how distorted the distribution of classes is. Impurity measurements are given in Equations (2)–(4).
Entropy   ( t ) = i = 0 c 1   p ( i | t ) l o g 2 ( i | t )
Gini   ( t ) = 1 i = 0 c 1 [ p ( i | t ) ] 2
Classification   error   ( t ) = 1 m a x i [ p ( i | t ) ]
The input features for the DT models can be selected using Entropy (t) as stated in Equation (2), Gini Index (Gini (t)) using Equation (3) and the Classification error (t) calculated using Equation (4). In the above Equations, c is the number of classes.

3.3.3. Random Forest (RF)

Random Forest (RF) [47] is a classifier that works as a group of decorrelated decision trees to perform classification. It was created using multiple decision trees. RF is best used for medical diagnosis, remote sensing, and object detection in images. RF finalizes the decision by estimating average votes from different trees. The matrices containing sample features of the training dataset are given as input to the algorithm. For each sample, a DT is created to make a classification decision and the significance of every feature in each tree is computed. The prediction in RF is taken as an average of the trees and the formula in RF is given in Equation (5):
F ( x ) = 1 j   j = 1 j c j f u l l + k = 1 k (   1 j j = 1 j c o n t r i b u t i o n j   ( x , k ) )
where F ( x ) is the prediction function, j is the number of trees in the RF model and k is the total number of features.

3.3.4. K-Nearest Neighbor (K-NN)

The K-NN algorithm [48] works on the principle of the similarity measure. The number of available cases or nearest neighbours that contribute to voting is represented by K and the model organizes new data depending on similarity with the available cases. The algorithm is mostly used in search applications where the task is to search for similar data objects. The probabilistic interpretation of K-NN is discussed using Equations (6)–(8).
P ( x ) = k N V
P ( x | y ) = k 1 N 1 V
In Equations (6) and (7), the nearest neighbour from class 1 is denoted by k1 and N1 signifies the total sum of instances in class 1.
p ( x ) = k N   C d   R K D ( x )
In Equation (8), R K   D (x) is the distance between the estimation point and the kth closest neighbour and C d represents the volume of the sphere in the D dimension.

3.3.5. Support Vector Machine (RBF Kernel)

Support vector machine methods analyze data and sort them into categories based on supervised learning. SVM (RBF kernel) follows the use of a labelled training dataset as input to learn the classifier and produce input-output mapping functions [49]. Support vector machines are versatile because diverse kernel functions can be listed for the judgment functions. It is possible to define common kernels as well as custom kernels. RBF stands for the Radial Basis function kernel that works in infinite dimensions. Radial kernel SVM was used to deal with overlapping data. The radial kernel is mathematically represented in Equation (9).
k = e γ ( a b ) 2 = exp ( | | x y | | 2 2 σ 2 )
The kernel k determines the influence of each observation in the training dataset as stated in Equation (9). γ (gamma) is determined by cross-validation scales, perceived as an inverse of the radius.

3.3.6. Support Vector Machine (Polynomial Kernel)

SVM (Polynomial Kernel) calculates a high-dimensional relationship between observations [50,51]. A polynomial kernel is mathematically represented in Equations (10) and (11).
k = ( a × b + r ) d
For d = n,
k = ( i = 1 n   x i   y i   + r ) n
In the above Equations (10) and (11), a and b define different observations in the dataset, r defines the polynomial coefficient, and d sets the degree of the polynomial. The polynomial method sacrifices the linearity and uses higher-degree polynomial equations for higher-dimensional data.

3.3.7. Support Vector Machine (Sigmoid Kernel)

SVM (Sigmoid Kernel) [52] is known by the name hyperbolic tangent kernel and also as a multilayer perceptron kernel. The sigmoid function is used for activation, mathematically represented as:
k ( x i ,   x j = tanh ( α x i t x j + c )
In Equation (12), c is the intercepting constant and α is the slope. xi is the ith instance and xj is the jth feature of the ith instance. The Support vector sigmoid kernel is interesting because of its origin in neuronal network theory and has been found to perform well.

3.3.8. Gradient Boost (GB)

GB [53] is also founded on the theory of sequential ensemble or collection of predictors where the basic learners are created progressively in such a way that the existing base learner should be more competent than the previous one. Predictions are combined by weighted average or vote to provide an overall prediction. The goal of boosting algorithms is to minimize the empirical loss functional as given in Equation (13):
L ( f ) = 1 m i = 1 m Φ   ( y i ,       f ( x i ) )
In Equation (15), L ( f ) is an empirical loss function and is viewed as Functional Gradient Descent rather than a vector-valued function to minimize the loss L. xi is a training sample of m size.

3.3.9. Extreme Gradient Boosting (XGB)

The XGB classifier [54] is commonly used for complex and more convoluted data-driven real-world problems. Boosting comes under the category of ensemble learning techniques where weak learners are combined to become strong learners so that the accuracy of the model is increased. Thus, boosting is an efficient method for increasing the performance of classification. It is possible to execute an ensemble in one of two ways: using a sequential ensemble model, also known as boosting, where the weak learners are created sequentially during the training phase or using a parallel ensemble model, also known as bagging, where the weak learners are produced in parallel during the training phase. Mathematically, boosting can be represented with the help of Equation (14).
  f 1 ( x ) < f 0 ( x ) + h 1 ( x )
where f0 is considered the initial model to predict the target variable. h1 is used to fit the residual from the preceding steps.
The performance can be further improved by building a new model after the residuals of f1 as shown in Equation (15).
f 2 ( x ) < f 1 ( x ) + h 2 ( x )
The same can be repeated for ‘m’ iterations until it gets minimized.

3.3.10. Multi-Layer Perceptron Neural Network (MLP-NN)

In this study, we presented a multilayer perceptron neural network (MLP-NN), a feedforward artificial neural network [55] creating outputs from a collection of inputs. A neural network is a computational model inspired by the biological brain. This artificial neuron is called a perceptron as shown in Figure 5, which takes input convolved with weights. X1 and X2 up to Xn are the inputs to the perceptron that are convolved with weights W1 and W2 up to Wn.
The activation function takes an input of the calculated weighted sum to pass through it to provide a threshold value. Thus, an artificial neural network involves n such perceptrons. The architectural framework of the proposed MLP-NN is shown in Figure 6. The multi-layer perceptrons (MLPs) with backpropagation learning techniques have input layers with rectified linear units (ReLu), hidden layers with multiple dense units and ReLu activation function and output dense layers with a softmax activation function. The input layer has 2000 × 1024 units with a bias of 1024. Furthermore, four hidden layers (1024 × 1024, 1024 × 512, 512 × 512, and 512 × 256) units are included. Finally, the diagnostic output is generated using the output layer with 253 × 3 units and bias 3.

3.4. Proposed Ensemble Learning Approach: XGB + DT + SVM

Based on their performance, the three top-performing learning models, including Extreme Gradient Boosting, Decision Tree and SVM-Poly kernel are chosen for the ensemble learning model in the proposed work. In machine learning, voting ensemble techniques often provide more accurate results than a unified framework due to challenges such as the representation problem, the computing difficulty, and the statistical problem, which can be overcome by employing ensemble approaches. The basic example of stacking is an ensemble learning strategy in which additional innovations are created by combining the projections of many important functional, also known as classifiers, to teach a meta-classifier.
In this study, we implemented an ensemble of SVM (Poly kernel), Extreme Gradient Boosting Ensemble architecture, and a Decision Tree. This approach increases the performance by combining the advantages of a variety of machine learning architectures combining bagging, boosting and kernel-based rectilinear models such as Decision Tree, Extreme Gradient Boosting and SVM Poly Kernel. XGB + DT + SVM concept incorporates the Variance minimizer from the Decision tree, sampling accuracy enhancer via Extreme Gradient Boosting and Kernel Space Optimization vector using Poly kernel Support Vector Machine to achieve a more accurate average prediction than a single prediction obtained using a single kernel [56]. Any predictor of our selection can serve as the meta-classifier. By splitting the training set into multiple splits for the weak classifiers distribution, it would reduce time complexity or training time, allowing the task to be accomplished easily. Stage one filters are trained with the first split of your training examples. After that, utilize the stage one filters that have been learned to generate predictions about the remaining training splits. After that, the meta-classifier is trained using output from week learners. The confidence score (CS), a regressive value, measures the classification strength using the same feature vectors for classification distinction. The aggregating criteria are a decision of voting for each estimator output that is concatenated. Voting is calculated on the predicted probability of the output class. The architecture of the proposed XGB + DT + SVM is illustrated in Figure 7.
By aggregating the projected class or predicted probability based on voting, this proposed technique significantly improved the model’s performance. Therefore, if many base models are provided to the ensemble voting classifier, it ensures that the mistake is resolved by any model. The challenges of classification include the summation of votes for crisp class labels from different types and the prediction of the most votes in the class. We considered ensemble voting to overcome the error that occurs due to an independent classifier [57,58]. A voting classifier estimator is experimented with to train three kernels (estimators) to predict based on aggregating the findings of each kernel (base estimator).

3.5. Grid Search-Based Optimization on ML Methods and Ensemble Technique

The grid Search tuning approach aims to calculate hyperparameter values optimally. In this study, an exhaustive search was conducted on these models’ specific parameter values. This is performed by choosing suitable hyperparameters and tuning them to control the behaviour of the algorithm. The efficiency of our models is optimized by tuning without causing excessive variance or overfitting. Determining the ideal collection of hyperparameters can be computationally challenging. The hyperparameter values were set before learning begins. Regularization parameters in SVMs, k in k-Nearest Neighbors, and hidden layer count in multi-layer perceptron neural networks are other examples. A parameter, on the other hand, constitutes the internal feature of the model and can be measured from the data.

4. Experimental Analysis

In this study, a total of 2125 neuroimaging (MRI) data are used, divided into three classes: AD, CN, and MCI. Further, three different forms of classifiers inspired by (machine learning, neural-inspired, and ensemble learning algorithms) are designed for AD classification.

4.1. Implementation Setup

We trained our models on the Google Colab framework, providing computational tools and flexibility to adjust utilization limits and hardware availability using Python 3.7, TensorFlow, and the Scikit-learn library. Additionally, we used many Python libraries, including Numpy, Pandas, Matplotlib, and Keras. We also used pydicom, a Python program for working with the DICOM files. TensorFlow and TensorFlow-GPU versions 2.0.0 has also been used.

4.2. Performance Evaluation Metrics

The objective of this study is to identify the effectiveness of different algorithms for accurate classification and minimum log loss. Here, different models are investigated to achieve the accuracy, sensitivity, specificity, and log loss values for each classifier measuring the performance of each classifier on the test-set. In the context of statistical classification, a confusion matrix, also known as an error matrix, is used for the visualization of an algorithm’s performance. When evaluating a classification model’s performance on a known set of test data, confusion matrices (also known as contingency tables) are frequently used. The confusion matrix is listed in Table 3.
(a)
Accuracy: Classification models can be evaluated based on their accuracy. The fraction of correct predictions made by the model divided by all predictions is referred to as its accuracy as shown in Equation (16).
A c c u r a c y = T P + T N T P + T N + F P + F N  
(b)
Sensitivity: The value is represented numerically by dividing the number of correct positive predictions by the total number of positive values, as given in Equation (17).
Sensitivity = T P   T P + F N
(c)
Specificity: The value is mathematically represented by dividing the number of successful negative predictions by the total number of negative values. Sensitivity and specificity are statistical indicators of test result performance. The specificity is computed using the formula given in Equation (18).
Specificity = T N   T N + F P
(d)
Logarithmic loss (log loss): Using a log loss parameter, classification models that yield a probability value between zero and one may be evaluated for their efficiency. Increasing the logarithmic loss value is associated with increasing the anticipated probability value. The log loss action considers the confidence of the prediction in determining how incorrect classifications can be penalized [59]. The mathematical model used for computing logarithmic loss is shown in Equation (19).
L o g   L o s s   = 1 N   i 1 [ y i   L o g   ( p i ) + ( 1 y i )   Log   ( 1   p i ) ]
where the number of instances is represented by N, yi indicates the output of ith instance and pi represents the probability of 1 and (1 − pi) is the probability of 0. For the total number of cases, log loss accumulates the likelihood of a sample assuming both states 0 and 1.
The next section presents comprehensive results that were obtained after applying different classification approaches.

5. Results

The current section Comparison of classic ML models with proposed architecture on the ADNI data. The details of the respective results obtained by the different classifiers are presented below.

5.1. Parameter Optimization Results

Grid search determined the ideal model hyperparameters that lead to the most accurate predictions. Table 4 lists the grid parameters and optimized parameters of the classification models implemented in this study.

5.2. Classification Results

5.2.1. Before Hyperparameter Tuning

ML models and proposed XGB + DT + SVM were investigated in the context of AD and MCI classification-based neuroimaging analysis. The performance analysis based on accuracy, sensitivity, specificity, and log loss for three classes (AD, CN, and MCI) is given in Table 5. The NB prediction obtained the lowest accuracy of 72.22%, 77.98% and 77.58%, respectively.
The prediction model of Among SVM Polynomial achieved accuracies of 87.84%, 86.21% and 81.48%, respectively. The EGB model and GB obtained accuracies for AD 81.56% and 72.25%, respectively. The classification accuracy of the MLP-NN model for AD, CN and MCI is 89.20%, 81.56% and 84.30%, respectively. XGB + DT + SVM achieved better accuracy at 87.00%, 90.51% and 90.51% for AD, CN and MCI, respectively, compared to other algorithms. Overall, XGB + DT + SVM achieved an average accuracy of 89.77%. The sensitivity values for each class show the test’s ability to correctly identify the class in test data. Among the ML algorithms, SVM Polynomial and the Proposed model have the lowest log loss value of 0.29615 and 0.27795, respectively. Whereas Naive Bayes has the highest log loss value of 13.69922.

5.2.2. After Hyperparameter Tuning

Additionally, Table 6 compares the overall information scores results obtained with optimized parameters for ML classifiers and the proposed ensemble (XGB + DT + SVM) model. GridSearch enhanced the classification performance of these models by choosing the optimized gird of parameters. MLP-NN has shown better classification accuracy of 89.71% among ML models. The proposed ensemble model XGB + DT + SVM with hyperparameter tuning obtained the highest classification average accuracy of 95.75%, which outperformed all the models implemented in this study. XGB + DT + SVM achieved better accuracy 96.12%, 95% and 96.15% for AD, CN and MCI, respectively, compared to other algorithms The conjunction of results after the hyperparameter tuning reduces the variance in the categorical sensitivity and specificity and keeps a reasonable increase for all the different models.
XGB + DT + SVM used traditional, bagging and boosting approaches. The advantage of using traditional techniques is to ensure wide-scale globalization of features. For example, Boosting ensures heard features can be used to decode samples while Bagging ensures there are no overfitting biases. Using an ensemble method, we were able to lower the standard deviation and maintain high precision.
Figure 8 further illustrates the comparative plot, which offers a visual benchmark for contrasting the ML models and proposed model XGB + DT + SVM based on classification accuracy (before and after hyperparameter tuning).
In Figure 8, the authors compared the performance improvement among the models before and after GridSearch. During hyperparameter optimization with GridSearch, the SVM polynomial kernel was chosen among the RBF, polynomial, and sigmoid kernels. Figure 8b represents the performance of models after using GridSearch which includes the SVM polynomial kernel along with other models, whereas, in Figure 8a, all three SVM kernels are included along with other implemented models. Overall, it is demonstrated that model tuning enhances the performance and classification accuracy of all the models.

5.3. Statistical Evaluation of Outcomes

In this experiment, k-fold CV, Friedman’s rank test, and t-test were used to statistically validate the performance attained by various models for classifying AD, CN, and MCI.

5.3.1. Ten-Fold Cross-Validation

We performed 10 iterations of the k-fold cross-validation (k = 10) method, where each fold was determined to have a mean response value that was within a certain tolerance of the original value.
Cross-validation using a 10-fold set to tune the model on various splits and find the optimal kernel space boundary. The dataset is divided into 10 folds of roughly the same size. Cross-validation is applied to evaluate the performance of each model after hyperparameter tuning.
Table 7 lists the 10-fold cross-validation for each model reported in this study after parameter tuning. The entire data was randomly divided into 10 folds (k = 10). A larger K value results in a less biased model, whereas a smaller k value is comparable to that of the train-test split strategy. Next, k − 1 folds were used to fit the model, and the kth fold was used for validation. This procedure was repeated until each K-fold functioned as a test set. The average of reported scores is then calculated to validate our models. The analysis from this procedure demonstrates the highest mean value is achieved by XGB + DT + SVM (98.84%), whereas Nave Bayes achieved the lowest (67.04%) mean value for 10-fold validation. Finally, the results obtained demonstrate that hyperparameter tuning, when implemented, yields marginally enhanced accuracies in comparison to the values without optimized parameters.

5.3.2. Friedman’s Rank Test

In this work, the significance of the differences between these classifiers is measured using the Friedman rank test [60]. The Friedman test compares methods based on their rankings and is a non-parameter test that is used to represent a comparison of approaches among one another. In our study, we choose a significance threshold of =0.05.
Models are ranked using Friedman’s test, with the best-performing models placed at the top of the list.
The test determines the model’s average rank by taking into account M r j , where M r j is the rank of the ith models on the jth fold of the dataset. Each model, according to the null hypothesis, displays identical behaviour and, hence, has equivalent precision. Friedman’s rank test is distributed by x F M 2 having the r − 1 degree of freedom can be defined in Equation (20) given below: As seen in Equation (20), the distribution of Friedman’s rank test is x F M 2 with r − 1 degrees of freedom.
x F M 2 = s r ( r + 1 ) [ j M j 2 + r   ( r + 1 ) 2 8 ]
If s > 10 or/and r > 5, the chi-square distribution can be used to estimate the critical value of Friedman’s test statistics; otherwise, the table is referred to.
The p-value obtained after Friedman’s statistics test at α = 0.05 demonstrates that the null hypothesis cannot be accepted and substantiates the claim that there is a meaningful distinction between the various models. The estimated average rank and p-value of models are shown in Table 8, which demonstrates that the proposed ensemble method XGB + DT + SVM (with highest rank value = 8) and MLP-NN (with second highest rank value = 6.65) has outstanding accuracy performance than other models used in this study.

5.3.3. T-Test

Further, the accuracy mean difference is evaluated using a t-test on two experimental settings: (i) before hyperparameter tuning, and (ii) after hyperparameter tuning. A t-test is a type of inferential statistic used to determine if there is a significant mean difference in performance accuracy [61]. The mean difference of accuracies obtained before and after hyperparameter tuning, i.e., 76.39% and 80.448% demonstrates the enhanced results of (ii) after hyperparameter tuning, with a high statistical difference of -4.058 at p < 0.05 as shown in Table 9.

5.4. ROC Demonstration of Models Classification Abilities

In this subsection, we have shown the graphical illustration for the models in three categories. Receiver operating characteristic (ROC) curve plots for all classifiers are presented to indicate positive and negative results. The ROC curve is plotted to visualize how the predicted probabilities are compared to the true status [62].
False positive rates are plotted on the X-axis, while true positive rates are plotted on the Y-axis. Anticipated values are shown as TN, TP, FN, and FP on the ROC curve. There are three lines (blue, orange, and green) on the AUC graphs. In each ROC curve, the orange line represents Alzheimer’s disease (AD), the green line represents Control Normal (CN), and the blue line represents Mild Cognitive Impairment (MCI) as shown in Figure 9. The trade-off between sensitivity (TPR) and specificity (1–FPR) is demonstrated in the receiver operating characteristic (ROC) curve. Better results are achieved where the curve is closer to the top-left corner.
XGB + DT + SVM has shown more accurate predictions than other models. MLP-NN showed impressive prediction compared to other ML models. The ROC curve is independent of class distribution. As a result, it is advantageous to test classifiers that predict unusual events such as disease prediction.

5.5. Comparison with Existing Studies

In this subsection, the efficacy of the offered models is demonstrated by comparing their performance to that of earlier research. A comprehensive analysis of machine learning algorithms and a proposed approach has been performed for the classification of Alzheimer’s disease.
To ensure the generalization of these models, the efficiency models presented in this study are compared using the accuracy score obtained in previous studies on chosen benchmark dataset [63,64,65,66].
Among all the methods analyzed, EGB and MLP-NN have shown better performance. The outcomes of the comparison of EGB and XGB + DT + SVM with previous studies are listed in Table 10. The graphical demonstration of comparison between the results obtained by different models presented in this work with previous studies is shown in Figure 10. Among classical ML models used in this framework, XGB has shown better performance for the classification of AD, CN and MCI with an accuracy of 85.42%. In addition, the proposed XGB + DT + SVM and MLP-NN models outperformed existing approaches with an accuracy of 88.64% and 90.75%, respectively.

6. Conclusions

Computer-aided detection (CAD) systems are being developed to help radiologists in interpreting medical imaging findings and minimizing inter-reader variability more precisely. Artificial intelligence and machine learning play an important role in the development of CAD systems as it is extensively used to uncover useful image features from complicated datasets and assist in more reliable diagnosis.
The master model is suggested in this study, and the main principle of this strategy is that when two or more learning procedures combine, the performance of the final model will be improved. Some state-of-the-art learning algorithms and the proposed master model were compared to evaluate the performance of these learning algorithms to classify Alzheimer’s disease from neuroimaging data taken from ADNI. This study tries to improve the understanding of these algorithms and enhance their scope and significance in Alzheimer’s disease classification.
Comparative result analysis showed that after applying GridSearch for parameter optimization, the classification performance improved significantly. Of the various considered algorithms with optimized parameters, MLP-NN outperforms other algorithms with the best classification accuracy of 89.64%. Overall, the proposed XGB + DT + SVM with hyperparameter tuning achieved the highest accuracy of 95.75%, outperforming all the other classifiers on high-resolution brain MRI data.
In this study, statistical analysis of the results was carried out to validate the performance of the proposed models. The analysis has been performed using K-Fold cross-validation, Friedman’s Rank test and t-test. These statistical validations ensure the accuracy, effectiveness and efficiency of the models.
Further, the results of the proposed model (XGB + DT + SVM) and other state-of-the-art models considered in this study such as the neural-inspired model (MLP-NN) and XGB were compared with the results of existing studies. The results show that the models proposed in this study have outperformed the results of the existing studies.

7. Future Scope

From the state-of-the-art approaches, we can conclude that certain gaps need to be overcome in further research and more investigations are required to perform future work in the direction. There are several challenges identified. Firstly, the experimentation can be performed using a large number of samples to yield improved performance. Secondly, the work can be extended by using different neuroimaging modalities to determine the effective one. Thirdly, deep learning algorithms such as 3D-Convolution Neural Networks (3D-CNN), Recurrent Neural Networks (RNN) and pre-trained transfer learning models such as VGG-16, VGG-19, ResNet can be utilized to address the existing prediction and classification problems for better diagnosis. Fourthly, the AD data can be experimented with by including more viewing planes such as sagittal and coronal instead of considering just one single plane. Fifthly, the use of nature-inspired algorithms can be experimented with for optimization and also extract the relevant features from the data. Moreover, the work can be extended by considering the severity levels of AD, i.e., early mild cognitive impairment (EMCI), late mild cognitive impairment (LMCI), static mild cognitive impairment (SMCI) and progressive mild cognitive impairment (PMCI).

Author Contributions

Conceptualization: Y.F.K.; methodology, Y.F.K. and B.K.; software: Y.F.K. and C.L.C.; validation: B.K. and G.S.; formal analysis: Y.F.K., B.K., G.S. and C.L.C.; investigation: Y.F.K. and B.K.; resources: Y.F.K. and B.K.; data Curation: Y.F.K.; writing—original draft preparation: Y.F.K. and G.S.; review and editing: B.K. and C.L.C. All authors have read and agreed to the published version of the manuscript.

Funding

The authors received funding from Vellore Institute of Technology, Vellore for this research work.

Informed Consent Statement

Not applicable for this study.

Data Availability Statement

The dataset available online can be downloaded from the links given as follows. ADNI DataSet at the URL: https://ida.loni.usc.edu/login.jsp?project=ADNI (accessed on 10 July 2022).

Acknowledgments

Authors would like to thank the Vellore Institute of Technology, Vellore for supporting this work.

Conflicts of Interest

The authors declare that they have no conflicts of interest to report regarding the present study.

Appendix A

The acronyms for labels used in Table 1 are enlisted in Table A1.
Table A1. Acronym for labels used in Table 1.
Table A1. Acronym for labels used in Table 1.
LabelsAcronymLabelsAcronym
AFActivation functions8L-CNN8-layer convolution neural network
AIBLAustralian Imaging, Biomarker & Lifestyle Flagship Study of AgeingMIRIADMinimal Interval Resonance Imaging in
Alzheimer’s Disease dataset
LDMILlandmark-based deep multi-instance learningVBMVoxel-based morphometry
CLMConventional landmark-based morphometryMMSRMulti-modal sparse representation
SPASparse patch-based approachDL AlgoDeep learning algorithm
FSFisher scoreRVBRegional Voxel-based method
SUVRStandardized uptake value ratioGGMLGraph-guided multitask-task learning
N-CKN-number (combination of multi-kernels)E-MKEnsemble multi-kernel
RFERecursive feature eliminationOASISOpen Access Series of Imaging Studies
RFESRandom forest-based ensemble strategyGIMGini importance measure
lMRIlongitudinal-MRIRFLDRegression forest-based landmark detection model
AALAutomated Anatomical LabelingFWFilter Wrapper method
N-CKN-number (combination of multi-kernels)DMTFSDiscriminative Multi-Task Feature Selection
FAFisher AlgorithmWAWrapper Algorithm
MK-SVMMulti kernel- SVMPM-KernelProbabilistic multi-kernel
p-MCIProgressive-MCIL-SVMLagrangian-SVM
s-MCIStatic-MCILR, RFLogistic regression, Random Forest

References

  1. Hussain, R.; Zubair, H.; Pursell, S.; Shahab, M. Neurodegenerative Diseases: Regenerative Mechanisms and Novel Therapeutic Approaches. Brain Sci. 2018, 8, 177. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Yu, G.; Liu, Y.; Shen, D. Graph-guided joint prediction of class label and clinical scores for the Alzheimer’s disease. Brain Struct. Funct. 2016, 221, 3787–3801. [Google Scholar] [CrossRef] [Green Version]
  3. Montie, H.L.; Durcan, T.M. The Cell and Molecular Biology of Neurodegenerative Diseases: An Overview. Front. Neurol. 2013, 4, 194. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Alzheimer’s Association. 2014 Alzheimer’s Disease Facts and Figures. Alzheimer’s Dement. 2014, 10, e47–e92. [Google Scholar]
  5. Alzheimer Association. Available online: https://www.alz.org/ (accessed on 5 June 2022).
  6. Alzheimer’s and Related Disorder Society of India. Available online: https://ardsi.org/ (accessed on 7 January 2022).
  7. Und Halbach, O.V.B.; Schober, A.; Krieglstein, K. Genes, proteins, and neurotoxins involved in Parkinson’s disease. Prog. Neurobiol. 2004, 73, 151–177. [Google Scholar] [CrossRef]
  8. Jones, D.T.; Townley, R.A.; Graff-Radford, J.; Botha, H.; Knopman, D.S.; Petersen, R.C.; Jack, C.R.; Lowe, V.J.; Boeve, B.F. Amyloid- and tau-PET imaging in a familial prion kindred. Neurol. Genet. 2018, 4, e290. [Google Scholar] [CrossRef] [Green Version]
  9. Nelson, T.J.; Sen, A. Apolipoprotein E particle size is increased in Alzheimer’s disease. Alzheimer’s Dement. Diagn. Assess. Dis. Monit. 2019, 11, 10–18. [Google Scholar] [CrossRef]
  10. Saar, G.; Koretsky, A.P. Manganese Enhanced MRI for Use in Studying Neurodegenerative Diseases. Front. Neural Circuits 2019, 12, 114. [Google Scholar] [CrossRef] [Green Version]
  11. Tenreiro, S.; Eckermann, K.; Outeiro, T.F. Protein phosphorylation in neurodegeneration: Friend or foe? Front. Mol. Neurosci. 2014, 7, 42. [Google Scholar] [CrossRef] [Green Version]
  12. Gorman, A.M. Neuronal cell death in neurodegenerative diseases: Recurring themes around protein handling. J. Cell. Mol. Med. 2008, 12, 2263–2280. [Google Scholar] [CrossRef] [Green Version]
  13. Davatzikos, C.; Bhatt, P.; Shaw, L.M.; Batmanghelich, K.N.; Trojanowski, J.Q. Prediction of MCI to AD conversion via MRI, CSF biomarkers, and pattern classification. Neurobiol. Aging 2011, 32, 2322.e19–2322.e27. [Google Scholar] [CrossRef] [PubMed]
  14. Noble, W.; Hanger, D.P.; Miller, C.C.J.; Lovestone, S. The Importance of Tau Phosphorylation for Neurodegenerative Diseases. Front. Neurol. 2013, 4, 83. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Emard, J.-F.; Thouez, J.-P.; Gauvreau, D. Neurodegenerative diseases and risk factors: A literature review. Soc. Sci. Med. 1995, 40, 847–858. [Google Scholar] [CrossRef]
  16. Rathore, S.; Habes, M.; Iftikhar, M.A.; Shacklett, A.; Davatzikos, C. A review on neuroimaging-based classification studies and associated feature extraction methods for Alzheimer’s disease and its prodromal stages. NeuroImage 2017, 155, 530–548. [Google Scholar] [CrossRef]
  17. Shimizu, S.; Hirose, D.; Hatanaka, H.; Takenoshita, N.; Kaneko, Y.; Ogawa, Y.; Sakurai, H.; Hanyu, H. Role of Neuroimaging as a Biomarker for Neurodegenerative Diseases. Front. Neurol. 2018, 9, 265. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Luo, C.; Song, W.; Chen, Q.; Zheng, Z.; Chen, K.; Cao, B.; Jang, J.; Li, J.; Huang, X.; Gong, Q.; et al. Reduced functional connectivity in early-stage drug-naive Parkinson’s disease: A resting-state fMRI study. Neurobiol. Aging 2014, 35, 431–441. [Google Scholar] [CrossRef]
  19. Yin, W.; Li, T.; Mucha, P.J.; Cohen, J.R.; Zhu, H.; Zhu, Z.; Lin, W. Altered neural flexibility in children with attention-deficit/hyperactivity disorder. Mol. Psychiatry 2022, 91, 945–955. [Google Scholar] [CrossRef]
  20. Xie, Y.; Xu, Z.; Xia, M.; Liu, J.; Shou, X.; Cui, Z.; Liao, X.; He, Y. Alterations in Connectome Dynamics in Autism Spectrum Disorder: A Harmonized Mega- and Meta-analysis Study Using the Autism Brain Imaging Data Exchange Dataset. Biol. Psychiatry 2021, 91, 945–955. [Google Scholar] [CrossRef]
  21. Casanova, R.; Whitlow, C.T.; Wagner, B.; Williamson, J.D.; Shumaker, S.A.; Maldjian, J.A.; Espeland, M.A. High Dimensional Classification of Structural MRI Alzheimer’s Disease Data Based on Large Scale Regularization. Front. Neuroinform. 2011, 5, 22. [Google Scholar] [CrossRef] [Green Version]
  22. Costafreda, S.G.; Dinov, I.; Tu, Z.; Shi, Y.; Liu, C.-Y.; Kloszewska, I.; Mecocci, P.; Soininen, H.; Tsolaki, M.; Vellas, B.; et al. Automated hippocampal shape analysis predicts the onset of dementia in mild cognitive impairment. NeuroImage 2011, 56, 212–219. [Google Scholar] [CrossRef] [Green Version]
  23. Battineni, G.; Chintalapudi, N.; Amenta, F.; Traini, E. A comprehensive machine-learning model applied to magnetic resonance imaging (mri) to predict alzheimer’s disease (ad) in older subjects. J. Clin. Med. 2020, 9, 2146. [Google Scholar] [CrossRef] [PubMed]
  24. Padilla, P.; López, M.; Górriz, J.M.; Ramirez, J.; Salas-Gonzalez, D.; Alvarez, I. NMF-SVM based CAD tool applied to functional brain images for the diagnosis of Alzheimer’s disease. IEEE Trans. Med. Imaging 2011, 31, 207–216. [Google Scholar] [CrossRef] [PubMed]
  25. Liu, X.; Tosun, D.; Weiner, M.W.; Schuff, N. Alzheimer’s Disease Neuroimaging Initiative, Locally linear embedding (LLE) for MRI based Alzheimer’s disease classification. Neuroimage 2013, 83, 148–157. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Gray, K.R.; Aljabar, P.; Heckemann, R.A.; Hammers, A.; Rueckert, D. Alzheimer’s Disease Neuroimaging Initiative, Random forest-based similarity measures for multi-modal classification of Alzheimer’s disease. NeuroImage 2013, 65, 167–175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Suk, H.I.; Shen, D. Deep Learning-Based Feature Representation for AD/MCI Classification. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nagoya, Japan, 22–26 September 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 583–590. [Google Scholar]
  28. Dukart, J.; Mueller, K.; Barthel, H.; Villringer, A.; Sabri, O.; Schroeter, M.L. Alzheimer’s Disease Neuroimaging Initiative. Meta-analysis based SVM classification enables accurate detection of Alzheimer’s disease across different clinical centers using FDG-PET and MRI. Psychiatry Res. Neuroimaging 2013, 212, 230–236. [Google Scholar] [CrossRef] [PubMed]
  29. Liu, J.; Li, M.; Lan, W.; Wu, F.X.; Pan, Y.; Wang, J. Classification of Alzheimer’s disease using whole brain hierarchical network. IEEE/ACM Trans. Comput. Biol. Bioinform. 2016, 15, 624–632. [Google Scholar] [CrossRef] [PubMed]
  30. Ye, T.; Zu, C.; Jie, B.; Shen, D.; Zhang, D. Discriminative multi-task feature selection for multi-modality classification of Alzheimer’s disease. Brain Imaging Behav. 2016, 10, 739–749. [Google Scholar] [CrossRef]
  31. Zu, C.; Initiative, T.A.D.N.; Jie, B.; Liu, M.; Chen, S.; Shen, D.; Zhang, D. Label-aligned multi-task feature learning for multimodal classification of Alzheimer’s disease and mild cognitive impairment. Brain Imaging Behav. 2015, 10, 1148–1159. [Google Scholar] [CrossRef] [Green Version]
  32. Sørensen, L.; Igel, C.; Pai, A.; Balas, I.; Anker, C.; Lillholm, M.; Nielsen, M.; Alzheimer’s Disease Neuroimaging Initiative. Differential diagnosis of mild cognitive impairment and Alzheimer’s disease using structural MRI cortical thickness, hippocampal shape, hippocampal texture, and volumetry. NeuroImage Clin. 2017, 13, 470–482. [Google Scholar] [CrossRef]
  33. Liu, J.; Wang, J.; Tang, Z.; Hu, B.; Wu, F.X.; Pan, Y. Improving Alzheimer’s disease classification by combining multiple measures. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017, 15, 1649–1659. [Google Scholar] [CrossRef]
  34. Wang, S.-H.; Phillips, P.; Sui, Y.; Liu, B.; Yang, M.; Cheng, H. Classification of Alzheimer’s Disease Based on Eight-Layer Convolutional Neural Network with Leaky Rectified Linear Unit and Max Pooling. J. Med. Syst. 2018, 42, 85. [Google Scholar] [CrossRef]
  35. Purushotham, S.; Tripathy, B.K. Evaluation of Classifier Models Using Stratified Tenfold Cross Validation Techniques. In Proceedings of the International Conference on Computing and Communication Systems, Vellore, India, 9–11 September 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 680–690. [Google Scholar]
  36. Ahmed, O.B.; Mizotin, M.; Benois-Pineau, J.; Allard, M.; Catheline, G.; Amar, C.B.; Alzheimer’s Disease Neuroimaging Initiative. Alzheimer’s disease diagnosis on structural MR images using circular harmonic functions descriptors on hippocampus and posterior cingulate cortex. Comput. Med. Imaging Graph. 2015, 44, 13–25. [Google Scholar] [CrossRef]
  37. Chincarini, A.; Bosco, P.; Calvini, P.; Gemme, G.; Esposito, M.; Olivieri, C.; Rei, L.; Squarcia, S.; Rodriguez, G.; Belotti, R.; et al. Local MRI analysis approach in the diagnosis of early and prodromal Alzheimer’s disease. Neuroimage 2011, 58, 469–480. [Google Scholar] [CrossRef] [PubMed]
  38. Salini, A.; Jeyapriya, U. A majority vote based ensemble classifier for predicting students academic performance. Int. J. Pure Appl. Math. 2018, 118, 24. [Google Scholar]
  39. Heidari, E.; Sobati, M.A.; Movahedirad, S. Accurate prediction of nanofluid viscosity using a multilayer perceptron artificial neural network (MLP-ANN). Chemom. Intell. Lab. Syst. 2016, 155, 73–85. [Google Scholar] [CrossRef]
  40. Liu, M.; Zhang, D.; Shen, D.; Alzheimer’s Disease Neuroimaging Initiative. Ensemble sparse classification of Alzheimer’s disease. NeuroImage 2012, 60, 1106–1116. [Google Scholar] [CrossRef] [Green Version]
  41. Elish, M.O. Assessment of Voting Ensemble for Estimating Software Development Effort. In Proceedings of the 2013 IEEE Symposium on Computational Intelligence and Data Mining (CIDM), Singapore, 16–19 April 2013; pp. 316–321. [Google Scholar]
  42. Yang, S.; Berdine, G. The receiver operating characteristic (ROC) curve. Southwest Respir. Crit. Care Chron. 2017, 5, 34–36. [Google Scholar] [CrossRef]
  43. Rezende, E.; Ruppert, G.; Carvalho, T.; Theophilo, A.; Ramos, F.; Geus, P.D. Malicious Software Classification Using VGG16 Deep Neural Network’s Bottleneck Features. In Information Technology-New Generations; Springer: Cham, Switzerland, 2018; pp. 51–59. [Google Scholar]
  44. Berrar, D. Bayes’ Theorem and Naive Bayes Classifier. In Encyclopedia of Bioinformatics and Computational Biology: ABC of Bioinformatics; Elsevier Science Publisher: Amsterdam, The Netherlands, 2015; pp. 403–412. [Google Scholar]
  45. Xu, S.; Li, Y.; Wang, Z. Bayesian multinomial Naïve Bayes classifier to text classification. In Advanced Multimedia and Ubiquitous Engineering; Springer: Singapore, 2017; pp. 347–352. [Google Scholar]
  46. Brijain, M.; Patel, R.; Kushik, M.R.; Rana, K. A survey on decision tree algorithm for classification. Int. J. Eng. Dev. Res. 2014, 2, 1–5. [Google Scholar]
  47. Jaiswal, J.K.; Samikannu, R. Application of Random Forest Algorithm on Feature Subset Selection and Classification and Regression. In Proceedings of the 2017 World Congress on Computing and Communication Technologies (WCCCT), Tiruchirappalli, India, 2–4 February 2017; pp. 65–68. [Google Scholar]
  48. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Cheng, D. Learning k for knn classification. ACM Trans. Intell. Syst. Technol. 2017, 8, 1–19. [Google Scholar] [CrossRef] [Green Version]
  49. Hamiane, M.; Saeed, F. SVM Classification of MRI Brain Images for Computer-Assisted Diagnosis. Int. J. Electr. Comput. Eng. 2017, 7, 2555. [Google Scholar] [CrossRef] [Green Version]
  50. Zhan, Y.; Chen, K.; Wu, X.; Zhang, D.; Zhang, J.; Yao, L.; Guo, X. Identification of Conversion from Normal Elderly Cognition to Alzheimer’s Disease using Multimodal Support Vector Machine. J. Alzheimers Dis. 2015, 47, 1057–1067. [Google Scholar] [CrossRef]
  51. Suthaharan, S. Support Vector Machine. In Machine Learning Models and Algorithms for Big Data Classification; Springer: Boston, MA, USA, 2016; pp. 207–235. [Google Scholar]
  52. Lin, H.T.; Lin, C.J. A study on sigmoid kernels for SVM and the training of non-PSD kernels by SMO-type methods. Neural Comput. 2003, 3, 16. [Google Scholar]
  53. Natekin, A.; Knoll, A. Gradient boosting machines, a tutorial. Front. Neurorobotics 2013, 7, 21. [Google Scholar] [CrossRef] [Green Version]
  54. Chakravarthy, D.G.; Kannimuthu, S. Extreme Gradient Boost Classification Based Interesting User Patterns Discovery for Web Service Composition. Mob. Netw. Appl. 2019, 24, 1883–1895. [Google Scholar] [CrossRef]
  55. Jiang, J.; Trundle, P.; Ren, J. Medical image analysis with artificial neural networks. Comput. Med. Imaging Graph. 2010, 34, 617–631. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Leon, F.; Floria, S.A.; Bădică, C. Evaluating the Effect of Voting Methods on Ensemble-Based Classification. In Proceedings of the 2017 IEEE International Conference on Innovations in Intelligent Systems and Applications (INISTA), Gdynia, Poland, 3–5 July 2017; pp. 1–6. [Google Scholar]
  57. Zhou, Z.-H. Ensemble Methods: Foundations and Algorithms; Chapman and Hall/CRC: Boca Raton, FL, USA, 2019. [Google Scholar]
  58. Zhang, Y.; Zhang, H.; Cai, J.; Yang, B. A Weighted Voting Classifier Based on Differential Evolution. Abstr. Appl. Anal. 2014, 2014, 376950. [Google Scholar] [CrossRef] [Green Version]
  59. Vovk, V. The Fundamental Nature of the Log Loss Function. In Fields of Logic and Computation II; Springer: Cham, Switzerland, 2015; pp. 307–318. [Google Scholar]
  60. Sharma, R.; Kaushik, B. Handwritten Indic scripts recognition using neuro-evolutionary adaptive PSO based convolutional neural networks. Sādhanā 2022, 47, 30. [Google Scholar] [CrossRef]
  61. Madsen, H.; Pinson, P.; Kariniotakis, G.; Nielsen, H.A.; Nielsen, T.S. Standardizing the performance evaluation of short-term wind power prediction models. Wind. Eng. 2005, 29, 475–489. [Google Scholar] [CrossRef] [Green Version]
  62. Hajian-Tilaki, K. Receiver Operating Characteristic (ROC) Curve Analysis for Medical Diagnostic Test Evaluation. Casp. J. Intern. Med. 2013, 4, 627–635. [Google Scholar]
  63. Ahmed, U.; Mukhiya, S.K.; Srivastava, G.; Lamo, Y.; Lin, J.C.-W. Attention-Based Deep Entropy Active Learning Using Lexical Algorithm for Mental Health Treatment. Front. Psychol. 2021, 12, 642347. [Google Scholar] [CrossRef]
  64. Shah, N.; Srivastava, G.; Savage, D.W.; Mago, V. Assessing Canadians Health Activity and Nutritional Habits Through Social Media. Front. Public Health 2020, 7, 400. [Google Scholar] [CrossRef] [PubMed]
  65. Priya, R.M.S.; Maddikunta, P.K.R.; Parimala, M.; Koppu, S.; Gadekallu, T.R.; Chowdhary, C.L.; Alazab, M. An effective feature engineering for DNN using hybrid PCA-GWO for intrusion detection in IoMT architecture. Comput. Commun. 2020, 160, 139–149. [Google Scholar] [CrossRef]
  66. Wang, W.; Chen, Q.; Yin, Z.; Srivastava, G.; Gadekallu, T.R.; Alsolami, F.; Su, C. Blockchain and PUF-Based Lightweight Authentication Protocol for Wireless Medical Sensor Networks. IEEE Internet Things J. 2021, 9, 8883–8891. [Google Scholar] [CrossRef]
  67. Hu, K.; Wang, Y.; Chen, K.; Hou, L.; Zhang, X. Multi-scale features extraction from baseline structure MRI for MCI patient classification and AD early diagnosis. Neurocomputing 2016, 175, 132–145. [Google Scholar] [CrossRef] [Green Version]
  68. Kavitha, C.; Mani, V.; Srividhya, S.R.; Khalaf, O.I.; Andrés Tavera Romero, C. Early-Stage Alzheimer’s Disease Prediction Using Machine Learning Models. Front. Public Health 2022, 10, 853294. [Google Scholar] [CrossRef]
  69. Shahbaz, M.; Ali, S.; Guergachi, A.; Niazi, A.; Umer, A. Classification of Alzheimer’s Disease Using Machine Learning Techniques. In Proceedings of the 8th International Conference on Data Science, Technology and Applications (DATA 2019), Prague, Czech Republic, 26–28 July 2019. [Google Scholar] [CrossRef]
Figure 1. Individuals affected with dementia (in millions) [Source: http://ardsi.org (accessed on 5 May 2022)].
Figure 1. Individuals affected with dementia (in millions) [Source: http://ardsi.org (accessed on 5 May 2022)].
Diagnostics 12 03193 g001
Figure 2. Framework for Alzheimer’s disease classification and prediction using various algorithms.
Figure 2. Framework for Alzheimer’s disease classification and prediction using various algorithms.
Diagnostics 12 03193 g002
Figure 3. Brain axial MRI (T1 and T2 weighted) data for (a) Alzheimer’s Disease, (b) Mild Cognitive Impairment and (c) Control Normal.
Figure 3. Brain axial MRI (T1 and T2 weighted) data for (a) Alzheimer’s Disease, (b) Mild Cognitive Impairment and (c) Control Normal.
Diagnostics 12 03193 g003
Figure 4. The flow chart depicts the flow of the study.
Figure 4. The flow chart depicts the flow of the study.
Diagnostics 12 03193 g004
Figure 5. The basic concept of a neural-inspired model.
Figure 5. The basic concept of a neural-inspired model.
Diagnostics 12 03193 g005
Figure 6. Layers and parameters used in multilayer perceptrons neural network (MLP-NN) architecture.
Figure 6. Layers and parameters used in multilayer perceptrons neural network (MLP-NN) architecture.
Diagnostics 12 03193 g006
Figure 7. Proposed XGB + DT + SVM Ensemble model.
Figure 7. Proposed XGB + DT + SVM Ensemble model.
Diagnostics 12 03193 g007
Figure 8. Comparative performance evaluations among classical ML models and XGB + DT + SVM ensemble based on classification accuracy (a) before using Grid Search, (b) after using Grid Search.
Figure 8. Comparative performance evaluations among classical ML models and XGB + DT + SVM ensemble based on classification accuracy (a) before using Grid Search, (b) after using Grid Search.
Diagnostics 12 03193 g008
Figure 9. ROC plots for models to visualize the correct predicted and false predicted values for the classification of AD, CN and MCI.
Figure 9. ROC plots for models to visualize the correct predicted and false predicted values for the classification of AD, CN and MCI.
Diagnostics 12 03193 g009
Figure 10. A graphical representation of comparison between best performing XGB proposed XGB + DT + SVM and MLP-NN with previous studies [63,64,65].
Figure 10. A graphical representation of comparison between best performing XGB proposed XGB + DT + SVM and MLP-NN with previous studies [63,64,65].
Diagnostics 12 03193 g010
Table 1. A literature review. The acronym used is provided in the Appendix A of the paper.
Table 1. A literature review. The acronym used is provided in the Appendix A of the paper.
AuthorData
Base
Imaging ModalityTotal
Subjects
Dataset DivisionFeature
Extraction and
Classification
Techniques
Classification
Accuracy
[25]ADNIs-MRI259sMCI = 139; pMCI = 120FW
LLE
sMCI/pMCI = 80.0%
[29]ADNIMRI710AD = 200; MCI = 280; pMCI = 120;
sMCI = 160; CN = 230
MK-BoostAD/CN = 94.65%
MCI/CN = 85.79%
sMCI/pMCI = 72.08%
[31]ADNIMRI, PET726AD = 101; CI = 223; sMCI = 93; pMCI = 140; CN= 169SVMAD/CN = 91.20%
MCI/CN = 76.40%
sMCI/pMCI = 74.20%
[32]ADNIs-MRI303AD = 158; CN = 145SVM-RFE
SVM
AD/CN = 90.76%
[35]ADNIMRI, PET654AD = 154; MCI = 346; CN = 154RFLD
SVM
AD/CN = 88.30%
MCI/CN = 79.02%
[36]ADNIs-MRI,
PET
340AD = 113; MCI = 110; CN = 117AAL, ROI
mSCDDL
AD/CN = 98.5%
MCI/CN = 82.8%
[37]ADNIMRI, PET427AD = 65; pMCI = 95; sMCI = 132; CN = 135MDS, PCA
SVM
AD/CN = 96.5%
MCI/CN = 91.74%
sMCI/pMCI =88.99%
[38]ADNIMRI, PET249AD = 70; MCI = 111; CN = 68PCA
SVM
AD/CN = 92%
MCI/CN = 84%
[39]OASIS, LHMRI196AD = 98; CN = 983AF
8L-CNN
AD/CN = 97.65%
[40]ADNIMRI710AD = 200; pMCI = 120; sMCI = 160; CN = 230ROI, FS
MK-BOOST
AD/CN = 95.24%
MCI/CN = 86.35%
sMCI/pMCI = 74.25%
[41]ADNI1MRI891AD = 199; pMCI = 167; sMCI = 226; CN = 229ROI, VBM,
CLM
LDMIL
AD/CN = 95.86%
sMCI/pMCI = 77.64%
ADNI2MRI636AD = 159; pMCI = 38; sMCI = 239;
CN = 200
MIRIADMRI69AD = 23; CN = 46AD/CN = 97.16%
[42]SELFrs-MRI68AD = 34; CN = 343LHPM-ICA
DL Algo.
AD/CN = 95.59%
[28]ADNIMRI
FDG-
PET
1628AD = 342; MCI = 418; CN = 866RVB, SUVR, ROI
L-SVM, LR, RF
AD/CN = 88%
MCI/CN = 80%
sMCI/pMCI = 73%
AIBL139AD = 72; MCI = 94; pMCI = 21; sMCI = 16; CN = 442
OASIS416AD = 100; CN = 93
Table 2. Detailed Sample-wise Distribution of Dataset.
Table 2. Detailed Sample-wise Distribution of Dataset.
DatasetTotalAlzheimer’s
Disease
Mild Cognitive
Impairment
Control
Normal
Brain
Axial MRI
T1 and T2 weighted
2127Images = 612Images = 538Images = 975
Class 1 (AD)Class 2 (MCI)Class 3 (CN)
Label = 0Label = 1Label = 2
Table 3. Overview of confusion matrix/contingency table.
Table 3. Overview of confusion matrix/contingency table.
Ground Truth/Actual
Predicted Actual PositiveActual Negative
Predicted PositiveTrue positive (TP)False Negative (FN)
Predicted NegativeFalse Positive (FP)True Negative (TN)
Table 4. Hyperparameter tuning to obtain optimized parameters using GridSearch.
Table 4. Hyperparameter tuning to obtain optimized parameters using GridSearch.
ModelGrid ParametersOptimized Parameters
Decision Tree parameters = {
‘max_depth’: [2,4,6,8,10,12,14],
‘min_samples_leaf’: [5,7,9,11,13,15],
‘criterion’: [“gini”, “entropy”],
‘min_samples_split’: [1,2,3,4,5,6,7,8]}
parameters = {
‘max_depth’: [10],
‘min_samples_leaf’: [15],
‘criterion’: [‘gini’],
‘min_samples_split’: [4]}
Random Forest model_params = {
‘n_estimators’: randint(4200),
‘max_features’: truncnorm(a = 0,
B = 1, loc = 0.25, scale = 0.1),
‘min_samples_split’:uniform (0.01,0.199)}
model_params = {
‘n_estimators’: [148],
‘max_features’: [0.276163],
‘min_samples_split’: [0.0392]}
K-Nearest Neighbor grid_params = {
‘n_neighbors’: [3,5,11,13,15],
‘weights’: [‘uniform’,’distance’],
‘metric’: [‘minkowski’,’euclidean’,’manhattan’]}
grid_params = {
‘n_neighbors’: [3],
‘weights’: [‘uniform’],
‘metric’: [‘minkowski’]}
Naïve Bayes parameters = {
‘Var_smoothing’:np.logspace (0,9,num = 1000)}
parameters = {
‘Var_smoothing’: [4.328]}
Extreme Gradient
Boosting
parameters = {
‘learning_rate’: [0.01,0.1],
‘max_depth’: [3,5,6,10],
‘subsample’: [0.5,0.7],
‘colsample_bytree’: [0.5,0.7],
‘n_estimators’: [100,200,500],
‘objective’: [‘reg:squarederror’]}
parameters = {
‘learning_rate’: [0.1],
‘max_depth’: [6],
‘subsample’: [0.5],
‘colsample_bytree’: [0.5],
‘n_estimators’: [100],
‘objective’: [‘reg:squarederror’]}
Gradient Boost parameters = {
“n_estimators”: [50,150],
“max_depth”: [1,3,5],
“learning_rate”: [0.01,0.1,0.5,1],
‘max_depth’: [3,5,6,10],
‘min_child_weight’: [1,3,5],
‘subsample’: [0.5,0.7],
‘colsample_bytree’: [0.5,0.7],
‘n_estimators’: [100,200,500],}
parameters = {
“n_estimators”: [80],
“max_depth”: [3],
“learning_rate”: [0.5], ‘max_depth’: [6],
‘min_child_weight’: [5],
‘subsample’: [0.5],
‘colsample_bytree’: [0.5],
‘n_estimators’: [100]}
Multi-layer Perceptron
Neural Network
(MLP-NN)
Parameters = {‘batch_size’: [20,30],
‘nb_epoch’: [50,100,150],
‘optimizer’: [‘adam’,’rmsprop’,’SGD’]
‘units’: [512,1024]}
parameters = {‘batch_size’: [20],
‘nb_epoch’: [50],
‘optimizer’: [‘adam’],
‘units’: [512]}
Proposed
Ensemble model
XGB + DT + SVM
parameters = {
‘learning_rate’: [0.01,0.1],
‘max_depth’: [3,5,6,10],
‘min_child_weight’: [1,3,5],
‘subsample’: [0.5,0.7],
‘colsample_bytree’: [0.5,0.7],
‘n_estimators’: [100,200,500],
‘objective’: [‘reg:squarederror’],
‘min_samples_leaf’: [5,7,9,11,13,15],
‘criterion’: [“gini”, “entropy”],
‘min_samples_split’: [1,2,3,4,5,6,7,8]
‘kernel’: [‘rbf’, ‘poly’, ‘sigmoid’]}
param_grid = {
“n_estimators”: [80],
“max_depth”: [3],
“learning_rate”: [0.5], ‘max_depth’: [6],
‘min_child_weight’: [5],
‘subsample’: [0.5],
‘colsample_bytree’: [0.5],
‘n_estimators’: [100]
‘kernel’: [‘rbf’]}
Table 5. Before hyperparameter tuning (parameter optimization) using GridSearch.
Table 5. Before hyperparameter tuning (parameter optimization) using GridSearch.
ModelAccuracy (%)Sensitivity (%)Specificity (%)Log Loss
ADCNMCIADCNMCIADCNMCI
DT87.1289.5677.8495.2292.3987.5749.0698.8290.240.74054
RF80.5982.5160.8975.7653.2679.5071.2981.5894.110.71971
K-NN77.4272.0966.3083.6146.6380.1261.1696.3976.960.38076
NB72.2275.8658.7572.0192.3991.3057.7495.4664.4013.6992
GB72.2577.9877.5889.7648.9188.1972.2790.0097.510.53312
XGB81.5689.0084.6392.4992.4997.5180.0594.2798.710.36665
SVM RBF71.4588.5986.3698.6354.8999.0079.6399.1196.050.33290
SVM Poly87.8486.2181.4883.9558.1594.4079.2089.2395.400.29615
SVM Sigmoid75.4871.4983.1490.4476.0896.2786.5193.5499.260.48208
MLP-NN89.2081.5684.3095.2264.6798.7582.4996.4798.510.36665
Proposed XGB + DT + SVM87.0090.5190.5191.8077.1798.7588.2694.2799.510.27795
Table 6. After hyperparameter tuning (parameter optimization) using GridSearch.
Table 6. After hyperparameter tuning (parameter optimization) using GridSearch.
ModelAccuracy (%)Sensitivity (%)Specificity (%)Log Loss
ADCNMCIADCNMCIADCNMCI
DT89.1292.5679.8497.2637.5091.9264.7797.3098.330.58883
RF82.5987.5167.8976.7957.0681.3674.6883.1792.960.71971
K-NN 85.8488.2185.4890.4469.0297.5183.7793.5698.741.31750
NB79.2283.8678.7585.32a75.4389.4450.0099.8771.420.27213
GB79.2575.9879.5891.4655.4393.7876.4393.1197.880.39457
XGB87.5689.8986.6393.8562.5096.2779.8879.8899.480.32501
SVM Poly88.0086.2185.0092.7577.5487.0086.0081.7684.340.31123
MLP- NN79.4589.7986.3690.7877.1796.2787.0993.7699.270.42902
Proposed
XGB + DT + SVM
96.129596.1593.8578.8098.7589.6795.5999.050.24926
Table 7. Ten-fold cross-validation after parameter optimization (hyperparameter tuning).
Table 7. Ten-fold cross-validation after parameter optimization (hyperparameter tuning).
FoldsDTRFK-NNNBSVM-RBFSVM-
Poly
SVM-
Sigmoid
GBXGBMLP-NNProposed
XGB + DT + SVM
1_fold0.718120.771810.838920.838920.832210.832210.657710.818790.838920.845630.98547
2_fold0.718120.798650.906040.892610.926170.926170.677820.798670.872480.932880.97564
3_fold0.671140.738250.865770.805360.879190.879190.704690.838920.825500.872480.97554
4_fold0.677850.832210.912750.899320.879190.879190.778520.865770.879190.879190.96587
5_fold0.718120.798650.879190.892610.872480.872480.630870.812080.838920.885900.98658
6_fold0.718120.758380.852340.825500.885900.885900.657710.805360.892610.892610.97857
7_fold0.677850.791940.899320.872480.906040.906040.664420.845630.879190.919460.98143
8_fold0.675670.790540.885130.824320.878370.878370.682430.824320.878370.878370.98543
9_fold0.682430.790540.844590.817560.891890.891890.621620.844590.844590.891890.99948
10_fold0.756750.768450.864860.858100.878370.878370.648640.804050.864860.885130.96854
Mean0.701410.783940.874890.852680.874240.874240.670430.825820.861460.888350.98845
Table 8. Utilizing Friedman’s Rank test to determine each model’s mean rank position and p-value.
Table 8. Utilizing Friedman’s Rank test to determine each model’s mean rank position and p-value.
Rank1st2nd3rd4th5th6th7th8thp-Value
ModelGBRFK-NNNBDTXGBMLP-NN XGB + DT + SVM4.36 × 10−12
Average rank23.055.2514.155.96.658
Table 9. Mean differences of accuracy before and after hyperparameter tuning using t-test.
Table 9. Mean differences of accuracy before and after hyperparameter tuning using t-test.
Before Hyperparameter
Tuning
After Hyperparameter
Tuning
Mean Differencet-Values
Accuracy (%)76.3980.448−4.058 **−3.225
Note: (**) indicate the significance level at a 5% level of significance.
Table 10. Comparative analysis of presented models with previous studies.
Table 10. Comparative analysis of presented models with previous studies.
S. NoStudy ReferenceDatasetTechniqueAccuracy
1.Kun Hu et al. [67]ADNISVM84.30%
2.Kavitha et al. [68]OASISVoting classifier 85.12%
3.Shahbaz, Muhammad et al. [69]ADNIGeneralized linear model (GLM)88.24%
4.Classical ML ModelADNIXGB85.42%
5.Neural NetworkADNIMLP-NN89.64%
6.Proposed Voting Ensemble ADNIXGB + DT + SVM95.75%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Khan, Y.F.; Kaushik, B.; Chowdhary, C.L.; Srivastava, G. Ensemble Model for Diagnostic Classification of Alzheimer’s Disease Based on Brain Anatomical Magnetic Resonance Imaging. Diagnostics 2022, 12, 3193. https://doi.org/10.3390/diagnostics12123193

AMA Style

Khan YF, Kaushik B, Chowdhary CL, Srivastava G. Ensemble Model for Diagnostic Classification of Alzheimer’s Disease Based on Brain Anatomical Magnetic Resonance Imaging. Diagnostics. 2022; 12(12):3193. https://doi.org/10.3390/diagnostics12123193

Chicago/Turabian Style

Khan, Yusera Farooq, Baijnath Kaushik, Chiranji Lal Chowdhary, and Gautam Srivastava. 2022. "Ensemble Model for Diagnostic Classification of Alzheimer’s Disease Based on Brain Anatomical Magnetic Resonance Imaging" Diagnostics 12, no. 12: 3193. https://doi.org/10.3390/diagnostics12123193

APA Style

Khan, Y. F., Kaushik, B., Chowdhary, C. L., & Srivastava, G. (2022). Ensemble Model for Diagnostic Classification of Alzheimer’s Disease Based on Brain Anatomical Magnetic Resonance Imaging. Diagnostics, 12(12), 3193. https://doi.org/10.3390/diagnostics12123193

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop