You are currently on the new version of our website. Access the old version .
DiagnosticsDiagnostics
  • Article
  • Open Access

20 February 2023

Alzheimer Disease Classification through Transfer Learning Approach

,
,
and
1
Department of Computer Science, National University of Computer and Emerging Sciences, Lahore 54770, Pakistan
2
Department of Computer Science, Forman Christian College, Lahore 54600, Pakistan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue AI/ML-Based Medical Image Processing and Analysis

Abstract

Alzheimer’s disease (AD) is a slow neurological disorder that destroys the thought process, and consciousness, of a human. It directly affects the development of mental ability and neurocognitive functionality. The number of patients with Alzheimer’s disease is increasing day by day, especially in old aged people, who are above 60 years of age, and, gradually, it becomes cause of their death. In this research, we discuss the segmentation and classification of the Magnetic resonance imaging (MRI) of Alzheimer’s disease, through the concept of transfer learning and customizing of the convolutional neural network (CNN) by specifically using images that are segmented by the Gray Matter (GM) of the brain. Instead of training and computing the proposed model accuracy from the start, we used a pre-trained deep learning model as our base model, and, after that, transfer learning was applied. The accuracy of the proposed model was tested over a different number of epochs, 10, 25, and 50. The overall accuracy of the proposed model was 97.84%.

1. Introduction

Alzheimer’s disease (AD) is a slow neurological disorder that destroys the thought process, and consciousness, of a human. It directly affects the development of mental ability and neurocognitive functionality [1]. AD is more common among older people, and sometimes it becomes the primary cause of their deaths [2,3]. People who have other medical problems at later age, such as diabetes, cardio-problems, and hypertension [1], are at higher risk of developing Alzheimer’s disease. There is no complete treatment for AD disease. However, early detection of this disease can help in taking preventative action at an early stage and can improve the symptoms of AD [4]. It is a primary reason for dementia in the elderly, due to the destruction of neurons related to the human memory [4], and to questioning, and learning functions [5]. This neurological disorder starts with slow deterioration and symptoms worsen with each passing day [6]. Memory issues are generally one of the first symptoms of AD [2], although preliminary symptoms may also vary from individual to individual [3]. Changes in different factors of thinking, inclusive of locating the proper words, spatial issues, and impaired judgment, might also become the early symptoms of AD [7]. The initial stage of mild cognitive impairment (MCI) is an early signal that a human may be an AD patient [8]. The number of AD patients is increasing day by day and it is expected that one out of eighty-five humans could be suffering from Alzheimer’s disease by 2050 [9,10].
The progression of Alzheimer’s disease can be diagnosed using clinical measures, but it is a very time-consuming process and expert persons are required to detect the symptoms [11,12]. Early diagnosis is very difficult to assess by an expert person, unless symptoms become very obvious. Early detection of AD can assist in lowering the risk of neuron disorder [13,14]. Early diagnosis can ensure that the patient is aware of the need to take precautionary measures to lower the risk of advancement of the disease from MCI to AD [1]. In recent times, different machine learning and deep learning methods have been proposed for the prediction of the stages of AD through self-regulating analysis of magnetic resonance imaging (MRI) images, which provide efficient and improved diagnosis results for AD [14,15,16]. The major factors, or parameters, that researchers use are the cortical thickness of the human brain, gray matter (GM) density in the brain, ventricle expansions, and brain shrivel. Many research studies claim that a correlation exists between grey matter reduction and certain brain diseases, like Alzheimer’s disease [13]. The hippocampus is the part of the brain that is affected at the initial stage of Alzheimer’s disease. White matter (WM), gray matter (GM), and cerebrospinal liquid are the major and most primitive tissues in human brain images. Out of these three fundamental tissues of the brain, researchers have discovered that GM shrivel corresponds more with corporeal diminishment in mild cognitive impairment [17].
Different statistical and machine learning methods, such as Support vector machine (SVM) [18] are used for automated recognition of Alzheimer’s disease. Lately, deep learning techniques, such as CNN and sparse auto-encoders, have surpassed the SVM learning techniques [11]. There are some limitations to using deep learning methods. The training from scratch of deep learning models requires a lot of computation and a large dataset of annotated medical images [19,20]. The collection of such a massive amount of medical images is a bottleneck for researchers. To overcome issues of large dataset requirements and computational issues associated with complex methods, the transfer learning approach can be used for different medical imaging modalities. Transfer learning is very useful, even for cross-domain applications, such as the fact that a model trained on images of natural things could be used for medical images, like X-rays, CT scans, MRIs, and many others [21]. The idea behind transfer learning is to utilize a pre-trained model on a new, small image, dataset of a different nature [22,23]. Transfer learning is an emerging deep learning technique in which an architecture devised for one task is reused as the initial point for a second task. The important motive behind transfer learning is the saving of knowledge attained while resolving a specific dilemma, followed by applying that gained knowledge to solve different problems.
In this research, a CNN-based transfer learning technique is used for the classification of neuro–medical AD scans into the following four categories: Alzheimer’s disease (AD), late mild cognitive impairment (LMCI), mild cognitive impairment (MCI), and normal cognition (NC). The elemental motivation for using transfer learning was to forward the features extracted from natural brain images to AD images, and to investigate a new approach for the categorization of AD that could be helpful for clinicians in proper diagnosis and decision-making. In this way, the patient can be guided in taking precautionary measures to lower the risk of MCI progressing to AD. The main objective is to produce improved results, even when using a small-sized dataset. Transfer learning gives us the flexibility to drastically reduce training time by reusing pre-trained models for new data and for obtaining better results without over-fitting. The proposed model uses a deep learning architecture as a base model where the last few convolutional layers are retained and then merged with fully connected layers. By applying the transfer learning approach with retraining of the last two layers, promising results are obtained on multi-class categorizations, such as AD, LMCI, EMCI, MCI, and NC. This research is primarily based on two-dimensional Gray matter (GM) images of the human brain, that are extracted from raw Magnetic resonance images (MRIs), these being more helpful in the detection of the early stage of AD. As illustrated in Figure 1, the MRI scans of the patients are converted to 2D slices in the pre-processing step, where the GM slices are focused for the extraction and training of the proposed model.
Figure 1. MRI to 2D GM slice.
The major research contributions are mentioned below:
  • A customized convolutional neural network with transfer learning is proposed for the classification of Alzheimer’s disease.
  • A new corpus, consisting of four different types of AD, is developed. Each type consists of 1254 images.
  • Extraction of 2D GM slices, using “SPM12”, which is very familiar in medical image pre-processing.
  • Higher accuracy 97.84%, with a lower number of epochs, is achieved for the multi-class classification of AD.
The rest of the paper is organized in such a way that Section 2 describes the background of the techniques used in this research. Section 3 provides a detailed view of the related work, in which some related techniques of AD classifications are discussed. Section 4 describes the methodology and details of the experimental results. Section 5 elaborates the discussions, while Section 6 contains the conclusion of the research and future directions.

2. Background

Early diagnosis of Alzheimer’s disease can assist in lowering the peculiar disorder of the neurons. In this research, a model that takes a 2D GM slice as input to classify the stage of AD is discussed. A brief introduction for each of the basic concepts used in this research is provided below.

2.1. Convolutional Neural Network

Among many algorithms that enable a machine or a computer to train, perceive and learn for different medical classification problems, one is the convolutional neural network (CNN) [11,15,16]. CNN has revolutionized machine learning and artificial intelligence domains to make them akin to human brains. One of the important features of CNN is its characteristics of learning the features of images with training. In the human brain, the neurons are attached to each other in a specific pattern. CNN models are designed in the same manner. CNN models are multi-layered structures that perform in a group. The design of CNN comprises the following four kinds of layers [24]:
  • Convolution
  • Pooling
  • Fully Connected
  • Softmax

2.2. Transfer Learning

Transfer learning is an emerging deep learning technique in which an architecture trained for one task is reused as the initial point for a second task. The basic motivation behind this concept is to reduce training time, and to overcome issues associated with the need for large training sets. Pre-trained models are used for new problems without compromising the performance results.

2.3. Gray Metter

Gray matter and white matter are the two major tissues of which the central nervous system is composed [2,3]. Gray matter is the main part of the brain that is used to process information sent in the form of signals by the sensory organs of the human body.

2.4. DenseNet

DenseNet is an efficient variant of a convolutional neural network, in which each layer has extra inputs from all previous layers and forwards its very own characteristic maps to all the next layers. Due to concatenation, each layer in the DenseNet obtains complete and comprehensive knowledge from all previous layers. In this architecture, each layer obtains extra inputs from all preceding layers and passes on its feature-maps to all subsequent layers, as illustrated in Figure 2 [25]. Instead of summation, as applies to ResNet, layers are combined using concatenation. Due to this layer-wise propagation of knowledge, DenseNet can be thinner and have better computational performance [26].
Figure 2. One Dense Block in DenseNet [26].

4. Methodology

In this research, the segmentation of Alzheimer’s MRI images and their multi-class classification through transfer learning is proposed. For Alzheimer’s detection through MRI, Convolution Neural Network (CNN) was customized by using the specific GM-segmented images of the brain. Instead of training the model from scratch, a pre-trained deep learning model, Densenet-169, was used as a base model. Gradually, transfer learning was applied to this base model for Alzheimer’s detection. An overview of the proposed methodology is illustrated in Figure 3 which shows how MRI images were pre-processed and GM slices extracted. These GM slices were used for training the pre-trained deep learner and, finally, multi-class classification was performed.
Figure 3. Proposed model workflow.

4.1. DataSet

The dataset [34], used in this research was acquired from the publicly available Alzheimer’s Disease Neuroimaging Initiative (ADNI) database [37]. In 2004, this dataset was collected to analyze multiple sensitive techniques on various scanned images, like PET scan, sMRI, MRI and fMRI, etc. To evaluate the initial stages of Alzheimer’s disease [5] this dataset is considered a benchmark dataset. It contains images of four categories i.e., Alzheimer’s Disease (AD), Non-Cognitive (NC), Late Mild Cognitive Impairment (LMCI), and Mild Cognitive Impairment (MCI). Each category has 1254 images in total.

4.2. Data Preprocessing

In preprocessing, a series of steps were applied to the T1 weighted MRI images extracted from the publicly available database ADNI. For pre-processing, Statistical Parametric Mapping 12 (SPM12) was used which is quite often used in practice for medical and natural images. All the MRI images which were in Neuroimaging Informatics Technology Initiative (NIfTI) format were pre-processed to identify and extract the Gray Matter (GM) slices. The proposed model was trained on GM to detect and examine the initial changes in AD patients. The major steps of pre-processing encompassed segmentation, skull stripping, spatial normalization, and re-scaling. All the major steps, performed in the pre-processing of MRI images, are illustrated in Figure 4.
Figure 4. Major Steps of Pre-processing on MRIs.

4.2.1. Skull Stripping

For the morphometric analysis and study of brain MRIs, a preliminary process of skull stripping was performed, in which the tissues of the brain, i.e., cerebellum and cortex, were segmented from the surrounding zone consisting of non-brain and skull area. Among the multiple automated skull stripping approaches [38], the most commonly used technique, histogram thresholding, was followed by the employment of certain morphological processes [39]. The morphological procedures included erosion, and were based on anisotropic filters. The Snake method of contouring was also applied, to remove the eyes. The gradual process of skull stripping involved background and noise removal, followed by contour identification of the brain and, finally, refinement of the brain contour.
Head segmentation required noise removal, which was performed with the help of a threshold level, considering that the maximum noise was generated by Rayleigh probability distribution. The noise was removed by applying Equation (1):
R n o i s e ( f ) = f σ 2 e x p ( f 2 2 σ 2 )
where f is the intensity of noise and σ is standard deviation of Rayleigh noise.
After that, the best-fit Rayleigh curve was subtracted from the histogram volume, as given in Equation (2):
G ( f ) = h ( f ) r ( f )
where h ( f ) is histogram volume and r ( f ) is the best-fit curve. The value for r ( f ) was determined by minimizing error as given in Equation (3):
ε τ = f = 0 τ 1 g ( f ) + f = τ r ( f )
The said procedure produced speckle, which was outside the head region. This error was removed by using morphological operations in a 5 × 5 kernel.
The initial brain mask was generated in three steps. The first one was the smoothing of the brain and attenuation of non-brain regions for which anisotropic diffusion, given in Equation (4), was used in a nonlinear way:
t M ( x , t ) = . ( C ( x , t ) M ( x , t ) )
where M is MRI, x represents its coordinates, t is the iteration step and C ( x , t ) is the function for diffusion, given in Equation (5).
Next, the MRI volume was defused with the help of automatic thresholding, and a diffusion function, given in Equation (5), was used to produce a binary mask:
C ( x , t ) = e x p ( ( | M ( x , t ) | 2 f ) 2 )
where f was constant for diffusion, which was set as 128.
Finally, non-brain regions were removed by using spatial and morphological information obtained from the head mask. All the holes within the mask region were filled by applying binary erosion with kernel size 10 × 10. As a result, the kernel separated the eyes from the brain. After erosion, all the regions where the centroid was outside the bounding box were eliminated, which was decided after applying heuristics. After erosion, binary dilation was applied with the same size as the kernel that was used for erosion. This helped in recovering the darkest pixels which were eliminated from the edge of the brain, due to threshold.
Two sample images before and after skull stripping are given in Figure 5 for axial and sagittal planes.
Figure 5. MRIs of Axial and Sagittal planes before and after skull striping.

4.2.2. Segmentation

Segmentation is one of the major steps of pre-processing. For the diagnosis of certain brain disorders, segmenting brain tissues in Gray Matter (GM), White Matter (WM) and Cerebrospinal Fluid (CSF) makes the diagnosis process efficient.
The human brain MRI needs to be segmented into three main parts i.e., WM, GM, and CSF [40]. MRIs are mainly generated with the help of T1 and T2 weighted scan sequences. The main differences between the two are in Repetition Time (TR) and Time to Echo (TE), which is shorter for T1 and longer for T2. The contrast and brightness of MRIs are determined by T1 and T2, which affect the intensity levels of GM, WM, and SCF. For segmenting brain MRIs in GM, WM and SCF intensity probability distribution (IPD) was used, along with bias regularization, which was set to an extremely small regularization value i.e., 0.0001. Cutoff of 60 mm was used as Full Width at Half Maximum (FWHM), which was considered the width of a line structure at half of its highest amplitude. Figure 6 illustrates the results of segmenting an original brain MRI into GM, WM and SCF.
Figure 6. (a) Original (b) Grey Matter (GM (c) White Matter (WM) (d) Cerebrospinal Fluid.

4.2.3. Normalization

Spatial normalization is a process of mapping images containing gray matter to the same reference space. In this research, Voxel-based morphometry [41] was employed for spatial normalization. In Neuroanatomy, Voxel-based morphometry measures the dissimilarity between local tissue clusters of the brain and several other brain images. The differences are calculated via a voxel-wise comparison between the two brain images and the images are mapped to a reference brain space. For this purpose, the Montreal Neurological Institute (MNI) space was used to identify the boundaries around the brain.

4.2.4. Rescaling

After the segmentation of raw MRI images, 256 × 240 shaped data samples were extracted. All these images were then re-scaled to 224 × 224 , that were, finally,. used to train, test, and evaluate the proposed model. The re-scaling adjusted numbers for comparing values even if they were out of scope. Furthermore, it reduced the number of trainable parameters of deep learners and they trained faster.

4.2.5. Smoothing

After rescaling, MRIs were smoothed by applying a 3D Gaussian kernel with voxel size (2 2 2). Each voxel was updated by the weighted mean value, attained from the adjacent voxels. The weight was decided based on the Gaussian shape that surrounded the voxel. The variance or width of the Gaussian curve determined how much smoothing was applied. With higher variance, more neighboring values were considered in calculating the mean value, which was expressed as FWHM. Figure 7 illustrates the Gaussian Kernel.
Figure 7. Gaussian Smoothing Kernel.

4.2.6. Augmentation

The last step of data pre-processing was augmentation which was adopted to avoid over-fitting. For generating augmented MRIs, five augmentation techniques, i.e., rescaling, rotation, zooming, horizontal flip, and vertical flip, were applied. In re-scaling, images were re-scaled to a range between 0 and 1 by multiplying each cell with 1 255 . The MRIs were rotated at 30 in an anti-clockwise manner. All the MRIs were also zoomed in with a ratio of 0.2. Finally, vertical and horizontal flips were applied and, in this way, five images in total were generated from one pre-processed MRI. Figure 8 illustrates the results of augmentation.
Figure 8. Effects of augmentation on MRI.

4.3. Architecture

After pre-processing, the pre-possessed images were provided to the proposed Convolutional Neural Network, shown in Figure 9. The proposed CNN model consisted of a 7 × 7 convolution layers, 3 × 3 max-pooling layers, four dense blocks with a different number of convolution layers (6, 12, 32, 32) in respective blocks, three transitional layers, and a fully connected classification layer. Three major components, i.e., the convolution layer, dense blocks, and transition layers, were used to extract the features from the pre-processed MRIs. The convolution layer with a kernel matrix of size 7 × 7 and a stride value set as 2, was used for the extraction of macro-features [42]. These extracted macro-level features were then forwarded to dense blocks, where more features were extracted. Each dense block contained 1 × 1 convolutions, which were used to reduce the number of input feature maps, followed by 3 × 3 convolution layers. In each dense block, the extracted features were passed to all the next layers. Between the dense blocks, there were transition layers, along with the activation functions, like batch normalization and ReLu, etc. The transition layer contained 1 × 1 convolution and a 2 × 2 max pooling layer with a stride value set as 2. Once the transition layer reduced the dimensions of the features, they were forwarded to the fully connected classification layer where Softmax was used for the classification. Initially, all the initial blocks of layers were kept frozen and the last two blocks were retrained.
Figure 9. Proposed Model Architecture.
The details of all hyperparameters are listed in Table 1.
Table 1. Hyperparameters of the proposed model.
The concept of dense connectivity used in this research is represented in Equation (6):
c l = H l [ ( c 0 , c 1 , c 2 , c l 1 ) ]
where H l is the non-linear transformation of L t h layer and [ c 0 , c 1 , c 2 , c 3 . . . . ] are the feature maps forwarded by all the previous layers to the L t h layer.
The growth rate at the L t h layer, which revealed how much information was added by the previous layers, was calculated via Equation (7):
g l = [ g 0 + g × ( l 1 ) ]
where the hyper parameter g is the growth rate.
Convolutional layers were used to apply the filter on the input images to extract the feature maps. Once the feature maps were extracted by applying different filters, they were passed through the ReLU activation function. To obtain the non-linear input to the unit M i j l , all the previous layers’ benefaction were summed up, as described in Equation (8):
M i j l = a = 0 m 1 b = 0 m 1 μ a b Y ( i + a ) ( j + b ) l 1
To reduce the dimensionality of the extracted feature maps, the max pooling layer was used. Considering the dimensions of feature maps as [ m h , m w , m c ] , representing height, width, and channels of the feature maps, the dimensions were reduced by applying Equation (9) and resultant feature maps were achieved.
m a x p = ( m h f + 1 ) s × ( m w f + 1 ) s × m c
where f represents the size of the filter and s represents the strides.
In the dense layers, a neuron received the input from all the neurons of its preceding layer to perform the matrix multiplication. The standard equation for the matrix multiplication in the dense layer is shown in Equation (10):
N . λ = n 11 n 12 . . . . . n 1 y p 1 n 21 n 22 . . . . . n 2 n p 2 n x 1 n x 2 . . . . . n x y p y
where the N represents matrix dimensions of x × y and another matrix P with dimensions 1 × y and λ is the matrix of the trained parameters of the previous layer.
The λ , representing the matrix of the trained parameters of the previous layer, was updated by using the back-propagation in the training phase. The back-propagation was used to adjust the weights w k associated with a layer k over the learning rate η , as shown in Equation (11):
w k = w k η × d w k
where dw is the partial derivatives of the loss function.
The partial derivative dw of the loss function of w was acquired by employing Equation (12):
d w k = L w k = 1 n d Z k A [ k 1 ] T
where Z k is the activation function and A k is the non-linear activation function at layer k [43].
Figure 10 illustrates all the trainable and untrainable parameters of the proposed model, DenseNet, along with the output shape at each layer.
Figure 10. Output size and model parameters at each layer of the proposed DenseNet.

4.4. Experimental Setup and Training

For the performance evaluation for Alzheimer’s classification, four transfer learning-based models were trained with the same architecture, as illustrated in Figure 9. The first model was trained without using the feature of transfer learning and was provided with the images without augmentation. The second one was again trained without the feature of transfer learning but was provided with augmented MRIs. The next two models were used with the feature of transfer learning and were trained on only original MRIs and augmented MRIs, respectively.
Gray matter (GM) images, as shown in Figure 11, were provided to the models for training. Three views of these MRIs, i.e., axial, coronal, and sagittal, were considered for training the models and, in total, 1254 images from each category, i.e., AD, NC, MCI, and LMCI, were used. In two steps, the data were divided into training, validation, and testing sets. In the first step, 20% of the data were split and separated for testing. The remaining 80% of the data were further divided into training and validation sets, as explained in Figure 12.
Figure 11. Gray matter images for proposed model.
Figure 12. Framework of the proposed methodology.
During the training of the model, the number of epochs was gradually increased to verify the training accuracy and validation loss. The gradual increase in training accuracy and reduction in training loss with a different number of epochs is illustrated in Table 2.
Table 2. Training accuracy and validation loss for transfer learning-based model, trained on augmented MRIs.

5. Results

The four deep learning-based models were tested over the 20% data from the ADNI dataset. To check the performance of the proposed models, multiple measures, i.e., sensitivity, Equation (13), specificity, Equation (14), and accuracy, Equation (15), were used. All of these measures were elaborated in terms of True Positive (TP), False Positive (FP), False Negative (FN), and False Positive (FP) in their respective equations.
Sensitivity = TP TP + FN
Specificity = TN TN + FP
Accuracy = TN + TP TP + TN + FP + FN
The class-wise outcomes of these measures are depicted in Table 3, while the confusion matrix of the four classes for the best achieved accuracy is given in Table 4 and Table 5. We can see that the proposed model performed well in the classification of all categories, especially in the case of AD vs NC, MCI, and LMCI, with higher accuracy of 99.68%, sensitivity of 99.65%, and specificity of 99.69%.
Table 3. Confusion Matrix.
Table 4. Class-wise performance comparison of proposed model.
Table 5. Accuracy achieved by the four models with and without Augmentation and Transfer Learning.

6. Discussions

Numerous studies have been conducted recently on early diagnosis of Alzheimer’s disease by utilizing deep learning techniques, especially CNN. Most of the diagnosing systems are developed by training deep learning models from scratch and with a huge dataset that is trained over a large number of epochs. This study proposed a transfer learning assisted fine-tuning approach for the detection of four stages of Alzheimer’s disease. The designed model was evaluated by using prepossessed GM sliced images with a different number of epochs. The performance of the proposed was evaluated based on three measures, which were sensitivity, accuracy, and specificity. All of these three measures play a huge role in the correct classification of healthy versus ill persons. Gray matter (GM) is the main fundamental part of the human brain which processes information sent in the form of signals by different sensory organs of the body. GM is more helpful in the early diagnosis of AD. This research focused on the segregation of AD patients from healthy people. The proposed model was gradually trained and tested with a different number of epochs. Initially, the model was trained and tested over 10 epochs, with 93.11% accuracy. In the next phase, the number of epochs increased from 10 to 25, and then in the next phase the epochs increased from 25 to 50, with an overall accuracy of 97.84%, which elaborated the effectiveness of the proposed study (Figure 13).
Figure 13. Average Class Accuracy of the four Models.
Table 6 depicts the comparison with the base article [32] and the other studies that have contributed to the early diagnosis of Alzheimer’s disease. However, the proposed model performs outstandingly with 97.84% accuracy in the multi-class classification of AD with a very less number of epochs.
Table 6. Comparison of the proposed model with existing models.
Early detection of Alzheimer’s disease, using MRIs, can be challenging, due to the presence of different artifacts. like noisy background, low contrast and partial volume [45]. To overcome these issues, more recent technologies, like Functional magnetic resonance imaging (fMRI), can be used in the future for the early detection of AD.

7. Conclusions

Alzheimer’s disease is a slow neurological disorder that destroys the thought processes, and consciousness, of a human and, mostly, the aptness to perform straightforward tasks. Many deep learning models have been proposed to detect and classify the stages of AD. Gary matter is the main fundamental part of the human brain that is mainly affected by Alzheimer’s disease and it plays a major role in the gradual destruction of neurons. In this research, a deep learning model, Dense-Net, with transfer learning, was applied to the MRI dataset. The main objective of this research was to classify the stages of Alzheimer’s disease, based on the extracted gray matter (GM), the main fundamental part of the human brain. MRI scans were segmented into 3 parts, GM, WM, and CSF, using the SPM12 for pre-processing. The 2D GM slices were used as input for the training and testing of the model. A pre-trained DenseNet model, with retraining of the last two blocks, was applied over the segmented GM slices. The proposed model provided promising results, with 97.84% accuracy in the multi-class classification of AD.
There were certain pros and cons of the proposed technique. DenseNet is well-reputed, due to its applications in medical imaging [46]. The analysis of medical images by DenseNet is remarkable, as it acquires a comprehensive and complete knowledge of image details from all previous layers. In addition to that, the propagation which takes place layer-wise can be made thinner to make it computationally economical without compromising accuracy, due to the fact that the final prediction depends on features obtained from all layers. As the proposed methodology was based on transfer learning, so the significant limitation of this approach was negative transfer, due to which, for initial training, the target and initial problems were supposed to be adequately homogeneous.
Still, considering the outstanding performance of the proposed model with respect to other CNN models, in the future, this model could be applied for the detection and classification of lungs, and other diseases.

Author Contributions

Conceptualization, N.R. and A.N.; methodology, N.R. and A.N.; software, N.R.; validation, M.T., A.N. and K.Z.; formal analysis, M.T. and K.Z.; investigation, N.R., A.N., M.T. and K.Z.; resources, A.N.; data curation, N.R.; writing-original draft preparation, N.R.; writing-review and editing, A.N. and M.T; visualization, N.R.; supervision, A.N. and K.Z.; project administration, A.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The dataset [34], used in this research, is acquired from the publicly available Alzheimer’s Disease Neuroimaging Initiative (ADNI) database [37].

Acknowledgments

We would like to acknowledge all data sources who made all datasets publicly available [34]. In addition to availability of the dataset, all codes for deep CNN architectures are publicly available.

Conflicts of Interest

There is no conflict of interest among the authors. This paper will be of interest to the readership of this Journal. As the corresponding author and first author of the research work, I hereby confirm that the manuscript is entirely original, has not been copyrighted, published, submitted, or accepted for publication elsewhere.

References

  1. Srivastava, S.; Ahmad, R.; Khare, S.K. Alzheimer’s disease and its treatment by different approaches: A review. Eur. J. Med. Chem. 2021, 216, 113320. [Google Scholar] [CrossRef] [PubMed]
  2. Alzheimer’s Association. 2015 Alzheimer’s disease facts and figures. Alzheimer’s Dement. 2015, 11, 332–384. [Google Scholar] [CrossRef] [PubMed]
  3. Alzheimer’s Association. 2016 Alzheimer’s disease facts and figures. Alzheimer’s Dement. 2016, 12, 459–509. [Google Scholar] [CrossRef] [PubMed]
  4. Breijyeh, Z.; Karaman, R. Comprehensive review on Alzheimer’s disease: Causes and treatment. Molecules 2020, 25, 5789. [Google Scholar] [CrossRef]
  5. Jack, C.R., Jr.; Bernstein, M.A.; Fox, N.C.; Thompson, P.; Alexander, G.; Harvey, D.; Borowski, B.; Britson, P.J.; Whitwell, J.L.; Ward, C.; et al. The Alzheimer’s disease neuroimaging initiative (ADNI): MRI methods. J. Magn. Reson. Imaging Off. J. Int. Soc. Magn. Reson. Med. 2008, 27, 685–691. [Google Scholar] [CrossRef]
  6. Khan, Y.F.; Kaushik, B. Computer vision technique for neuro-image analysis in neurodegenerative diseases: A survey. In Proceedings of the IEEE 2020 International Conference on Emerging Smart Computing and Informatics (ESCI), Pune, India, 9–11 March 2020; pp. 346–350. [Google Scholar]
  7. Khan, Y.F.; Kaushik, B. Neuro-image classification for the prediction of alzheimer’s disease using machine learning techniques. In Proceedings of the International Conference on Machine Intelligence and Data Science Applications; Springer: Berlin/Heidelberg, Germany, 2021; pp. 483–493. [Google Scholar]
  8. Ungar, L.; Altmann, A.; Greicius, M.D. Apolipoprotein E, gender, and Alzheimer’s disease: An overlooked, but potent and promising interaction. Brain Imaging Behav. 2014, 8, 262–273. [Google Scholar] [CrossRef]
  9. Brookmeyer, R.; Johnson, E.; Ziegler-Graham, K.; Arrighi, H.M. Forecasting the global burden of Alzheimer’s disease. Alzheimer’s Dement. 2007, 3, 186–191. [Google Scholar] [CrossRef]
  10. Naseer, A.; Yasir, T.; Azhar, A.; Shakeel, T.; Zafar, K. Computer-aided brain tumor diagnosis: Performance evaluation of deep learner CNN using augmented brain MRI. Int. J. Biomed. Imaging 2021, 2021, 5513500. [Google Scholar] [CrossRef]
  11. Kundaram, S.S.; Pathak, K.C. Deep learning-based Alzheimer disease detection. In Proceedings of the 4th International Conference on Microelectronics, Computing And Communication Systems; Springer: Berlin/Heidelberg, Germany, 2021; pp. 587–597. [Google Scholar]
  12. Naseer, A.; Tamoor, M.; Azhar, A. Computer-aided COVID-19 diagnosis and a comparison of deep learners using augmented CXRs. J. X-ray Sci. Technol. 2021, 30, 1–21. [Google Scholar] [CrossRef]
  13. Dadar, M.; Manera, A.L.; Ducharme, S.; Collins, D.L. White matter hyperintensities are associated with grey matter atrophy and cognitive decline in Alzheimer’s disease and frontotemporal dementia. Neurobiol. Aging 2022, 111, 54–63. [Google Scholar] [CrossRef]
  14. Aderghal, K.; Khvostikov, A.; Krylov, A.; Benois-Pineau, J.; Afdel, K.; Catheline, G. Classification of Alzheimer disease on imaging modalities with deep CNNs using cross-modal transfer learning. In Proceedings of the 2018 IEEE 31st International Symposium on Computer-Based Medical Systems (CBMS), Karlstad, Sweden, 18–21 June 2018; pp. 345–350. [Google Scholar]
  15. Ebrahimi, A.; Luo, S.; Initiative, A.D.N. Disease Neuroimaging Initiative for the Alzheimer’s. Convolutional neural networks for Alzheimer’s disease detection on MRI images. J. Med. Imaging 2021, 8, 024503. [Google Scholar] [CrossRef] [PubMed]
  16. Farooq, A.; Anwar, S.; Awais, M.; Rehman, S. A deep CNN based multi-class classification of Alzheimer’s disease using MRI. In Proceedings of the 2017 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 18–20 October 2017; pp. 1–6. [Google Scholar]
  17. Khedher, L.; Ramírez, J.; Górriz, J.M.; Brahim, A.; Segovia, F.; Alzheimer’s Disease Neuroimaging Initiative. Early diagnosis of Alzheime’s disease based on partial least squares, principal component analysis and support vector machine using segmented MRI images. Neurocomputing 2015, 151, 139–150. [Google Scholar] [CrossRef]
  18. Plant, C.; Teipel, S.J.; Oswald, A.; Böhm, C.; Meindl, T.; Mourao-Miranda, J.; Bokde, A.W.; Hampel, H.; Ewers, M. Automated detection of brain atrophy patterns based on MRI for the prediction of Alzheimer’s disease. Neuroimage 2010, 50, 162–174. [Google Scholar] [CrossRef] [PubMed]
  19. Tamoor, M.; Younas, I. Automatic segmentation of medical images using a novel Harris Hawk optimization method and an active contour model. J. X-ray Sci. Technol. 2021, 29, 721–739. [Google Scholar] [CrossRef] [PubMed]
  20. Malik, Y.S.; Tamoor, M.; Naseer, A.; Wali, A.; Khan, A. Applying an adaptive Otsu-based initialization algorithm to optimize active contour models for skin lesion segmentation. J. X-ray Sci. Technol. 2022, 30, 1–16. [Google Scholar] [CrossRef] [PubMed]
  21. Gao, M.; Bagci, U.; Lu, L.; Wu, A.; Buty, M.; Shin, H.C.; Roth, H.; Papadakis, G.Z.; Depeursinge, A.; Summers, R.M.; et al. Holistic classification of CT attenuation patterns for interstitial lung diseases via deep convolutional neural networks. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2018, 6, 1–6. [Google Scholar] [CrossRef]
  22. Liu, X.; Wang, C.; Bai, J.; Liao, G. Fine-tuning pre-trained convolutional neural networks for gastric precancerous disease classification on magnification narrow-band imaging images. Neurocomputing 2020, 392, 253–267. [Google Scholar] [CrossRef]
  23. Naseer, A.; Zafar, K. Meta-feature based few-shot Siamese learning for Urdu optical character recognition. In Computational Intelligence; Wiley: Hoboken, NJ, USA, 2022. [Google Scholar]
  24. Sezer, A.; Sezer, H.B. Convolutional neural network based diagnosis of bone pathologies of proximal humerus. Neurocomputing 2020, 392, 124–131. [Google Scholar] [CrossRef]
  25. Kim, H. Organize Everything I Know. 2018. Available online: https://oi.readthedocs.io/en/latest/computer{_}vision/cnn/densenet.html (accessed on 25 November 2022).
  26. Tsang, S.H. Review: DenseNet—Dense Convolutional Network (Image Classification). 2018. Available online: https://towardsdatascience.com/review-densenet-image-classification-b6631a8ef803 (accessed on 25 November 2022).
  27. Hon, M.; Khan, N.M. Towards Alzheimer’s disease classification through transfer learning. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas, MO, USA, 13–16 November 2017; pp. 1166–1169. [Google Scholar]
  28. Sarraf, S.; Tofighi, G. Deep learning-based pipeline to recognize Alzheimer’s disease using fMRI data. In Proceedings of the IEEE 2016 Future Technologies Conference (FTC), San Francisco, CA, USA, 6–7 December 2016; pp. 816–820. [Google Scholar]
  29. Liu, S.; Liu, S.; Cai, W.; Che, H.; Pujol, S.; Kikinis, R.; Feng, D.; Fulham, M.J. Multimodal neuroimaging feature learning for multiclass diagnosis of Alzheimer’s disease. IEEE Trans. Biomed. Eng. 2014, 62, 1132–1140. [Google Scholar] [CrossRef]
  30. Tufail, A.B.; Ma, Y.K.; Zhang, Q.N. Binary classification of Alzheimer’s disease using sMRI imaging modality and deep learning. J. Digit. Imaging 2020, 33, 1073–1090. [Google Scholar] [CrossRef]
  31. Liu, S.; Liu, S.; Cai, W.; Pujol, S.; Kikinis, R.; Feng, D. Early diagnosis of Alzheimer’s disease with deep learning. In Proceedings of the 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI), Beijing, China, 29 April–2 May 2014; pp. 1015–1018. [Google Scholar]
  32. Maqsood, M.; Nazir, F.; Khan, U.; Aadil, F.; Jamal, H.; Mehmood, I.; Song, O.y. Transfer learning assisted classification and detection of Alzheimer’s disease stages using 3D MRI scans. Sensors 2019, 19, 2645. [Google Scholar] [CrossRef] [PubMed]
  33. Mehmood, A.; Maqsood, M.; Bashir, M.; Shuyuan, Y. A deep Siamese convolution neural network for multi-class classification of Alzheimer disease. Brain Sci. 2020, 10, 84. [Google Scholar] [CrossRef] [PubMed]
  34. Mehmood, A.; Yang, S.; Feng, Z.; Wang, M.; Ahmad, A.S.; Khan, R.; Maqsood, M.; Yaqub, M. A transfer learning approach for early diagnosis of Alzheimer’s disease on MRI images. Neuroscience 2021, 460, 43–52. [Google Scholar] [CrossRef] [PubMed]
  35. Qiu, S.; Joshi, P.S.; Miller, M.I.; Xue, C.; Zhou, X.; Karjadi, C.; Chang, G.H.; Joshi, A.S.; Dwyer, B.; Zhu, S.; et al. Development and validation of an interpretable deep learning framework for Alzheimer’s disease classification. Brain 2020, 143, 1920–1933. [Google Scholar] [CrossRef]
  36. Ramzan, F.; Khan, M.U.G.; Rehmat, A.; Iqbal, S.; Saba, T.; Rehman, A.; Mehmood, Z. A deep learning approach for automated diagnosis and multi-class classification of Alzheimer’s disease stages using resting-state fMRI and residual neural networks. J. Med. Syst. 2020, 44, 1–16. [Google Scholar] [CrossRef]
  37. Weiner, D.M.W. Alzheimer Disease Neuroimaging Initiative. 2004. Available online: https://adni.loni.usc.edu/ (accessed on 25 November 2022).
  38. Kalavathi, P.; Prasath, V. Methods on skull stripping of MRI head scan images—A review. J. Digit. Imaging 2016, 29, 365–379. [Google Scholar] [CrossRef]
  39. Atkins, M.S.; Mackiewich, B.T. Fully automatic segmentation of the brain in MRI. IEEE Trans. Med. Imaging 1998, 17, 98–107. [Google Scholar] [CrossRef]
  40. Young, J.; Modat, M.; Cardoso, M.J.; Mendelson, A.; Cash, D.; Ourselin, S. Alzheimer’s Disease Neuroimaging Initiative. Accurate multimodal probabilistic prediction of conversion to Alzheimer’s disease in patients with mild cognitive impairment. NeuroImage Clin. 2013, 2, 735–745. [Google Scholar] [CrossRef]
  41. Ashburner, J.; Friston, K. Voxel-Based Morphometry—The Methods. Neuroimage 2000, 11 Pt 1, 805–821. [Google Scholar] [CrossRef]
  42. Bose, S.R.; Pearline, A.S.; Bose, R. Real-Time Plant Species Recognition Using Non-Averaged Densenet-169 Deep Learning Paradigm; SSRN: Rochester, NY, USA, 2022. [Google Scholar]
  43. Vulli, A.; Srinivasu, P.N.; Sashank, M.S.K.; Shafi, J.; Choi, J.; Ijaz, M.F. Fine-Tuned DenseNet-169 for Breast Cancer Metastasis Prediction Using FastAI and 1-Cycle Policy. Sensors 2022, 22, 2988. [Google Scholar] [CrossRef]
  44. Shanmugam, J.V.; Duraisamy, B.; Simon, B.C.; Bhaskaran, P. Alzheimer’s disease classification using pre-trained deep networks. Biomed. Signal Process. Control 2022, 71, 103217. [Google Scholar] [CrossRef]
  45. Chamarajan, G.; Charishma, Y.; Ghomi, S.M.T.; Marchiori, E. Alzheimer’s Disease: A Survey. Int. J. Artif. Intell. 2021, 8, 33–39. [Google Scholar]
  46. Zhou, T.; Ye, X.; Lu, H.; Zheng, X.; Qiu, S.; Liu, Y. Dense Convolutional Network and Its Application in Medical Image Analysis. BioMed Res. Int. 2022, 2022, 2384830. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.