Next Article in Journal
Association of Gulf War Illness-Related Symptoms with Military Exposures among 1990–1991 Gulf War Veterans Evaluated at the War-Related Illness and Injury Study Center (WRIISC)
Previous Article in Journal
Cognitive Test Scores and Progressive Cognitive Decline in the Aberdeen 1921 and 1936 Birth Cohorts
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Classification of Alzheimer’s Disease Based on External-Attention Mechanism and Fully Convolutional Network

1
School of Information Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
2
Department of Clinical Engineering, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou 310019, China
3
Department of Clinical Science, Intervention and Technology, Karolinska Institutet, 171 77 Stockholm, Sweden
4
School of Applied Sciences, Macao Polytechnic Institute, Macao SAR 999078, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(3), 319; https://doi.org/10.3390/brainsci12030319
Submission received: 12 January 2022 / Revised: 11 February 2022 / Accepted: 16 February 2022 / Published: 26 February 2022
(This article belongs to the Topic Age-Related Neurodegenerative Diseases and Stroke)

Abstract

:
Automatic and accurate classification of Alzheimer’s disease is a challenging and promising task. Fully Convolutional Network (FCN) can classify images at the pixel level. Adding an attention mechanism to the Fully Convolutional Network can effectively improve the classification performance of the model. However, the self-attention mechanism ignores the potential correlation between different samples. Aiming at this problem, we propose a new method for image classification of Alzheimer’s disease based on the external-attention mechanism. The external-attention module is added after the fourth convolutional block of the fully convolutional network model. At the same time, the double normalization method of Softmax and L1 norm is introduced to obtain a better classification performance and richer feature information of the disease probability map. The activation function Softmax can increase the degree of fitting of the neural network to the training set, which transforms linearity into nonlinearity, thereby increasing the flexibility of the neural network. The L1 norm can avoid the attention map being affected by especially large (especially small) eigenvalues. The experiments in this paper use 550 three-dimensional MRI images and use five-fold cross-validation. The experimental results show that the proposed image classification method for Alzheimer’s disease, combining the external-attention mechanism with double normalization, can effectively improve the classification performance of the model. With this method, the accuracy of the MLP-A model is 92.36%, the accuracy of the MLP-B model is 98.55%, and the accuracy of the fusion model MLP-C is 98.73%. The classification performance of the model is higher than similar models without adding any attention mechanism, and it is better than other comparison methods.

1. Introduction

Alzheimer’s disease (AD) is a progressively developing degenerative disease of the brain and nervous system. With the global escalation of the aging process, the incidence of Alzheimer’s disease is increasing every year. Elderly people with Alzheimer’s disease will experience a series of brain damages such as gradual memory loss, inconvenience of movement, decline in language expression and cognitive difficulties as the disease continues to worsen [1]. A large number of clinical studies have shown that drug intervention and care for early AD patients can delay the development of the disease and stabilize the patient’s condition. Therefore, the early and accurate judgment of patients with suspected AD has important practical significance.
At present, researchers use machine learning and deep learning to replace traditional methods for the auxiliary diagnosis of Alzheimer’s disease [2,3,4]. The methods of traditional machine learning for the classification of Alzheimer’s disease generally extract features from collected medical image data manually or semi-manually, and then send them to traditional classifiers for classification [5,6]. The algorithms of classification based on traditional methods mainly include two stages of feature extraction and classification, and sometimes also include feature selection and feature fusion [7,8,9]. In the process of image feature extraction, there are different methods such as Histogram of Oriented Gradient (HOG), Local Binary Pattern (LBP), and Principal Component Analysis (PCA) [10,11]. HOG constitutes a feature by calculating and counting the gradient direction histogram of the local area of the image. LBP is an operator used to describe the local texture features of the image; it has the advantages of rotation invariance and grayscale invariance. PCA is an effective algorithm for eliminating redundancy and simplifying datasets; it can remove redundant image features [12,13].
The deep learning uses the characteristics of its network to extract image features, discover hidden laws from it, and then achieve classification and recognition. Therefore, deep learning has achieved breakthrough results in target detection, face recognition, image classification and other fields [14,15,16,17]. In recent years, deep learning methods for Alzheimer’s disease classification have been continuously emerging, such as: the bottom-up unsupervised learning method of Stacked Auto Encoder (SAE), Deep Boltzmann Machine (DBM), and a top-down supervised learning method of deep convolutional neural network [18,19,20]. Suk et al. [21] used DBM to extract multi-modal features from PET and MRI data in the ADNI database, and used a 3D patch to pair potential hierarchical feature representations to classify AD and NC images; they got good results. Shi et al. [22] used a deep polynomial network (DPN) to classify AD and NC images of MRI and PET data respectively, and further proposed a multi-modal stacked deep polynomial network (MM-SDPN) to perform binary classification tasks, finally the accuracy of their experiment reached 96.93%. Recently, Tomassini et al. [23] proposed an end-to-end 3D convolutional long short-term memory network framework (LSTM) for early diagnosis of AD from full-resolution sMRI images.
In the process of Alzheimer’s disease research, the selection of different classifiers, model structures and appropriate attention mechanisms all play a crucial role in image classification, image recognition, and image segmentation [24]. For example, decision tree is a very common classification method [25]. It is a tree structure, each internal node represents a judgment on an attribute, each branch represents the output of a judgment result, and finally each leaf node represents a classification result. At the same time, in order to preserve the inherent characteristics of the original image and improve the good characteristics of the image in disease detection and classification, researchers usually use the latest visual sensing equipment, which can clearly observe tens of thousands of pixels in the image [26]. The vision sensor is the direct source of the machine vision system, it mainly consists of auxiliary equipment such as a graphics sensor and a light projector, which can obtain the original images that the machine vision system needs to process [27]. In addition, some researchers have studied the activation functions and pooling functions of the convolutional neural network, to compare the impact on the classification performance of Alzheimer’s disease [28,29]. Even image preprocessing is also an effective way to improve the classification performance of subsequent experiments, including template registration of images and various image filtering [30].
We conduct image classification of Alzheimer’s disease in order to better distinguish the difference between patients and normal people, and we are eager to apply it in clinical experiments in the future, but there are various uncertain problems in the research. Therefore, most researchers use neutrosophic statistics to expand and solve the uncertainty of various problems. Neutrosophic statistics refers to the statistical analysis of data samples with uncertainty, which is an extension of classical statistics and is suitable for situations where the data come from complex processes or uncertain environments [31].
In the field of computer vision, the attention mechanisms can effectively extract the feature of images. The attention mechanisms have various implementations, roughly divided into soft attention and hard attention [32,33,34]. The attention mechanism selects the focal position of the image, yielding more discriminative feature representations and bringing continuous performance improvements to the model. The soft attention mechanism means that when selecting information, it calculates the weighted average of the N input information instead of selecting only one information from the N information, and then inputs it into the neural network [35]. While the hard attention mechanism refers to selecting the information in a certain position of the input sequence, such as randomly selecting a piece of information or selecting the information with the highest probability. The visual attention mechanism can be used to pay attention to key areas in the image to obtain high-level information of image features.
The self-attention mechanism (SA) was proposed by Zhang et al. [36], and they used the weight matrix of three branches to capture the internal feature correlation of a single sample, thereby reducing the dependence on external information. But the self-attention mechanism has quadratic complexity and ignores the potential correlation between different samples. So, Guo et al. [37] proposed an external attention mechanism in 2021 to solve this problem, and they adopted two external matrices, Mk and Mv, to model the potential correlation between samples. Meanwhile, the external attention mechanism has linear complexity and implicitly considers the correlation among all data samples. Recently, Jiao et al. [38] proposed a feature fusion model for AD classification, which can comprehensively utilize multiple types of data to improve the classification performance.
Based on the fully convolutional network, this paper proposes a new method for image classification of Alzheimer’s disease that combines the external-attention mechanism with double normalization [39]. First, we obtain the feature information of the disease probability map through the FCN model, and then select the region of interest (ROI) according to the MCC heatmap of the FCN model, finally combine with age, gender, MMSE as the input of the MLP model to classify AD and NC images. The contributions of this paper are: (1) We propose a method for image classification of Alzheimer’s disease based on external-attention mechanism and fully convolutional network; (2) and add a self-attention module to the FCN model as a comparative experiment to highlight the effectiveness and efficiency of the external-attention mechanism; (3) In the normalization process of the attention map, the double normalization method of Softmax and L1 norm is used to replace the original Softmax, which can improve the classification performance in a small range.

2. Materials and Methods

2.1. Datasets

In order to prove the effectiveness of the self-attention mechanism and the external-attention mechanism, this paper uses T1-weighted Alzheimer’s disease MRI images in the ADNI dataset for experiments. The specific details of the dataset are shown in Table 1. The dataset of the experiments selects individual scan images over 55 years old, which includes a total of 550 1.5 T three-dimensional MRI images, of which 307 images belong to Alzheimer’s disease (AD) patients and 243 images belong to normal cognitive persons.

2.2. Image Preprocessing

We use the FLIRT tool in the FSL software package to align the brain magnetic resonance image with the MNI152 public template. We use the matrix that the image is invariant to affine transformation to determine the parameters of the transformation function, and then transform the original image into a standard form of image according to the transformation function determined by this parameter, finally the image size is 182 × 218 × 182. FLIRT uses coordinate rotation, translation, scaling, and shearing to match two images together, and the cost function O ( w ) is expressed in the form of quadratic summation, the intensity difference between input image and the public template is used as the optimization objective.
O ( w ) = i = 1 N ( g ( x i ( x i , w ) ) f ( x i ) ) 2 ,
f ( x i ) represents the intensity of the public template, g ( x i ) represents the intensity of the input image, x i represents the value of x i after affine transformation.
After the image is registered, we normalize the voxel values of all images. Then we control these voxel values and other outliers within a certain range to avoid the interference of background information. The flowchart of image preprocessing is shown in Figure 1.

2.3. Experimental Settings and Evaluation Criteria

The experiments in this paper use the Pytorch deep learning framework and GeForce RTX 3090 GPU processor. The FCN model uses Adam optimizer and cross-entropy loss function. In addition, the experimental parameters are set: batch size is 10, learning rate is 0.0001, and the number of training iterations is 3000. The validation set is verified every 20 iterations, then the optimal model and weights are saved. Finally, the optimal model is tested with the test set to obtain the classification performance of the FCN model and the feature information of the disease probability map.
We use accuracy and Matthews correlation coefficient (MCC) to evaluate the classification performance of the FCN model. In addition, we also record the age, gender and MMSE of AD patients and normal cognitive persons in this dataset, as the input of the MLP model’s classification experiments. As for the MLP model, we use accuracy (marked as Accu), sensitivity (marked as Sens), specificity (marked as Spec), F1 score and MCC to evaluate its classification performance. F1 score is an indicator used to measure the two-class model in statistics. It takes into account the accuracy and recall of the classification model at the same time, and can be regarded as the harmonic average of the model’s accuracy and recall. MCC comprehensively considers true positives, true negatives, false positives, and false negatives, and is a relatively balanced indicator in deep learning. In order to ensure the accuracy and reliability of the experiments, we use five-fold cross-validation for the experiments, repeat the experiments five times for the FCN model, and three times for the MLP model. The final classification performance of the models is represented by the mean and standard deviation.
A c c u = T P + T N T P + T N + F P + F N
S e n s = T P T P + F N
S p e c = T N T P + F P
F 1 = 2 T P 2 T P + F N + F P ,
M C C = T P × T N F P × F N [ ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N ) ] 0.5 .
Among them, true positives (TP) represent the correct predictions of positive samples, true negatives (TN) represent the correct predictions of negative samples, false negatives (FN) represent the false predictions of positive samples, and false positives (FP) represent the false predictions of negative samples.

2.4. Methods

2.4.1. Self-Attention Mechanism

A simplified diagram of the self-attention mechanism is shown in Figure 2. Wf, Wg and Wh are the weight matrices of the 1 × 1 × 1 convolutional layer, in order to learn the local dependencies of the image, and it also learns the long-distance global dependencies.
We take the feature map after the last convolutional block of the FCN model, passing through three branches of the 1 × 1 × 1 convolutional layers f(x), g(x), h(x) and the number of channels of the three branches is C, where H, W and D represent the length, width, and depth of the feature map respectively. After that, we transpose the output of matrix f(x) and multiply it with the output of matrix g(x), then multiply it by Softmax for normalization to obtain the attention feature map. Finally, we multiply the attention feature map with the output of matrix h(x), then pass it through the 1 × 1 × 1 convolutional layer to integrate the output into a self-attention feature map.
f ( x ) = W f x ,
g ( x ) = W g x ,
h ( x ) = W h x ,
x C × D × H × W is the original feature map before the input of three branches.
s i j = f ( x i ) T g ( x j ) ,
β j , i = exp ( s i j ) i = 1 N exp ( s i j ) ,
f ( x i ) means the values of all channels at the i-th pixel position, g ( x j ) means the values of all channels at the j-th pixel position. β j , i means the degree of attention of the model to the i-th region when synthesizing the j-th region.
The output of the self-attention module is defined as:
O j = v ( i = 1 N β j , i h ( x i ) ) , v ( x i ) = W v x i ,
We multiply the output O j of the self-attention module with a weight coefficient γ and then add it to the input feature map x i to get the final output y i of the self-attention module:
y i = γ O j + x i .
Among them, γ is a learnable parameter, and its function is to enable the network to learn the proportion of global dependence in the feature map by itself.
The tensor dimension’s changes of the feature map in the self-attention module are shown in Figure 3. In the process of feature map convolution, in order to make each eigenvalue in the image interact with each other, we change the number of channels and matrix dimensions of the convolution. D × H × W means the total pixels of the feature map.

2.4.2. External-Attention Mechanism

A simplified diagram of the external attention mechanism is shown in Figure 4. Here, the linear transformation refers to the 1 × 1 × 1 convolutional layer processing, and the Norm means double normalization; M k and M v are two external matrices.
Due to the self-attention mechanism only considering the value of a single sample to make the attention map, so the external attention mechanism uses two external matrices M k and M v which are different from the self-attention mechanism ( M k and M v are linear layer without bias), to model the similarity between the i-th pixel and the j-th pixel. Of course, the matric M is learnable, and it can also model the potential connections between different samples in the whole dataset as the training process progresses.
A = N o r m ( F M k T ) ,
F o u t = A M v .
F C × D × H × W is the output feature map of the last convolutional block, Norm means to normalize F M k T and F o u t is the output feature map of the FCN model after adding the external-attention module.
In the self-attention module, Softmax is used to normalize the attention map to make Σ j α i , j = 1 . However, the attention map is calculated by matrix multiplication, which is very sensitive to the size of the input feature. When a certain eigenvalue is very large or very small, its dot product to other eigenvalues will also become very large or very small. Therefore, the external attention module uses the double normalization of Softmax and L1 norm to solve this problem, that is, first applies Softmax to the columns of the attention map, and then applies L1 norm to the rows.
a ˜ i , j = F M k T ,
α ^ i , j = exp ( a ˜ i , j ) Σ k exp ( a ˜ k , j ) ,
α i , j = α ^ i , j Σ k α ^ i , k .
a ˜ i , j represents the input feature map F multiplying with the transposed external matrix M k . α ^ i , j represents normalizing the columns of the attention map a ˜ i , j by Softmax, and α i , j represents normalizing the rows of the attention map α ^ i , j by L1 norm. Among them, the k of a ˜ k , j and α ^ i , k represents the number of channels in the linear layer.
The tensor dimension’s changes of the feature map in the external attention module are shown in Figure 5. First, we use 1 × 1 × 1 convolutional layer to change the tensor dimension, and then multiply it with the transposed external matrix M k . The k means the number of channels of the external matrix and the D × H × W means the total pixels of the feature map. After that, we apply the double normalization to normalize the attention map. Then, we multiply the attention map with the external matrix M v to put it into a new 1 × 1 × 1 convolutional layer. Finally, we can obtain the external-attention feature map with the same tensor dimension as the original feature map.

2.4.3. Model’s Framework

The FCN model consists of four convolutional blocks and two fully connected layers. Among them, the convolutional block includes 3D convolutional layer, 3D maxpool layer, 3D batch normalization, Leaky ReLU and Dropout, as shown in the Figure 6. The last two fully connected layers play a role in improving the efficiency of the model in the classification task. The network is trained by randomly initializing weights. As shown in Figure 7, we adopt a method of randomly sampling patches of the 3D-MRI images to train the FCN model, that is, training patches with a random sampling size of 47 × 47 × 47 from the 3D-MRI images. And the size of each patch is the same as the receptive field of FCN model.
Because convolution operation reduces the output size of the input layer, each Patch will generate two scalar values after being trained by the FCN model. Then they are converted into the Alzheimer’s disease probability and normal cognitive probability of the corresponding pixel under the action of the activation function Softmax. The disease state of the brain’s local structure is displayed through each pixel’s risk probability value of Alzheimer’s disease. The corresponding feature information of the disease probability map will be used as auxiliary information of the MLP model for classification experiments.
In order to conduct comparative experiments with the FCN model, we also use the same network for the CNN model, as shown in Figure 8. The specific parameter settings of the hidden layer of the CNN model are shown in Table 2.
After that, we build the MLP model’s framework, as shown in Figure 9. The MLP model consists of two fully connected layers; batch normalization, Leaky ReLU and Dropout. For the MLP model, we select the probability value of Alzheimer’s disease from the feature information of the disease probability map, and select the region of interest (ROI) based on the MCC value of the FCN model, then combine with the age, gender, MMSE of the MRI image as the input of the MLP model to reclassify MRI images.
Among them, the MCC value can show the overall classification performance of the FCN model, and the MCC heatmap can show that the FCN model has a higher classification accuracy for certain pixel positions of the 3D-MRI image. Therefore, we select these high-accuracy regions as regions of interest (ROI).
According to Figure 9, the MLP-A model indicates that only the feature information of the disease probability map of the FCN model is used to classify MRI images; The MLP-B model indicates that only the image information of age, gender and MMSE are used to classify MRI images; The MLP-C model indicates combining the feature information of the disease probability map of the FCN model with age, gender, MMSE to classify MRI images.
In addition, the specific FCN model’s parameter settings and the changes of output patch size in our experiments are shown in the Table 3.

3. Experiments and Results

First, this paper conducts experiments on the original FCN model and MLP model, then adds the self-attention module and the external-attention module respectively for multiple experiments. Finally, we use the mean and standard deviation to represent the classification performance of the model. The double normalization of Softmax and L1 norm are used to replace Softmax in the external-attention module. The MCC heatmap of the FCN model is shown in Figure 10.
The feature information of the disease probability map generated by the FCN model is shown in the Figure 11. Red and blue indicate the probability of suffering from Alzheimer’s disease in different parts of the brain. The dividing line between the two is 0.5.
We have summarized the changes in the accuracy of the FCN models and the MLP models in different situations. The changes of the FCN models’ accuracy and the MLP models’ accuracy after adding the self-attention mechanism and the external-attention mechanism are shown in Figure 12.
In detail, the changes of the MLP models’ accuracy after adding the self-attention mechanism and the external-attention mechanism, as well as the double normalization are shown in Figure 13.
Of course, we also plot the changes of the MLP models’ MCC value after adding the self-attention mechanism, the external attention mechanism and double normalization, as shown in Figure 14.
Next, we list the various experimental results of the FCN models and the MLP models, which are the final classification performance (mean and standard deviation), including accuracy, sensitivity, specificity, F1 score, MCC. The classification performance of the FCN model without any attention module is shown in Table 4.
As a comparative experiment, the experimental results of the CNN model and the MLP fusion model are shown in Table 5. Among them, the fusion model means that MLP model combines the feature information of the CNN model with age, gender, and MMSE to classify MRI images.
Comparing the experimental results of the MLP-C in Table 4 with the fusion model in Table 5. It can be found that selecting the region of interest (ROI) of the MLP model through the MCC heatmap of the FCN model, and combining with the feature information of the disease probability map can improve the classification performance better than the MLP fusion model.
We use accuracy and MCC to evaluate the classification performance of the FCN model, as shown in Table 6.
At here, the MCC is calculated by using each pixel in the 3D-MRI image as a sample. After each pixel in the 3D-MRI image is trained by the FCN model, a predicted probability value of Alzheimer’s disease will be generated. The prediction of each pixel is compared with the input label, and then the corresponding pixel is marked as TP, TN, FP, FN.
The classification performance of the MLP models after adding the self-attention module is shown in Table 7.
The results in Table 6 show that after the self-attention module is integrated into the FCN model, the accuracy increases of about 1.155%, and the MCC value increases of about 2.454%. Comparing the experimental results in Table 4 and Table 7, it can be found that after adding the self-attention module, for the MLP models, the accuracy increases by about 0.57% to 2.62%, and the MCC value increases by about 0.60% to 3.03%. This shows the effectiveness of the self-attention mechanism and it can improve the classification performance of the model.
On the other hand, the experimental results in Table 6 show that after adding the external-attention module and double normalization, compared with the original FCN model, the accuracy increases by about 3.366%, and the MCC value increases by about 5.195%. Furthermore, double normalization compares with Softmax, the accuracy increases by about 0.803%, and the MCC value increases by about 0.432%.
After adding the external-attention module, the classification performance of the MLP models by using double normalization or Softmax respectively are shown in Table 8 and Table 9.
Table 8 and Table 9 compare the classification performance difference between using double normalization with only using Softmax in the external-attention mechanism. Experimental results show that double normalization can increase the accuracy of the MLP models by about 0.22% to 1.12%, and the MCC value by about 0.21% to 2.39%. This shows that double normalization can improve the classification performance in a small range, highlighting the effectiveness of double normalization.
The classification index sensitivity of the model represents the proportion of all positive samples that are paired and measures the model’s ability to discriminate against positive samples. From the experimental results, after adding an external-attention mechanism to the FCN model and combining with double normalization, the sensitivity of MLP-A model classification is 92.6%, the sensitivity of MLP-B model classification is 99.02%, and the sensitivity of MLP-C model classification is 99.29%. This also shows that our proposed model has a high discriminative ability for images of Alzheimer’s disease patients.

4. Discussion

The above experimental results show that the external-attention mechanism can generate richer feature information of a disease probability map for the MLP models, thereby improving the classification performance of the model. In addition, the MLP models combine with age, gender, and MMSE, which are more conducive to the accurate judgment of image classification, in the case of selecting the region of interest (ROI) according to the FCN model.
The reason we choose to add the self-attention module to the FCN model as a comparative experiment is because the external-attention mechanism changes the weight matrix on the basis of the self-attention mechanism. The self-attention mechanism calculates the direct interaction between any two locations, allowing the network to focus on areas that are scattered in different locations. However, this self-attention mechanism only considers the correlations within a single sample, it ignores the potential connections between samples. Therefore, the external-attention mechanism uses a learnable external matrix to establish potential correlations between samples. In addition, the Softmax in the self-attention module normalizes the attention map, but the attention map is calculated by matrix multiplication, which is sensitive to the size of the input feature and susceptible to particularly large or small feature values. Therefore, the double normalization in the external-attention module first applies Softmax to the columns, and then applies L1 norm to the rows to solve this problem. The experimental results show that after adding external-attention mechanism to the fully convolutional network and combining with double normalization, the classification performance of the MLP models is better than other comparison methods. However, due to the many unknown factors of the deep learning model, the limitation of our study is that it is difficult to quantify and visually analyze the correspondence between models and results, which makes it difficult to apply in clinical practice. Therefore, we expect further research on the interpretability of the deep learning model in order to improve the confidence of the classification results.
Finally, we compare with the models of other references on Alzheimer’s disease, as shown in Table 10. Different classification techniques have their own advantages and disadvantages. For example, SVM uses the inner product kernel function to replace the nonlinear mapping to the high-dimensional space, and its goal is to divide the feature space into the optimal classification hyperplane [40]. However, its disadvantage is that it is difficult to implement large-scale training samples, and it is sensitive to the choice of parameter adjustment and function.

5. Conclusions and Future Work

In this paper, the self-attention mechanism models the correlations within the samples to obtain the corresponding attention feature maps, which plays a certain role in improving the classification performance of the model. Moreover, this paper proposes a new method for the image classification of Alzheimer’s disease based on the external-attention mechanism and double normalization, which embeds the external-attention module after the last convolutional block of the FCN model. The detailed experimental results show that the external-attention mechanism is effective and efficient in improving classification performance. We used the double normalization of Softmax and L1 norm to replace the original Softmax, so that the classification accuracy of the MLP model can increase by about 0.22% to 1.12%. The core of this paper is to integrate the external-attention mechanism into the FCN model to obtain richer and more detailed feature information of a disease probability map.
In the future, we will consider using other deep learning methods or new efficient attention mechanisms to continuously tap the potential of fully convolutional networks. In the image preprocessing, we will try to use different image denoising and smoothing methods to effectively remove noise from the original MRI image, and further improve the classification performance of the model. In addition, we will try different experimental approaches using 2D slices and 3D patches to compare the classification performance of the two [43], which has special implications for Alzheimer’s disease classification research.

Author Contributions

Conceptualization and methodology, M.J.; formal analysis and investigation, Y.L.; resources, T.L.; data curation, J.Z.; writing—original draft preparation, B.Y.; writing—review and editing, M.J. and Y.L.; visualization, W.K.; project administration, M.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (61672466, 62011530130, 62101497), Joint Fund of Zhejiang Provincial Natural Science Foundation (LSZ19F010001), the Key Research and Development Program of Zhejiang Province (2020C03060), the 521 Talents project of Zhejiang Sci-Tech University, and this work is also supported by Science Foundation of Zhejiang Sci-Tech University under Grant No. 19032111-Y.

Institutional Review Board Statement

Ethical review and approval were waived for this study, due to all research data are from open-source datasets.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alzheimer’s Disease & Related Dementias: Alzheimer’s Disease Fact Sheet. Available online: https://www.nia.nih.gov/health/alzheimers-disease-fact-sheet (accessed on 8 July 2021).
  2. Kruthika, K.R.; Rajeswari; Maheshappa, H.D. CBIR system using Capsule Networks and 3D CNN for Alzheimer’s disease diagnosis. Inform. Med. Unlocked 2019, 14, 59–68. [Google Scholar] [CrossRef]
  3. Zhang, Y.; Wang, S.; Xia, K.; Jiang, Y.; Qian, P. Alzheimer’s disease multiclass diagnosis via multimodal neuroimaging embedding feature selection and fusion. Inf. Fusion 2021, 66, 170–183. [Google Scholar] [CrossRef]
  4. Fan, Z.; Xu, F.; Qi, X. Classification of Alzheimer’s disease based on brain MRI and machine learning. Neural Comput. Appl. 2020, 32, 1928–1936. [Google Scholar] [CrossRef]
  5. Zeng, N.; Qiu, H.; Wang, Z.; Liu, W.; Zhang, H.; Li, Y. A new switching-delayed-PSO-based optimized SVM algorithm for diagnosis of Alzheimer’s disease. Neurocomputing 2018, 320, 195–202. [Google Scholar] [CrossRef]
  6. Jo, T.; Nho, K.; Saykin, A.J. Deep learning in Alzheimer’s disease: Diagnostic classification and prognostic prediction using neuroimaging data. Front. Aging Neurosci. 2019, 11, 220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Chen, Y.; Xia, Y. Iterative sparse and deep learning for accurate diagnosis of Alzheimer’s disease. Pattern Recognit. 2021, 116, 107944. [Google Scholar] [CrossRef]
  8. Beheshti, I.; Demirel, H. Alzheimer’s Disease Neuroimaging Initiative. Feature-ranking-based Alzheimer’s disease classification from structural MRI. Magn. Reson. Imaging 2016, 34, 252–263. [Google Scholar] [CrossRef]
  9. Lin, W.M.; GAO, Q.Q.; Du, M. Multiclass diagnosis of stages of Alzheimer’s disease using linear discriminant analysis scoring for multimodal data. Comput. Biol. Med. 2021, 134, 104478. [Google Scholar] [CrossRef] [PubMed]
  10. Jalal, A.; Kamal, S.; Kim, D. Depth Silhouettes Context: A new robust feature for human tracking and activity recognition based on embedded HMMs. In Proceedings of the 12th IEEE International Conference on Ubiquitous Robots and Ambient Intelligence, KINTEX, Goyang City, Korea, 28 October 2015; pp. 294–299. [Google Scholar]
  11. Kamal, S.; Jalal, A.; Kim, D. Depth Images-based Human Detection, Tracking and Activity Recognition Using Spatiotemporal Features and Modified HMM. J. Electr. Eng. Technol. 2016, 11, 1921–1926. [Google Scholar] [CrossRef] [Green Version]
  12. Setiawan, F.; Lin, C.W. Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features. Brain Sci. 2021, 11, 902. [Google Scholar] [CrossRef] [PubMed]
  13. Ahmad, F.; Dar, W.M. Classification of Alzheimer’s Disease Stages: An Approach Using PCA-Based Algorithm. American. J. Alzheimers Dis. Other Dement. 2018, 33, 433–439. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, M.; Cheng, D.; Yan, W. Classification of Alzheimer’s Disease by Combination of Convolutional and Recurrent Neural Networks Using FDG-PET Images. Front. Neuroinform. 2018, 12, 35. [Google Scholar]
  15. Turkson, R.E.; Hong, Q. Classification of Alzheimer’s Disease Using Deep Convolutional Spiking Neural Network. Neural Processing Lett. 2021, 53, 2649–2663. [Google Scholar] [CrossRef]
  16. Tufail, A.B.; Zhang, Q.N.; Ma, Y.K. Binary Classification of Alzheimer Disease using sMRI Imaging modality and Deep Learning. J. Digit. Imaging 2020, 33, 1073–1090. [Google Scholar] [CrossRef] [PubMed]
  17. Ramzan, F.; Khan, M.U.G.; Rehmat, A.; Iqbal, S.; Saba, T.; Rehman, A.; Mehmood, Z. A Deep Learning Approach for Automated Diagnosis and Multi-Class Classification of Alzheimer’s Disease Stages Using Resting-State fMRI and Residual Neural Networks. J. Med. Syst. 2019, 44, 37. [Google Scholar] [CrossRef] [PubMed]
  18. Payan, A.; Montana, G. Predicting Alzheimer’s disease: A neuroimaging study with 3D convolutional neural networks. arXiv 2015, arXiv:1502.02506v1. [Google Scholar]
  19. Vu, T.D.; Yang, H.J.; Nguyen, V.Q. Multimodal learning using convolution neural network and Sparse Autoencoder. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju Island, Korea, 13–16 February 2017; pp. 309–312. [Google Scholar]
  20. Liu, S.; Cai, W. Multimodal Neuroimaging Feature Learning for Multiclass Diagnosis of Alzheimer’s Disease. IEEE Trans. Biomed. Eng. 2015, 62, 1132–1140. [Google Scholar] [CrossRef] [Green Version]
  21. Suk, H.I.; Lee, S.W.; Shen, D. Hierarchical feature representation and multimodal fusion with deep learning for AD/MCI diagnosis. NeuroImage 2014, 101, 569–582. [Google Scholar] [CrossRef] [Green Version]
  22. Shi, J.; Zheng, X.; Li, Y. Multimodal Neuroimaging Feature Learning with Multimodal Stacked Deep Polynomial Networks for Diagnosis of Alzheimer’s Disease. IEEE J. Biomed. Health Inform. 2018, 22, 173–183. [Google Scholar] [CrossRef]
  23. Tomassini, S.; Falcionelli, N.; Sernani, P. An End-to-End 3D ConvLSTM-based Framework for Early Diagnosis of Alzheimer’s Disease from Full-Resolution Whole-Brain sMRI Scans. In Proceedings of the 2021 IEEE 34th International Symposium on Computer-Based Medical Systems (CBMS), Aveiro, Portugal, 7–9 June 2021; pp. 74–78. [Google Scholar]
  24. Nguyen, T.V.; Pham, N.; Tran, T.; Le, B. Higher Order Conditional Random Field for Multi-Label Interactive Image Segmentation. In Proceedings of the 2012 IEEE RIVF International Conference on Computing & Communication Technologies, Research, Innovation, and Vision for the Future, Ho Chi Minh City, Vietnam, 27 February–1 March 2012; pp. 1–4. [Google Scholar]
  25. Azam, M.; Aslam, M.; Pfeiffer, K.P. Three Steps Strategy to Search For Optimum Classification Trees. Commun. Stat. Simul. Comput. 2016, 45, 548–565. [Google Scholar] [CrossRef]
  26. Puwein, J.; Ballan, L.; Ziegler, R. Joint camera pose estimation and 3D human pose estimation in a multi-camera setup. In Proceedings of the IEEE ACCV, Singapore, 1–5 November 2014; pp. 473–487. [Google Scholar]
  27. Jalal, A.; Kim, Y.; Kamal, S.; Farooq, A.; Kim, D. Human daily activity recognition with joints plus body features representation using Kinect sensor. In Proceedings of the IEEE International Conference on Informatics, Electronics and Vision, Fukuoka, Japan, 15–18 June 2015; pp. 1–6. [Google Scholar]
  28. Stergiou, A.; Poppe, R.; Kalliatakis, G. Refining activation downsampling with SoftPool. arXiv 2021, arXiv:2101.00440v1. [Google Scholar]
  29. Wang, S.H.; Phillips, P.; Sui, Y. Classification of Alzheimer’s Disease Based on Eight-Layer Convolutional Neural Network with Leaky Rectified Linear Unit and Max Pooling. J. Med. Syst. 2018, 42, 85. [Google Scholar] [CrossRef] [PubMed]
  30. Lenhart, L.; Seiler, S.; Pirpamer, L.; Goebel, G.; Potrusil, T.; Wagner, M.; Scherfler, C. Anatomically Standardized Detection of MRI Atrophy Patterns in Early-Stage Alzheimer’s Disease. Brain Sci. 2021, 11, 1491. [Google Scholar] [CrossRef] [PubMed]
  31. Smarandache, F. Introduction to Neutrosophic Statistics. arXiv 2014, arXiv:1406.2000v1. [Google Scholar]
  32. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2019), Long Beach, CA, USA, 16–20 June 2019; pp. 3146–3154. [Google Scholar]
  33. Li, X.; Zhong, Z.; Wu, J.; Yang, Y.; Lin, Z.; Liu, H. Expectation- maximization attention networks for semantic segmentation. In Proceedings of the 2019 IEEE CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 20–26 October 2019; pp. 9166–9175. [Google Scholar]
  34. Woo, S.; Park, J.; Lee, J.Y. CBAM: Convolutional Block Attention Module. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  35. Zhang, H.; Zu, K.; Lu, J. EPSANet: An Efficient Pyramid Split Attention Block on Convolutional Neural Network. arXiv 2021, arXiv:2105.14447v2. [Google Scholar]
  36. Zhang, H.; Goodfellow, I.; Metaxas, D. Self-Attention Generative Adversarial Networks. arXiv 2018, arXiv:1805.08318v1. [Google Scholar]
  37. Guo, M.H.; Liu, Z.N.; Mu, T.J. Beyond Self-attention: External Attention using Two Linear Layers for Visual Tasks. arXiv 2021, arXiv:2105.02358v2. [Google Scholar]
  38. Jiao, Z.; Chen, S.; Shi, H.; Xu, J. Multi-Modal Feature Selection with Feature Correlation and Feature Structure Fusion for MCI and AD Classification. Brain Sci. 2022, 12, 80. [Google Scholar] [CrossRef]
  39. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
  40. Vantigodi, S.; Babu, R.V. Real-time human action recognition from motion capture data. In Proceedings of the 2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG), Jodhpur, India, 18–21 December 2013; pp. 1–4. [Google Scholar]
  41. Ullah, H.; Onik, Z.A.; Islam, R. Alzheimer’s Disease and Dementia Detection from 3D Brain MRI Data Using Deep Convolutional Neural Networks. In Proceedings of the 3rd International Conference for Convergence in Technology (I2CT), Pune, India, 6–7 April 2018; pp. 1–3. [Google Scholar]
  42. Hinrichs, C.; Singh, V.; Mukherjee, L. Spatially augmented LPboosting for AD classification with evaluations on the ADNI dataset. Neuroimage 2009, 48, 138–149. [Google Scholar] [CrossRef] [Green Version]
  43. Wen, J.; Thibeau-Sutre, E.; Diaz-Melo, M.; Samper-González, J.; Routier, A. Convolutional neural networks for classification of Alzheimer’s disease: Overview and reproducible evaluation. Med. Image Anal. 2020, 63, 101694. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The flowchart of image preprocessing.
Figure 1. The flowchart of image preprocessing.
Brainsci 12 00319 g001
Figure 2. Diagram of the self-attention mechanism.
Figure 2. Diagram of the self-attention mechanism.
Brainsci 12 00319 g002
Figure 3. The feature maps in the self-attention module.
Figure 3. The feature maps in the self-attention module.
Brainsci 12 00319 g003
Figure 4. Diagram of the external-attention mechanism.
Figure 4. Diagram of the external-attention mechanism.
Brainsci 12 00319 g004
Figure 5. The feature maps in the external-attention module.
Figure 5. The feature maps in the external-attention module.
Brainsci 12 00319 g005
Figure 6. The FCN model’s framework.
Figure 6. The FCN model’s framework.
Brainsci 12 00319 g006
Figure 7. Randomly sampling 3D-MRI image’s patches for training the FCN Model.
Figure 7. Randomly sampling 3D-MRI image’s patches for training the FCN Model.
Brainsci 12 00319 g007
Figure 8. The CNN model’s framework.
Figure 8. The CNN model’s framework.
Brainsci 12 00319 g008
Figure 9. The MLP model’s framework.
Figure 9. The MLP model’s framework.
Brainsci 12 00319 g009
Figure 10. (A) The MCC value can show the overall classification performance of the FCN model. From the MCC heatmap, it can be observed that some locations have higher MCC values (that is, these locations have higher classification accuracy). The MLP model uses these specific locations as a region of interest (ROI). (BD) represents the MCC value of the FCN model in the individual axial, coronal and sagittal directions.
Figure 10. (A) The MCC value can show the overall classification performance of the FCN model. From the MCC heatmap, it can be observed that some locations have higher MCC values (that is, these locations have higher classification accuracy). The MLP model uses these specific locations as a region of interest (ROI). (BD) represents the MCC value of the FCN model in the individual axial, coronal and sagittal directions.
Brainsci 12 00319 g010
Figure 11. (A) The disease probability map generated by the FCN model highlights the brain regions at high risk of Alzheimer’s disease. The first two samples were clinically diagnosed as patients with Alzheimer’s disease, and the latter two samples were clinically confirmed as normal cognitive persons. (BD) shows the axial, coronal and sagittal disease probability map of patients who are clinically diagnosed with Alzheimer’s disease. Red indicates that the risk of Alzheimer’s disease is >0.5, and blue indicates <0.5.
Figure 11. (A) The disease probability map generated by the FCN model highlights the brain regions at high risk of Alzheimer’s disease. The first two samples were clinically diagnosed as patients with Alzheimer’s disease, and the latter two samples were clinically confirmed as normal cognitive persons. (BD) shows the axial, coronal and sagittal disease probability map of patients who are clinically diagnosed with Alzheimer’s disease. Red indicates that the risk of Alzheimer’s disease is >0.5, and blue indicates <0.5.
Brainsci 12 00319 g011
Figure 12. (a) The changes of the FCN models’ accuracy; (b) The changes of the MLP models’ accuracy.
Figure 12. (a) The changes of the FCN models’ accuracy; (b) The changes of the MLP models’ accuracy.
Brainsci 12 00319 g012
Figure 13. (a) No attention mechanism is added to models; (b) Models add self-attention mechanism; (c) Models add external-attention mechanism and Softmax; (d) Models add external-attention mechanism and double normalization.
Figure 13. (a) No attention mechanism is added to models; (b) Models add self-attention mechanism; (c) Models add external-attention mechanism and Softmax; (d) Models add external-attention mechanism and double normalization.
Brainsci 12 00319 g013
Figure 14. (a) Models without attention mechanism; (b) Models add self-attention mechanism; (c) Models add external-attention mechanism and Softmax; (d) Models add external-attention mechanism and double normalization.
Figure 14. (a) Models without attention mechanism; (b) Models add self-attention mechanism; (c) Models add external-attention mechanism and Softmax; (d) Models add external-attention mechanism and double normalization.
Brainsci 12 00319 g014aBrainsci 12 00319 g014b
Table 1. Detailed information of the experimental dataset.
Table 1. Detailed information of the experimental dataset.
DatasetADNI
Research objectADNC
Number of samples307243
Average age76.3 (57–92)79.4 (65–87)
Gender (Male/Female)180/127101/142
Average MMSE22.8 (19–27)28.6 (26–30)
Table 2. The parameter settings of CNN model.
Table 2. The parameter settings of CNN model.
Input LayerDetailed DescriptionOutput Size
Input (1, 182, 218, 182)
3D convolutional layer1channel 20, kernel 7, stride 2, padding 0(20, 88, 106, 88)
3D maxpool layer1kernel 3, stride 2, padding 0(20, 43, 52, 43)
3D batch normalization1eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU1; Dropout1Negative slope = 0.01; p = 0.1
3D convolutional layer2channel 40, kernel 4, stride 1, padding 0(40, 40, 49, 40)
3D maxpool layer2kernel 2, stride 2, padding 0(40, 20, 24, 20)
3D batch normalization2eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU2; Dropout2Negative slope = 0.01; p = 0.1
3D convolutional layer3channel 80, kernel 3, stride 1, padding 0(80, 18, 22, 18)
3D maxpool layer3kernel 2, stride 2, padding 0(80, 9, 11, 9)
3D batch normalization3eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU3; Dropout3Negative slope = 0.01; p = 0.1
3D convolutional layer4channel 160, kernel 3, stride 1, padding 0(160, 7, 9, 7)
3D maxpool layer4kernel 2, stride 1, padding 0(160, 6, 8, 6)
3D batch normalization4eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU4; Dropout4Negative slope = 0.01; p = 0.1
Flatten (46, 80)
Dropout5p = 0.1
Fully connected layer1channel 30(30)
Leaky ReLU5; Dropout6Negative slope = 0.01; p = 0.1
Fully connected layer2channel 2(2)
Table 3. The parameter settings of FCN model.
Table 3. The parameter settings of FCN model.
Input LayerDetailed DescriptionOutput Patch Size
Input (1, 47, 47, 47)
3D convolutional layer1channel 20, kernel 4, stride 1, padding 0(20, 44, 44, 44)
3D maxpool layer1kernel 2, stride 1, padding 0(20, 43, 43, 43)
3D batch normalization1eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU1; Dropout1Negative slope = 0.01; p = 0.1
3D convolutional layer2channel 40, kernel 4, stride 1, padding 0(40, 40, 40, 40)
3D maxpool layer2kernel 2, stride 2, padding 0(40, 20, 20, 20)
3D batch normalization2eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU2; Dropout2Negative slope = 0.01; p = 0.1
3D convolutional layer3channel 80, kernel 3, stride 1, padding 0(80, 18, 18, 18)
3D maxpool layer3kernel 2, stride 2, padding 0(80, 9, 9, 9)
3D batch normalization3eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU3; Dropout3Negative slope = 0.01; p = 0.1
3D convolutional layer4channel 160, kernel 3, stride 1, padding 0(160, 7, 7, 7)
3D maxpool layer4kernel 2, stride 1, padding 0(160, 6, 6, 6)
3D batch normalization4eps = 1 × 10−5, momentum = 0.1, affine = True
Leaky ReLU4; Dropout4Negative slope = 0.01; p = 0.1
Fully connected layer1channel 30, kernel 6, stride 1, padding 0(30, 1, 1, 1)
Leaky ReLU5; Dropout5Negative slope = 0.01; p = 0.1
Fully connected layer2channel 2, kernel 1, stride 1, padding 0(2, 1, 1, 1)
Table 4. The classification performance of the MLP models without any attention module.
Table 4. The classification performance of the MLP models without any attention module.
FCNAccuSensSpecF1MCC
MLP-A0.8685 ± 0.01400.8444 ± 0.03560.8943 ± 0.04810.8693 ± 0.01310.7408 ± 0.0288
MLP-B0.9688 ± 0.01030.9526 ± 0.02210.9642 ± 0.01750.9691 ± 0.01040.9485 ± 0.0197
MLP-C0.9728 ± 0.01430.9643 ± 0.01800.9674 ± 0.01390.9757 ± 0.01310.9534 ± 0.0177
Table 5. The experimental results of the CNN model and the MLP fusion model.
Table 5. The experimental results of the CNN model and the MLP fusion model.
AccuSensSpecF1MCC
CNN0.8636 ± 0.02370.8875 ± 0.01530.8696 ± 0.03640.8287 ± 0.04880.7549 ± 0.0368
MLP fusion model0.9188 ± 0.02210.9439 ± 0.03400.8918 ± 0.02800.9232 ± 0.02150.8389 ± 0.0446
Table 6. The classification performance of the FCN model.
Table 6. The classification performance of the FCN model.
AccuMCC
FCN0.58449 ± 0.01290.16132 ± 0.0264
FCN + SA0.59604 ± 0.00520.18586 ± 0.0076
FCN + EA + Softmax0.61012 ± 0.01050.20895 ± 0.0090
FCN + EA + Double normalization0.61815 ± 0.00690.21327 ± 0.0138
Table 7. The classification performance of the MLP models after adding the self-attention module.
Table 7. The classification performance of the MLP models after adding the self-attention module.
FCN + SAAccuSensSpecF1MCC
MLP-A0.8947 ± 0.01030.8696 ± 0.01620.9107 ± 0.01340.8846 ± 0.01230.7711 ± 0.0189
MLP-B0.9745 ± 0.01010.9649 ± 0.02480.9689 ± 0.02200.9751 ± 0.01000.9531 ± 0.0193
MLP-C0.9788 ± 0.01130.9704 ± 0.02410.9712 ± 0.01520.9792 ± 0.01130.9584 ± 0.0219
Table 8. The classification performance of the MLP models after adding the external-attention module and double normalization.
Table 8. The classification performance of the MLP models after adding the external-attention module and double normalization.
FCN + EA +
Double Normalization
AccuSensSpecF1MCC
MLP-A0.9236 ± 0.01930.9260 ± 0.01410.9356 ± 0.01630.9292 ± 0.01820.8492 ± 0.0180
MLP-B0.9855 ± 0.00450.9902 ± 0.00800.9796 ± 0.01290.9869 ± 0.00400.9707 ± 0.0090
MLP-C0.9873 ± 0.00690.9929 ± 0.00410.9828 ± 0.01850.9889 ± 0.00760.9749 ± 0.0130
Table 9. The classification performance of the MLP models after adding the external-attention module and Softmax.
Table 9. The classification performance of the MLP models after adding the external-attention module and Softmax.
FCN + EA + SoftmaxAccuSensSpecF1MCC
MLP-A0.9124 ± 0.01120.9116 ± 0.01110.9234 ± 0.01560.9111 ± 0.01000.8253 ± 0.0140
MLP-B0.9828 ± 0.01470.9867 ± 0.00890.9701 ± 0.02460.9836 ± 0.01290.9686 ± 0.0296
MLP-C0.9851 ± 0.00990.9906 ± 0.00720.9741 ± 0.02040.9876 ± 0.00880.9721 ± 0.0198
Table 10. Compare with the classification models of other researchers.
Table 10. Compare with the classification models of other researchers.
AuthorType of DatasetMethodsNumber of SamplesAccuracy
Liu S et al. [20]MRIStacked auto-encoder (SAE) + region-level engineered features180 AD/204 NC0.79
Shi J et al. [22]MRIDeep Polynomial Network (DPN)51 AD/52 NC0.9076
Tomassini S et al. [23]MRIBased on long short-term memory network (LSTM)213 AD/214 NC0.86
Ullah H et al. [41]MRIDeep Convolutional Network (3D-CNN)416 (AD + NC)0.8025
Hinrichs C et al. [42]MRISVM + Linear Program boost (LP) + voxel-level engineered features183 (AD + NC)0.82
Suk H I et al. [21]MRIDeep Boltzmann Machine93 AD/101 NC0.9238
PET0.9220
Our proposed methodsMRIFCN + SA307 AD/243 NC0.9788
FCN + EA + softmax0.9851
FCN + EA + double normalization0.9873
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiang, M.; Yan, B.; Li, Y.; Zhang, J.; Li, T.; Ke, W. Image Classification of Alzheimer’s Disease Based on External-Attention Mechanism and Fully Convolutional Network. Brain Sci. 2022, 12, 319. https://doi.org/10.3390/brainsci12030319

AMA Style

Jiang M, Yan B, Li Y, Zhang J, Li T, Ke W. Image Classification of Alzheimer’s Disease Based on External-Attention Mechanism and Fully Convolutional Network. Brain Sciences. 2022; 12(3):319. https://doi.org/10.3390/brainsci12030319

Chicago/Turabian Style

Jiang, Mingfeng, Bin Yan, Yang Li, Jucheng Zhang, Tieqiang Li, and Wei Ke. 2022. "Image Classification of Alzheimer’s Disease Based on External-Attention Mechanism and Fully Convolutional Network" Brain Sciences 12, no. 3: 319. https://doi.org/10.3390/brainsci12030319

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop