Next Article in Journal
Advancing the Role of Neuroimmunity and Genetic Susceptibility in Gulf War Illness
Next Article in Special Issue
Gross Total Resection Promotes Subsequent Recovery and Further Enhancement of Impaired Natural Killer Cell Activity in Glioblastoma Patients
Previous Article in Journal
Application of Network Analysis to Uncover Variables Contributing to Functional Recovery after Stroke
Previous Article in Special Issue
Endovascular Treatment of ICAS Patients: Targeting Reperfusion Rather than Residual Stenosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Deep Learning Radiomics Model to Discriminate AD, MCI and NC: An Exploratory Study Based on Tau PET Scans from ADNI †

1
Nuclear Medicine and Molecular Imaging Key Laboratory of Sichuan Province, Luzhou 646000, China
2
Department of Nuclear Medicine, Affiliated Hospital of Southwest Medical University, Luzhou 646000, China
3
Institute of Nuclear Medicine, Southwest Medical University, Luzhou 646000, China
4
School of Pharmacy, Southwest Medical University, Luzhou 646000, China
5
School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China
6
Institute of Biomedical Engineering, School of Life Science, Shanghai University, Shanghai 200444, China
*
Authors to whom correspondence should be addressed.
For the Alzheimer’s Disease Neuroimaging Initiative.
The two authors contribute equally to this work.
Brain Sci. 2022, 12(8), 1067; https://doi.org/10.3390/brainsci12081067
Submission received: 31 July 2022 / Revised: 9 August 2022 / Accepted: 10 August 2022 / Published: 12 August 2022
(This article belongs to the Special Issue Challenges and Perspectives of Neurological Disorders)

Abstract

:
Objective: We explored a novel model based on deep learning radiomics (DLR) to differentiate Alzheimer’s disease (AD) patients, mild cognitive impairment (MCI) patients and normal control (NC) subjects. This model was validated in an exploratory study using tau positron emission tomography (tau-PET) scans. Methods: In this study, we selected tau-PET scans from the Alzheimer’s Disease Neuroimaging Initiative database (ADNI), which included a total of 211 NC, 197 MCI, and 117 AD subjects. The dataset was divided into one training/validation group and one separate external group for testing. The proposed DLR model contained the following three steps: (1) pre-training of candidate deep learning models; (2) extraction and selection of DLR features; (3) classification based on support vector machine (SVM). In the comparative experiments, we compared the DLR model with three traditional models, including the SUVR model, traditional radiomics model, and a clinical model. Ten-fold cross-validation was carried out 200 times in the experiments. Results: Compared with other models, the DLR model achieved the best classification performance, with an accuracy of 90.76% ± 2.15% in NC vs. MCI, 88.43% ± 2.32% in MCI vs. AD, and 99.92% ± 0.51% in NC vs. AD. Conclusions: Our proposed DLR model had the potential clinical value to discriminate AD, MCI and NC.

1. Introduction

Alzheimer’s disease (AD) is the most prevalent cause of dementia and the most significant disease threatening the health of the elderly [1]. In the early stages of AD, patients often exhibit mild cognitive damage, i.e., mild memory loss mild executive function decrements (e.g., amyloid and tau pathological mechanism), and visuospatial impairment [2,3]. Mild cognitive impairment (MCI) is an intermediate step between normal aging and dementia [4], where patients start to appear memory impairment or other cognitive abnormalities, but have not reached the severity of dementia. Mild cognitive impairment subjects are at high-risk step for dementia [5]. Therefore, it is important to discriminate AD, MCI and normal control (NC) individuals [6,7].
tau positron emission tomography (tau-PET) imaging technology has become increasingly popular for the clinical diagnosis of AD and MCI [8,9,10]. The degree of brain tau accumulation, as an objective biomarker, is strongly correlated with the severity of AD. Johnson et al. found that abnormally high cortical tau binding in the inferior temporal gyrus was associated with clinical impairment [11]. Zhao et al. found that typical deposits of tau appeared in the amygdala, entorhinal cortex, fusiform and parahippocampus in AD brains [12]. La Joie et al. included 28 AD patients and 25 patients with a non-AD clinical neurodegenerative diagnosis and found that tau-PET standard uptake value (SUVR) in the whole brain showed excellent discrimination power (area under curve (AUC) = 0.92–0.94) for diagnosing AD and MCI [13]. Sun et al. proposed a random forest diagnostic model for the classification of NC, MCI and AD and achieved an accuracy of 81.6% [14]. However, existing diagnosis models still have shortcomings, such as the need to manually extract features from the region of interest (ROI) and to encode the extracted features, which often requires tedious processes. Thus, an alternative approach is needed.
Deep learning radiomics (DLR) methods may be the alternative approach. DLR techniques are able to learn high-dimensional features from medical images autonomously and overcome shortcomings such as the cumbersome manual coding in traditional methods [15,16]. In recent years, DLR models have been used in AD studies [17,18]. For instance, Basaia et al. used a deep neural network to classify AD and MCI based on cross-sectional structural resonance imaging (MRI) images. The classification accuracy between AD and NC was 98.2%, and the accuracy of progression from MCI to AD was 74.9% [19]. Lee et al. employed a DLR model for AD classification based on MRI images and achieved an accuracy of 95.35% [20]. Pan et al. proposed a novel convolutional neural network (CNN) architecture called a multi-view separable pyramid network (MiSePyNet) and achieved a classification accuracy of 83.05% in predicting the progression from MCI to AD [21]. Lu et al. used multiscale neural networks to identify subjects with pre-symptomatic AD and achieved an accuracy of 82.51% based on 18F-fluorodeoxyglucose positron emission tomography (FDG-PET) images [22]. The above results showed the feasibility of DLR models for diagnosing AD and MCI. However, whether DLR models could be used to analyze tau-PET images is still unknown. Therefore, in this study, we assumed that the DLR technique was also feasible for application to tau-PET images and would be useful for the diagnosis of AD and MCI. To test the above hypothesis, we employed a novel DLR model and validated it in an exploratory study.

2. Methods and Materials

Figure 1 shows the whole experimental process of this study, which includes the following six steps: (1) subject enrollment; (2) tau-PET image preprocessing, including registration, smoothing, and numerical normalization; (3) deep learning (DL) model pre-training. During this session, several classical CNN models were selected and compared, and the best one was finally selected for the next step; (4) extraction of DLR features; (5) classification; (6) comparative experiments.

2.1. Subjects

The data used in this study were obtained from the ADNI cohort, which was jointly funded by the National Institutes of Health and the National Institute on Aging in 2004. ADNI is currently the definitive data center for AD-related disease research. In order to obtain the pathogenesis of AD and find treatments, ADNI aims to study the pathogenesis of AD and discover clinical, imaging, genetic and biochemical biomarkers that can be used for the early detection of AD by collecting and organizing longitudinal data from AD patients; the database currently has more than 2000 neuroimaging data. Specific information is available on ADNI’s official website: http://adni.loni.usc.edu/about/, accessed on 12 November 2021.
In this study, a total of 211 NC subjects and 197 MCI and 117 AD patients were collected. All acquired subjects had both T1-weighted MRI images and tau-PET images. Of these, 189 NC subjects and 173 MCI and 101 AD patients were used to train and validate the DLR model. A separate 20 NC subjects and 18 MCI and 12 AD patients were used as an independent external test group. The remaining 2 NC subjects and 6 MCI and 4 AD patients were not included in the training or testing groups because images were found to be mutilated during pre-processing inspections. Demographic information (including gender, age and education) and T1-weighted MRI and tau-PET (AV 1451) images were collected for all participants. All subjects were also screened with the following neuropsychological examinations: the Clinical Dementia Rating-Sum of Boxes (CDR-SB), the MMSE, the MoCA-B, the 11-item and 13-item AD assessment cognitive scale (Alzheimer’s disease assessment scale-cognitive, ADAS) and the ADAS delayed word recall (ADASQ4) subscale. Figure 2 shows the flow chart of the data inclusion/exclusion criteria.
The inclusion criteria of MCI were according to the criteria proposed by Jak and Bondi in 2014 [5]: (1) Scores obtained in at least one cognitive domain (memory, language or speed, executive function) were below the standard deviation of the age/education corrected normative mean; (2) scores in each of the three cognitive domains of memory, language and speed/executive function were found to be impaired; (3) Scores on the Functional Activities Questionnaire (FAQ) ≥ 9. The diagnosis of AD was primarily based on guidelines provided by the National Institute on Aging (NIA) and the Alzheimer’s Association (AA) working group. The ADNI institutional review board reviewed and approved the ADNI data collection protocol [7].

2.2. Images Acquisition and Preprocessing

The image acquisition process is described on the ADNI website at http://adni.loni.usc.edu/about/, accessed on 1 June 2021. All tau-PET images were preprocessed using SPM12 software (https://www.fil.ion.ucl.auk/spm/software/spm12/, accessed on 20 September 2021.) implemented in MATLAB 2019b. The preprocessing steps were as follows.
First, the DICOM images were uniformly converted to NIFTI format (.nii) using an image conversion tool for subsequent processing. The converted images were 3D image data with spatial structure information of the brain and retained the characteristic information between tissue structures. Second, since subjects might have some head tilt problems during tau-PET image acquisition, the original correction function in SPM12 was used in this experiment reduce external differences. Furthermore, the T1 MRI images were used to align the tau-PET images so that the corresponding points at spatially uniform locations in the two types of images corresponded to each other. Smoothing and numerical normalization were performed in the next step. After completing the above processing, the images were smoothed to suppress the noise, and the numerical normalization could eliminate the differences between different instruments and reduce the number of subsequent calculations. In this experiment, the images were normalized according to the tau-PET precipitated area in the cerebellar cortex region. After the above processing, 3D image data with a size of 91 × 109 × 91 voxels in the standard space were obtained. To speed up the training time of the DLR models, all images were further normalized to −1 to 1 interval. In the unidirectional slicing condition, the 3D images were axially sliced into 91 single-channel images of size of 91 × 109 voxels, and the slices were filled and resampled to 224 × 224 voxels using linear interpolation due to the need to retain as much information as possible and to satisfy the model input conditions.

2.3. The Proposed DLR Model

The proposed DLR model is depicted in Figure 3. The model consists of the following steps: (1) DLR model pre-training. Five classical CNN networks were used for model pre-training. After comparison, we aimed to select the model with the best classification performance; (2) DLR feature extraction and fusion. Based on the pre-trained model, DLR features were extracted before the final maximum pooling layer and combined with clinical features; (3) classifiers: based on the features extracted above, support vector machine (SVM) was employed as the final classifier to obtain the classification results. The details of the model will be illustrated in the next sections.

2.3.1. DLR Model Pre-Training

In recent years, CNN models have been increasingly applied to medical imaging data and shown great potential in the classification tasks. In this study, we pre-trained five common CNN models, including AlexNet, ZF-Net, ResNet18, ResNet34, and InceptionV3 models.
(1)
AlexNet demonstrates the excellent performance of deep CNN models. ReLU is used as the activation function for its network structure, which employs interleaved pooling in CNN models [23].
(2)
ZF-Net is an improved CNN model based on AlexNet. Deconvolution is used to analyze feature behavior and then to improve classification performance [24].
(3)
The Inception models have more complex network structures and unique network characteristics in comparison with AlexNet and ZF-Net. The Inception structure is designed to use multiple convolutional or pooling operations to form a network module. Inception V3, as the classic version of the Inception series, uses convolutional decomposition and regularization to enhance the classification performance [25].
(4)
The ResNet framework introduces the residual network structure to solve the gradient disappearance or gradient explosion problem [26]. Different ResNet models, such as ResNet18, ResNet34, ResNet50 and ResNet101, are depending on the number of hidden layers. Figure 4 shows the structures of ResNet18 and ResNet34.
The whole model pre-training process was divided into two parts: forward propagation and backward propagation. Before building the model, all tau-PET images were sliced and tiled into two-dimensional images and adjusted to 224 × 224 pixels. Then, all data were labeled using unique thermal coding. In the model pre-training step, all data were passed into the network and then converged using the stochastic gradient descent (SGD) algorithm and back propagated to update the model parameters. The final output of the model pre-training process was used as the classification result.
In the model pre-training step, we set the learning rate to 1 × 10−2 and updated the model parameters using an SGD optimizer with a batch size of 8. The number of training iterations was set to 100. In addition, we performed data enhancement in the training/validation group by flipping the images horizontally and adding Gaussian noise to the input images to prevent the overfitting problem. The above experiments were performed on GPU (graphics processing unit, RTX3090 accelerated by PyCharm 3.6 (JetBrains from the Czech Republic, website: https://www.jetbrains.com/pycharm/, accessed on 17 February 2022)).

2.3.2. DLR Feature Extraction and Fusion

In contrast to traditional methods relying on manually ROI segmentation, DLR methods can automatically leverage tau-PET images to obtain high-dimensional DLR features through supervised learning. After obtaining the best DL pre-trained model, we replaced the final maximum pooling layer and fully connected layer with the SVM as the final classifier. We extracted features from the last convolutional layer of each convolutional network. These features were treated as DLR features. Then, the clinical information (gender, age, education, CDR-SB, MoCA-B, MMSE, etc.) and the DLR features were combined as the input for the SVM classifier.

2.3.3. Classifier

SVM was used as the classifier in this study. SVM is essentially a linear classifier that maximizes intervals in feature space and is a binary classification model that has been widely used with statistical and regression analysis species [27]. We used a linear kernel as the kernel function.

2.4. Comparative Experiments

To demonstrate the superiority of the proposed DLR model, we compared the DLR model with three existing models, including: (1) The clinical model. This model includes demographic information and neuropsychological cognitive assessment tests as features for classification; (2) the standard uptake value ratio (SUVR) model. We calculated SUVR values of 10 tau-PET Meta ROIs as features for classification. The ten ROIs included inferior temporal lobe, lingual gyrus, middle temporal lobe, occipital lobe, parietal lobe, hippocampus, parahippocampus, posterior cingu late gyrus, precuneus and fusiform [28]; (3) the radiomics model. The radiomics features of the above 10 tau-PET Meta ROIs were extracted as features for classification. In this experiment, we used the Radiomics Toolkit (https://github.com/mvallieres/radiomics, accessed on 17 February 2022) to extract radiomics features. The feature extraction steps included wavelet band-pass filtering, isotropic resampling, Lloyd–Max quantization, feature computation, and so on [29,30]. In the comparison experiments, each model was with 10-fold cross-validation 200 times.

2.5. Statistical Analysis

In this study, chi-square tests and nonparametric rank sum tests were introduced to compare differences in demographic characteristics between the training/validation group and the test group. We used SPSS version 25.0 software (SPSS Inc., Chicago, IL, USA) for all statistical analyses. Statistical results with p values < 0.05 were considered significantly different.

3. Results

3.1. Subject Demographics

Table 1 shows the demographic results. There were differences in MoCA-B (p = 0.042) and age (p = 0.03) in the NC group; ADAS11 (p = 0.020), ADAS13 (p = 0.034) and ADASQ4 (p = 0.044) in the MCI group and no significant differences in the AD group.
All data are expressed as mean ± standard deviation. MMSE, Mini-mental State Examination; MoCA-B, Montreal cognitive assessment-basic; CDR-SB, clinical dementia rating sum of boxes; ADAS11 and ADAS13, the 11-item and 13-item AD assessment scale cognitive; ADASQ4, the ADAS delayed word recall subscale.
For age, education, MMSE, MoCA-B, CDR-SB, ADAS11, ADAS13 and ADASQ4, a nonparametric rank sum test was performed to compare differences in demographic and clinical characteristics between the training/validation and test groups under each label, i.e., NC, MCI and AD; gender was tested by chi-square between the two groups under each label.

3.2. Pre-Training Results of Candidate DL Models

Table 2, Table 3 and Table 4 present the classification performance of the five candidate DL models. The performance evaluation metrics include accuracy, sensitivity and specificity. ResNet18 had the highest classification performance in NC vs. MCI, while ResNet34 had the highest classification performance in MCI vs. AD and NC vs. AD. Therefore, ResNet34 was selected to extract the corresponding DLR features.

3.3. Comparative Experiments

3.3.1. NC vs. MCI

Table 5 shows the classification results of the four models in NC vs. MCI. The DLR model performed the best classification performance, with an accuracy of 90.76% ± 2.15%, sensitivity of 94.17% ± 1.81% and specificity of 87.74% ± 2.54% in the test group. The remaining three models performed obviously lower than the DLR model with accuracy of 75.68% ± 2.63%, 72.02% ± 4.12% and 81.61% ± 3.23%; sensitivity of 62.32% ± 4.52%, 68.95% ± 9.22% and 83.11% ± 3.14%; and specificity of 86.67% ± 2.92%, 74.76% ± 7.11% and 80.31% ± 6.38%.
Figure 5 provides the ROC curves of the four models. The AUC (mean ± SD) for the DLR model reached 0.922 ± 0.021 and achieved the best performance among these models.

3.3.2. MCI vs. AD

Table 6 shows the classification performance of the four models in MCI vs. AD. The DLR model showed accuracy of 88.43% ± 2.32%, sensitivity of 91.25% ± 2.05% and specificity of 86.56% ± 2.86% in the test group. The remaining three models performed obviously lower than the DLR model with accuracy of 78.33% ± 4.27%, 79.68% ± 5.72% and 77.16% ± 2.95%; sensitivity of 62.67% ± 9.12%, 65.63% ± 10.97% and 88.17% ± 9.25%; and specificity of 86.67% ± 2.92%, 88.95% ± 2.99% and 68.91% ± 7.64%, respectively.
Figure 6 provided the ROC curves of these four models. The AUC (mean ± SD) for the DLR model reached 0.928 ± 0.024 and achieved the best performance among these models.

3.3.3. NC vs. AD

Table 7 shows the classification performance of the four models in MCI vs. AD. The DLR model showed an accuracy of 99.92% ± 0.51%, sensitivity of 99.78% ± 0.13%, and specificity of 99.99% ± 0.14% in the test group. The remaining three models performed obviously lower than the DLR model with accuracy of 90.66% ± 0.85%, 87.58% ± 3.63% and 96.98% ± 0.21%; sensitivity of 74.96% ± 0.59%, 74.17% ± 9.43% and 92.78% ± 3.13%; and specificity of 99.63% ± 1.33%, 95.24% ± 3.17% and 99.56% ± 2.17%.
Figure 7 provided the ROC curves of these four models. The AUC (mean ± SD) for the DLR model reached 0.996 ± 0.002 and achieved the best performance among these models.

4. Discussion

DLR has been becoming a hot topic nowadays. Because of its excellent performance in image recognition and processing, DLR models have been commonly used in computer-aided disease diagnostic fields such lesion detection, quantitative lesion diagnosis, treatment decision and prognosis expectation. In this study, we proposed a DLR model based on tau-PET images to distinguish NC, MCI and AD. In contrast with other traditional models, such as the SUVR model, the traditional radiomics model and the clinical model, the DLR model achieved the best classification results.
To date, many studies have focused on the classification among NC, MCI and AD using machine learning or DL models. For instance, Lange et al. performed a voxel-based statistical analysis using FDG-PET images and achieved an AUC of 0.728 in the classification of AD and NC [31]. Zhou et al. fused MRI and FDG-PET images and used radiomics analysis to achieve an accuracy of 0.733 in the classification of MCI and NC [32]. Shu et al. used radiomics features based on MRI images to classify MCI and AD and achieved an accuracy of 0.807 [33].
Compared with previous studies, our proposed DLR model achieved superior classification results (90.76% ± 2.15% in NC vs. MCI, 88.43% ± 2.32% in MCI vs. AD, and 99.92% ± 0.51% in NC vs. AD). The reasons may be as follows: (1) The DLR model is able to extract deeper image feature information from the pre-processed tau-PET images. As it does not require an additional ROI segmentation step, it decreases errors and biases caused by ROI segmentation; (2) the DLR model is subject to unavoidable external influences such as individual differences and different parameters of imaging acquisition. In our experiment, the DLR features and clinical information were combined together, so bias caused by individual heterogeneities may be eliminated.
Although the DLR model achieved good classification results, several limitations still exist. First, more supporting data are needed to verify the stability of our proposed DLR model. In this research, all data were obtained from the ADNI database. It is worth exploring whether our model works well with other databases. We only used the ADNI database, and the robustness of the results needs to be further verified. In the future, we plan to incorporate other ethnic group data to further verify the effectiveness of our model.
Second, we only adopted five classical deep convolutional networks to obtain the final DLR model. Although the ResNet models performed well in this classification experiment, there may be other more suitable models that can be applied. Moreover, we used whole-brain tau-PET images to train the model, and it remains to be explored whether extracting ROIs would yield better results. In addition to this, in this experiment, the 3D tau-PET images were segmented and sliced according to the axial direction. Whether better results can be obtained using 3D images and convolutional networks at the 3D level requires further validation and experiments. Finally, the model is trained on tau-PET images. Combining with other modalities, such as amyloid PET, MRI and FDG-PET images may improve the classification accuracy. In summary, the DLR model we proposed in this study provides a certain help with the clinical diagnosis and differentiation of NC, MCI and AD. Through this tau-Pet image-based DLR-assisted diagnosis of MCI, early intervention can be carried out for the MCI population, which can improve the cognitive function of patients, allow for early treatment and delay the conversion to dementia.

5. Conclusions

In this study, we developed a tau-PET-based DLR method for the subgroup diagnosis of NC, MCI and AD. This study shows that the proposed DLR method can improve the diagnostic performance of MCI and AD patients and provide the possibility of MCI-to-AD conversion prediction. In the future, the DLR method will propose practical applications for the computer-aided diagnosis of MCI and AD. We believe that more image modalities based on our proposed DLR method will be applied in the differential diagnosis of NC, MCI and AD.

Author Contributions

Y.Z. and J.Z. carried out the whole experiment together and and co-wrote the manuscript. During the experiment, Y.Z. conducted the data collecting and management. J.Z. focused on data analysis and programming the code to obtain the final results. During the writing phase, Y.Z. mainly focused on the Introduction and Discussion sections, and J.Z. focused on the Methods and Results sections. Y.C. and J.J. presented the idea and performed the conceptualization and supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Open Project Program of Nuclear Medicine and Molecular Imaging Key Laboratory of Sichuan Province (HYX21004).

Institutional Review Board Statement

The data for this article was collected from ADNI. The ADNI study was conducted according to Good Clinical Practice guidelines, the Declaration of Helsinki, US 21CFR Part 50—Protection of Human Subjects, and Part 56—Institutional Review Boards, and pursuant to state and federal HIPAA regulations. Each participating site obtained ethical approval from their Institutional Review Board before commencing subject enrolment.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. The human data was acquired from the publicly available ADNI database which meet the ethics requirements.

Data Availability Statement

The subject data used in this study were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu). Meanwhile, data supporting the findings of this study are available from the corresponding authors upon reasonable request.

Acknowledgments

Data sources and collection for this paper were supported by the Alzheimer’s Disease Neuroimaging Initiative (ADNI; National Institutes of Health Grant U01 AG024904) and the DODADNI (Department of Defense, award number W81XWH-12–2-0012). ADNI is funded by the National Institute of Aging and the National Institute of Biomedical Imaging and Bioengineering and through generous contributions from the following: AbbVie, Alzheimer’s Association; Alzheimer’s Drug Discovery Foundation; Araclon Biotech; BioClinica Inc; Biogen; Bristol-Myers Squibb Company; CereSpir Inc.; Eisai Inc.; Elan Pharmaceuticals Inc.; Eli Lilly and Company; EuroImmun; F.Hoffmann-La Roche Ltd. and its affiliated company Genentech Inc; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research & Development, LLC.; Johnson & Johnson Pharmaceutical Research & Development LLC.; Lumosity; Lundbeck; Merck & Co., Inc.; Meso Scale Diagnostics, LLC; NeuroRx Research; Neu-rotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI clinical sites in Canada.

Conflicts of Interest

The authors declare that this study was conducted without any commercial or financial relationship that could be considered a potential conflict of interest.

References

  1. Alzheimer’s Association. 2017 Alzheimer’s disease facts and figures. Alzheimers Dement. 2017, 13, 325–373. [Google Scholar] [CrossRef]
  2. Morris, J.C.; Storandt, M.; Miller, J.P.; McKeel, D.W.; Price, J.L.; Rubin, E.H.; Berg, L. Mild Cognitive Impairment Represents Early-Stage Alzheimer Disease. Arch. Neurol. 2001, 58, 397–405. [Google Scholar] [CrossRef]
  3. Petersen, R.C. Mild cognitive impairment as a diagnostic entity. J. Intern. Med. 2004, 256, 183–194. [Google Scholar] [CrossRef]
  4. Winblad, B.; Palmer, K.; Kivipelto, M.; Jelic, V.; Fratiglioni, L.; Wahlund, L.-O.; Nordberg, A.; Backman, L.J.; Albert, M.S.; Almkvist, O.; et al. Mild cognitive impairment—Beyond controversies, towards a consensus: Report of the International Working Group on Mild Cognitive Impairment. J. Intern. Med. 2004, 256, 240–246. [Google Scholar] [CrossRef] [PubMed]
  5. Bondi, M.W.; Edmonds, E.C.; Jak, A.J.; Clark, L.R.; Delano-Wood, L.; McDonald, C.R.; Nation, D.A.; Libon, D.J.; Au, R.; Galasko, D.; et al. Neuropsychological Criteria for Mild Cognitive Impairment Improves Diagnostic Precision, Biomarker Associations, and Progression Rates. J. Alzheimer’s Dis. 2014, 42, 275–289. [Google Scholar] [CrossRef]
  6. Albert, M.S.; DeKosky, S.T.; Dickson, D.; Dubois, B.; Feldman, H.H.; Fox, N.C.; Gamst, A.; Holtzman, D.M.; Jagust, W.J.; Petersen, R.C.; et al. The diagnosis of mild cognitive impairment due to Alzheimer’s disease: Recommendations from the National Institute on Aging-Alzheimer’s Association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimers Dement. 2011, 7, 270–279. [Google Scholar] [CrossRef] [PubMed]
  7. McKhann, G.M.; Knopman, D.S.; Chertkow, H.; Hyman, B.T.; Jack, C.R., Jr.; Kawas, C.H.; Klunk, W.E.; Koroshetz, W.J.; Manly, J.J.; Mayeux, R.; et al. The diagnosis of dementia due to Alzheimer’s disease: Recommendations from the National Institute on Aging-Alzheimer’s association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimers Dement. J. Alzheimers Assoc. 2011, 7, 263–269. [Google Scholar] [CrossRef]
  8. Campese, N.; Palermo, G.; Del Gamba, C.; Beatino, M.F.; Galgani, A.; Belli, E.; Del Prete, E.; Della Vecchia, A.; Vergallo, A.; Siciliano, G.; et al. Progress regarding the context-of-use of tau as biomarker of Alzheimer’s disease and other neurodegenerative diseases. Expert Rev. Proteom. 2021, 18, 27–48. [Google Scholar] [CrossRef]
  9. Jack, C.R., Jr.; Bennett, D.A.; Blennow, K.; Carrillo, M.C.; Dunn, B.; Haeberlein, S.B.; Holtzman, D.M.; Jagust, W.; Jessen, F.; Karlawish, J.; et al. NIA-AA Research Framework: Toward a biological definition of Alzheimer’s disease. Alzheimer Dement. 2018, 14, 535–562. [Google Scholar] [CrossRef]
  10. Lloret, A.; Esteve, D.; Cervera-Ferri, A.; Lopez, B.; Nepomuceno, M.; Monllor, P. When Does Alzheimer′s Disease Really Start? The Role of Biomarkers. Int. J. Mol. Sci. 2019, 20, 5536. [Google Scholar] [CrossRef]
  11. Johnson, K.A.; Schultz, A.; Betensky, R.A.; Becker, J.A.; Sepulcre, J.; Rentz, D.M.; Mormino, E.C.; Chhatwal, J.; Amariglio, R.; Papp, K.; et al. Tau positron emission tomographic imaging in aging and early Alzheimer disease. Ann. Neurol. 2015, 79, 110–119. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, Q.; Liu, M.; Ha, L.; Zhou, Y.; Initiative, A.D.N.; Weiner, M.W.; Aisen, P.; Petersen, R.; Jack, C.R.J.; Jagust, W.; et al. Quantitative 18F-AV1451 Brain Tau PET Imaging in Cognitively Normal Older Adults, Mild Cognitive Impairment, and Alzheimer’s Disease Patients. Front. Neurol. 2019, 10, 486. [Google Scholar] [CrossRef] [PubMed]
  13. La Joie, R.; Bejanin, A.; Fagan, A.M.; Ayakta, N.; Baker, S.L.; Bourakova, V.; Boxer, A.L.; Cha, J.; Karydas, A.; Jerome, G.; et al. Associations between [18F]AV1451 tau PET and CSF measures of tau pathology in a clinical sample. Neurology 2017, 90, e282–e290. [Google Scholar] [CrossRef] [PubMed]
  14. Sun, H.; Wang, A.; Ai, Q.; Wang, Y. A New-Style Random Forest Diagnosis Model for Alzheimer’s Disease. J. Med. Imaging Health Inform. 2020, 10, 705–709. [Google Scholar] [CrossRef]
  15. Liu, Z.; Wang, S.; Dong, D.; Wei, J.; Fang, C.; Zhou, X.; Sun, K.; Li, L.; Li, B.; Wang, M.; et al. The Applications of Radiomics in Precision Diagnosis and Treatment of Oncology: Opportunities and Challenges. Theranostics 2019, 9, 1303–1322. [Google Scholar] [CrossRef] [PubMed]
  16. Suzuki, K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef] [PubMed]
  17. Li, Y.; Wei, D.; Liu, X.; Fan, X.; Wang, K.; Li, S.; Zhang, Z.; Ma, K.; Qian, T.; Jiang, T.; et al. Molecular subtyping of diffuse gliomas using magnetic resonance imaging: Comparison and correlation between radiomics and deep learning. Eur. Radiol. 2021, 32, 747–758. [Google Scholar] [CrossRef]
  18. Park, J.E.; Kickingereder, P.; Kim, H.S. Radiomics and Deep Learning from Research to Clinical Workflow: Neuro-Oncologic Imaging. Korean J. Radiol. 2020, 21, 1126. [Google Scholar] [CrossRef] [PubMed]
  19. Basaia, S.; Agosta, F.; Wagner, L.; Canu, E.; Magnani, G.; Santangelo, R.; Filippi, M. Automated classification of Alzheimer’s disease and mild cognitive impairment using a single MRI and deep neural networks. NeuroImage Clin. 2018, 21, 101645. [Google Scholar] [CrossRef] [PubMed]
  20. Lee, B.; Ellahi, W.; Choi, J.Y. Using Deep CNN with Data Permutation Scheme for Classification of Alzheimer’s Disease in Structural Magnetic Resonance Imaging (sMRI). IEICE Trans. Inf. Syst. 2019, E102D, 1384–1395. [Google Scholar] [CrossRef]
  21. Pan, X.; Phan, T.-L.; Adel, M.; Fossati, C.; Gaidon, T.; Wojak, J.; Guedj, E. Multi-View Separable Pyramid Network for AD Prediction at MCI Stage by 18F-FDG Brain PET Imaging. IEEE Trans. Med. Imaging 2020, 40, 81–92. [Google Scholar] [CrossRef]
  22. Lu, D.; Popuri, K.; Ding, G.W.; Balachandar, R.; Beg, M.F. Multiscale deep neural network based analysis of FDG-PET images for the early diagnosis of Alzheimer’s disease. Med. Image Anal. 2018, 46, 26–34. [Google Scholar] [CrossRef] [PubMed]
  23. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. NIPS 2012, 60, 84–90. [Google Scholar] [CrossRef]
  24. Zeiler, M.; Fergus, R. Visualizing and Understanding Convolutional Networks; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8689, pp. 818–833. [Google Scholar]
  25. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Amari, S.; Wu, S. Improving support vector machine classifiers by modifying kernel functions. Neural Netw. 1999, 12, 783–789. [Google Scholar] [CrossRef]
  28. Chotipanich, C.; Nivorn, M.; Kunawudhi, A.; Promteangtrong, C.; Boonkawin, N.; Jantarato, A. Evaluation of Imaging Windows for Tau PET Imaging Using 18F-PI2620 in Cognitively Normal Individuals, Mild Cognitive Impairment, and Alzheimer’s Disease Patients. Mol. Imaging 2020, 19, 1–8. [Google Scholar] [CrossRef] [PubMed]
  29. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images Are More than Pictures, They Are Data. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef]
  30. Kumar, V.; Gu, Y.; Basu, S.; Berglund, A.; Eschrich, S.A.; Schabath, M.B.; Forster, K.; Aerts, H.J.W.L.; Dekker, A.; Fenstermacher, D.; et al. Radiomics: The process and the challenges. Magn. Reson. Imaging 2012, 30, 1234–1248. [Google Scholar] [CrossRef]
  31. Lange, C.; Suppa, P.; Frings, L.; Brenner, W.; Spies, L.; Buchert, R. Optimization of Statistical Single Subject Analysis of Brain FDG PET for the Prognosis of Mild Cognitive Impairment-to-Alzheimer’s Disease Conversion. J. Alzheimer’s Dis. 2016, 49, 945–959. [Google Scholar] [CrossRef]
  32. Zhou, H.; Jiang, J.; Lu, J.; Wang, M.; Zhang, H.; Zuo, C.; Initiative, A.D.N. Dual-Model Radiomic Biomarkers Predict Development of Mild Cognitive Impairment Progression to Alzheimer’s Disease. Front. Neurosci. 2019, 12, 1045. [Google Scholar] [CrossRef]
  33. Shu, Z.-Y.; Mao, D.-W.; Xu, Y.-Y.; Shao, Y.; Pang, P.-P.; Gong, X.-Y. Prediction of the progression from mild cognitive impairment to Alzheimer’s disease using a radiomics-integrated model. Ther. Adv. Neurol. Disord. 2021, 14, 1–13. [Google Scholar] [CrossRef]
Figure 1. The whole experimental process in this study.
Figure 1. The whole experimental process in this study.
Brainsci 12 01067 g001
Figure 2. The flow chart of the data inclusion/exclusion criteria.
Figure 2. The flow chart of the data inclusion/exclusion criteria.
Brainsci 12 01067 g002
Figure 3. The framework of the proposed DLR model.
Figure 3. The framework of the proposed DLR model.
Brainsci 12 01067 g003
Figure 4. The fundamental structure of ResNet18 and ResNet34. “7 × 7” and “3 × 3” indicate the size of the convolution kernel, “conv” indicates convolution, “Avg Pool” indicates average pooling, and “FC” indicates fully connected layer. “64”, “128”, “128”, “256” and “512” represent the numbers of channels, and “/2” means stride of 2.
Figure 4. The fundamental structure of ResNet18 and ResNet34. “7 × 7” and “3 × 3” indicate the size of the convolution kernel, “conv” indicates convolution, “Avg Pool” indicates average pooling, and “FC” indicates fully connected layer. “64”, “128”, “128”, “256” and “512” represent the numbers of channels, and “/2” means stride of 2.
Brainsci 12 01067 g004
Figure 5. ROC curves for the four models in NC vs. MCI.
Figure 5. ROC curves for the four models in NC vs. MCI.
Brainsci 12 01067 g005
Figure 6. ROC curves for the four models in MCI vs. AD.
Figure 6. ROC curves for the four models in MCI vs. AD.
Brainsci 12 01067 g006
Figure 7. ROC curves for the four models in NC vs. AD.
Figure 7. ROC curves for the four models in NC vs. AD.
Brainsci 12 01067 g007
Table 1. Demographic information in this study.
Table 1. Demographic information in this study.
NC GroupsMCI GroupsAD Groups
NC1 (Train)NC2 (Test)MCI1 (Train)MCI2 (Test)AD1 (Train)AD2 (Test)
Gender (M/F)69/12110/1195/8313/662/435/7
Age (year)73.17 ± 7.6476.76 ± 6.85 a73.88 ± 7.4670.93 ± 8.2175.39 ± 7.9476.76 ± 9.38
Education16.66 ± 2.3417.10 ± 2.0416.39 ± 2.5616.05 ± 2.3715.49 ± 2.5915.33 ± 2.61
MMSE29.11 ± 1.2329.14 ± 1.0627.78 ± 2.2127.37 ± 2.3121.36 ± 4.9821.58 ± 3.55
MoCA-B26.36 ± 2.5525.00 ± 2.73 a23.24 ± 3.5423.68 ± 3.1616.15 ± 5.0316.20 ± 4.64
CDR-SB0.06 ± 0.230.10 ± 0.201.45 ± 1.032.37 ± 1.885.92 ± 3.325.13 ± 2.22
ADAS118.69 ± 2.578.93 ± 1.8612.59 ± 4.1114.25 ± 3.08 a22.10 ± 7.2925.03 ± 5.98
ADAS1312.46 ± 4.1213.25 ± 2.9919.01 ± 6.1121.51 ± 5.10 a32.57 ± 8.6436.11 ± 7.20
ADASQ42.51 ± 1.723.25 ± 2.004.77 ± 2.265.89 ± 2.35 a8.06 ± 1.328.25 ± 1.76
a indicated that the p value was less than 0.05 in comparison results between the training/validation and test groups under the same label.
Table 2. Classification performance in NC vs. MCI.
Table 2. Classification performance in NC vs. MCI.
ModelAccuracy (%)Sensitivity (%)Specificity (%)
Training/Validation Groups
AlexNet94.26 ± 2.6093.80 ± 2.9294.70 ± 4.34
ZF-Net94.28 ± 3.9994.97 ± 3.6493.66 ± 7.46
ResNet1895.78 ± 2.5094.99 ± 4.7096.51 ± 2.98
ResNet3495.32 ± 2.6294.06 ± 3.7496.49 ± 4.55
InceptionV393.82 ± 3.9493.02 ± 4.9394.54 ± 6.39
Test Group
AlexNet81.25 ± 3.067947 ± 2.8682.86 ± 3.83
ZF-Net83.14 ± 3.2478.37 ± 3.8986.67 ± 5.95
ResNet1887.25 ± 2.2187.37 ± 2.2487.14 ± 2.32
ResNet3487.00 ± 2.1485.79 ± 2.1288.10 ± 2.52
InceptionV380.50 ± 3.5877.89 ± 4.2482.86 ± 5.79
The bold means this model performed best among others.
Table 3. Classification performance in MCI vs. AD.
Table 3. Classification performance in MCI vs. AD.
ModelAccuracy (%)Sensitivity (%)Specificity (%)
Training/Validation Groups
AlexNet93.18 ± 4.3689.33 ± 10.5595.41 ± 3.59
ZF-Net93.55 ± 5.1991.37 ± 7.2094.77 ± 5.16
ResNet1893.72 ± 3.4090.47 ± 8.1695.63 ± 4.22
ResNet3495.28 ± 2.5094.76 ± 4.9695.59 ± 3.03
InceptionV397.45 ± 2.7895.26 ± 6.7798.75 ± 2.64
Test Group
AlexNet79.68 ± 5.1264.17 ± 7.3289.47 ± 4.81
ZF-Net79.68 ± 2.4062.50 ± 3.7890.52 ± 2.34
ResNet1882.26 ± 1.7873.33 ± 2.7287.89 ± 2.14
ResNet3482.26 ± 1.5477.50 ± 2.4885.26 ± 2.12
InceptionV379.68 ± 2.1474.17 ± 3.3283.16 ± 3.48
The bold means this model performed best among others.
Table 4. Classification performance in NC vs. AD.
Table 4. Classification performance in NC vs. AD.
ModelAccuracy (%)Sensitivity (%)Specificity (%)
Training/Validation Groups
AlexNet97.36 ± 2.9895.67 ± 7.2598.27 ± 2.44
ZF-Net98.30 ± 2.4297.89 ± 5.0998.53 ± 2.08
ResNet1897.17 ± 2.0596.87 ± 4.4297.32 ± 2.60
ResNet3498.10 ± 1.8196.14 ± 5.8799.14 ± 1.38
InceptionV394.37 ± 3.5391.29 ± 8.6696.16 ± 3.13
Test Group
AlexNet94.24 ± 0.9684.17 ± 2.63100.0 ± 0.00
ZF-Net93.64 ± 2.6582.57 ± 7.34100.0 ± 0.00
ResNet1896.97 ± 2.9191.70 ± 3.50100.0 ± 0.00
ResNet3496.97 ± 2.1691.70 ± 2.83100.0 ± 0.00
InceptionV395.08 ± 3.1489.58 ± 5.3098.21 ± 0.96
The bold means this model performed best among others.
Table 5. The classification performance in NC vs. MCI.
Table 5. The classification performance in NC vs. MCI.
ModelAccuracy (%)Sensitivity (%)Specificity (%)
Training/Validation Groups
SUVR model69.36 ± 7.9463.53 ± 12.0374.73 ± 10.74
Traditional radiomics model69.05 ± 7.2263.34 ± 13.3474.28 ± 7.07
Clinical model74.84 ± 8.4880.13 ± 14.1671.58 ± 12.05
DLR model98.46 ± 1.7198.47 ± 1.6698.44 ± 1.76
Test Group
SUVR model75.68 ± 2.6362.32 ± 4.5286.67 ± 2.92
Traditional radiomics model72.02 ± 4.1268.95 ± 9.2274.76 ± 7.11
Clinical model81.61 ± 3.2383.11 ± 3.1480.31 ± 6.38
DLR model90.76 ± 2.1594.74 ± 1. 8187.74 ± 2.54
The bold means this model performed best among others.
Table 6. The classification performance in MCI vs. AD.
Table 6. The classification performance in MCI vs. AD.
ModelAccuracy (%)Sensitivity (%)Specificity (%)
Training/Validation Groups
SUVR model74.41 ± 8.1555.39 ± 15.6486.06 ± 8.88
Traditional radiomics model70.20 ± 7.83 57.79 ± 13.6481.58 ± 9.83
Clinical model90.84 ± 4.9584.59 ± 11.0494.45 ± 5.00
DLR model96.27 ± 1.1694.90 ± 4.8197.89 ± 2.11
Test Group
SUVR model78.33 ± 4.2762.67 ± 9.1286.67 ± 2.92
Traditional radiomics model79.68 ± 5.7265.63 ± 10.9788.95 ± 2.99
Clinical model77.16 ± 2.9588.17 ± 9.2568.91 ± 7.64
DLR model88.43 ± 2.3291.25 ± 2.0586.56 ± 2.86
The bold means this model performed best among others.
Table 7. The classification performance in NC vs. AD.
Table 7. The classification performance in NC vs. AD.
ModelAccuracy (%)Sensitivity (%)Specificity (%)
Training/Validation Groups
SUVR model86.06 ± 6.1873.43 ± 14.2193.04 ± 6.06
Traditional radiomics model78.65 ± 0.0857.67 ± 16.6487.06 ± 8.80
Clinical model91.96 ± 5.4499.06 ± 2.9886.65 ± 8.04
DLR model99.31 ± 1.5098.00 ±4.43100.0 ± 0.00
Test Group
SUVR model90.66 ± 0.8574.96 ± 0.5999.63 ± 1.33
Traditional radiomics model85.58 ± 3.6374.17 ± 9.4395.24 ± 3.17
Clinical model96.98 ± 0.2192.78 ± 3.1399.56 ± 2.17
DLR model99.92 ± 0.5199.78 ± 0. 1399.99 ± 0.14
The bold means this model performed best among others.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, Y.; Zhang, J.; Chen, Y.; Jiang, J. A Novel Deep Learning Radiomics Model to Discriminate AD, MCI and NC: An Exploratory Study Based on Tau PET Scans from ADNI. Brain Sci. 2022, 12, 1067. https://doi.org/10.3390/brainsci12081067

AMA Style

Zhao Y, Zhang J, Chen Y, Jiang J. A Novel Deep Learning Radiomics Model to Discriminate AD, MCI and NC: An Exploratory Study Based on Tau PET Scans from ADNI. Brain Sciences. 2022; 12(8):1067. https://doi.org/10.3390/brainsci12081067

Chicago/Turabian Style

Zhao, Yan, Jieming Zhang, Yue Chen, and Jiehui Jiang. 2022. "A Novel Deep Learning Radiomics Model to Discriminate AD, MCI and NC: An Exploratory Study Based on Tau PET Scans from ADNI" Brain Sciences 12, no. 8: 1067. https://doi.org/10.3390/brainsci12081067

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop