Next Article in Journal
Role of Obesity in Inflammation and Remodeling of Asthmatic Airway
Next Article in Special Issue
PathoLive—Real-Time Pathogen Identification from Metagenomic Illumina Datasets
Previous Article in Journal
Quality of Multicenter Studies Using MRI Radiomics for Diagnosing Clinically Significant Prostate Cancer: A Systematic Review
Previous Article in Special Issue
Recommendations of scRNA-seq Differential Gene Expression Analysis Based on Comprehensive Benchmarking
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Reproducible Deep-Learning-Based Computer-Aided Diagnosis Tool for Frontotemporal Dementia Using MONAI and Clinica Frameworks

by
Andrea Termine
1,†,
Carlo Fabrizio
1,†,
Carlo Caltagirone
2,
Laura Petrosini
3,* and
on behalf of the Frontotemporal Lobar Degeneration Neuroimaging Initiative
1
Data Science Unit, IRCCS Santa Lucia Foundation, 00143 Rome, Italy
2
Department of Clinical and Behavioral Neurology, IRCCS Santa Lucia Foundation, 00179 Rome, Italy
3
Experimental and Behavioral Neurophysiology, IRCCS Santa Lucia Foundation, 00143 Rome, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
The FTLDNI investigators are listed below in the Acknowledgments.
Life 2022, 12(7), 947; https://doi.org/10.3390/life12070947
Submission received: 25 May 2022 / Revised: 16 June 2022 / Accepted: 21 June 2022 / Published: 23 June 2022
(This article belongs to the Special Issue Computational Analysis of Biomedical Data)

Abstract

:
Despite Artificial Intelligence (AI) being a leading technology in biomedical research, real-life implementation of AI-based Computer-Aided Diagnosis (CAD) tools into the clinical setting is still remote due to unstandardized practices during development. However, few or no attempts have been made to propose a reproducible CAD development workflow for 3D MRI data. In this paper, we present the development of an easily reproducible and reliable CAD tool using the Clinica and MONAI frameworks that were developed to introduce standardized practices in medical imaging. A Deep Learning (DL) algorithm was trained to detect frontotemporal dementia (FTD) on data from the NIFD database to ensure reproducibility. The DL model yielded 0.80 accuracy (95% confidence intervals: 0.64, 0.91), 1 sensitivity, 0.6 specificity, 0.83 F1-score, and 0.86 AUC, achieving a comparable performance with other FTD classification approaches. Explainable AI methods were applied to understand AI behavior and to identify regions of the images where the DL model misbehaves. Attention maps highlighted that its decision was driven by hallmarking brain areas for FTD and helped us to understand how to improve FTD detection. The proposed standardized methodology could be useful for benchmark comparison in FTD classification. AI-based CAD tools should be developed with the goal of standardizing pipelines, as varying pre-processing and training methods, along with the absence of model behavior explanations, negatively impact regulators’ attitudes towards CAD. The adoption of common best practices for neuroimaging data analysis is a step toward fast evaluation of efficacy and safety of CAD and may accelerate the adoption of AI products in the healthcare system.

1. Introduction

Computer-Aided Diagnosis (CAD) tools aim to help improve physicians’ performance in disease detection, with the main objective of detecting early pathological signs that humans may fail to find [1]. CAD applications have been developed for numerous medical imaging modalities and diseases [1,2] and they provide diagnosis probabilities by analyzing the data. As such, CAD tools would help the radiologists to draw their conclusions, supporting their interpretation and decision-making processes [3]. Interestingly, it has been shown that the performance of the radiologist can be equalized by CADs and even improved by joining human collective reasoning with machine predictions [2,4].
CAD on medical imaging data relies on image processing methods and Artificial Intelligence (AI) classification systems. Deep Learning (DL) is the best suited AI technique for this purpose, strong in its role in computer vision. Highly difficult tasks requiring human or superhuman ability have been solved by DL over the last few years, defining new standards in protein structure prediction [5], image generation from natural language description [6], and general-purpose learning in complex domains [7]. Applications of DL methods in the medical field mainly include systems for an earlier or more accurate disease diagnosis, assessment for the risk of conversion to a more severe disease status and disease subtypes identification [8]. DL is based on Artificial Neural Networks that analyze input data by mimicking brain functioning, with several layers of nodes as neurons applying complex transformation functions to data [9]. In the case of a CAD tool, the information is processed and the model produces an output response, such as the prediction of a class probability (e.g., case or control). Implementation and investment in DL for medicine have been growing over the past few years [10]. This is due to its potential of providing new reliable methods to enhance healthcare practices and finally foster precision medicine.
Nonetheless, the acceptance of AI in standard clinical settings is still lagging [11]. One of the factors hindering AI-based CAD spread in hospitals is that DL models have a lack of interpretability, which is a primary concern for medical practitioners. Not knowing how the model made its choice weakens AI trustability. DL models are considered uninterpretable due to the complexity of the transformation they apply when processing the data, and for this reason they are often called black-box models. Recently, the development of explainable AI methods has been tackling this issue. In fact, explainable AI aims to enhance its interpretability by providing insights about models’ behavior [12]. Attention maps are visual tools that help explain deep convolutional neural networks, showing which input regions are the most influential for the network when making a prediction. Producing and interpreting attention maps when developing DL-based CAD tools could strongly enhance their acceptance and finally their diffusion.
Frontotemporal dementia (FTD) is a neurodegenerative clinical syndrome where behavior, executive functions, and language show progressive deficits [13]. FTD ranks third in the prevalence of dementia, after Alzheimer’s disease (AD) and dementia with Lewy bodies [14]. FTD encloses clinical syndromes whose histopathological characteristics are the neuronal loss, gliosis, and progressive neurodegeneration predominantly affecting the frontal and temporal lobes [15,16,17]. Diagnosing FTD is a difficult task and requires a longer period of time compared to AD. This is due to a subtle and insidious onset for FTD, as memory problems, typically the first sign of dementia, are often lacking and the majority of FTD patients have no initial complaints [18,19]. Moreover, FTD is highly heterogeneous in its manifestations and multiple disease variants have been identified, for which international diagnostic criteria have been proposed [20,21,22]. Differential diagnosis of FTD with other forms of dementia such as AD is even harder, and investigations of how long it takes to make a diagnosis of FTD have shown that a comprehensive assessment is essential to differentiate FTD from other diagnoses [23,24]. Even if there is currently no disease-modifying treatment for FTD, its early diagnosis has the potential to improve patient management by timely planning useful pharmacological treatment strategies for symptom control [25]. These are useful in helping caregivers cope with the high levels of distress they are experiencing due to the high prevalence of psychopathology in FTD patients [26]. Understanding how FTD leads to cognitive and behavioral symptoms is critical, and non-invasive brain stimulation methods such as Transcranial Magnetic Stimulation (TMS) and transcranial Direct Current Stimulation (tDCS) have been used to shed light on the mechanisms involved [27,28,29].
Brain imaging in vivo has been crucial to unraveling FTD pathophysiology, as it made it possible to identify several biomarkers associated with the disease. Neuromorphological hallmarks of FTD are gray matter volume loss of the prefrontal cortex, insula and anterior cingulate cortex [30,31]. Additionally, there is a white matter integrity loss over time, and brain atrophy increases with disease progression [32,33]. Magnetic resonance measurements of ventricular volumetric changes in FTD patients showed that their size increases with time, and this correlates with cognitive impairment severity [34]. In particular, ventricular expansion has been found to represent an informative marker to discriminate the behavioral variant of FTD from other FTD variants and from other forms of dementia [35].
The development of a reliable CAD tool needs the data to be carefully annotated, organized, and managed, especially when it is based on Magnetic Resonance Imaging (MRI). Classification systems exploiting brain imaging have been successfully used to capture structural changes in the human brain [36,37] and detect dementia up to flawless performance [38]. Extensive datasets improve the solidity of AI algorithms’ training and many data-sharing initiatives have grown in the field of neurodegenerative disease research in the last 20 years [39]. Such initiatives foster neurodegenerative disease research, by putting aside the need for years-long data collection and providing reliable data. This speeds up both hypothesis testing and data-driven research that exploits AI for data analysis. A key element that enables the use of such techniques is the adoption of academy and industry-wide data standards. In order to make the shared data productive, the FAIR (Findable, Accessible, Interoperable, Reusable) principles have been proposed to promote Open Science practices for data sharing initiatives [40,41]. FAIR data can be highly precious, especially when hundreds of subjects are available and the methods for data acquisition are standardized and reliable. The spread of research-useful data is aided by database management initiatives such as the Image and Data Archive (IDA), hosted by the Laboratory of NeuroImaging (LONI), that contains data from more than 160 studies. IDA shares the Neuroimaging in Frontotemporal Dementia (NIFD) database, which is one the biggest data sharing initiatives on FTD to date.
To date, few attempts have been made to discriminate FTD from Normal Controls (NC) using AI methods [36,37,42,43,44,45,46,47]. Most of the studies used MRI-derived numerical features, such as gray/white matter volume or cortical thickness quantification, with traditional ML algorithms such as Support Vector Machine (SVM) and logistic regression. To the best of our knowledge, there is only one published study that used 3D MRI data for FTD classification with a pretrained Convolutional Neural Network (CNN) achieving high accuracy [42]. Given that only a few of the previous classification attempts used NIFD data, sample sizes usually ranged from 12 to ~450 subjects, and preprocessing and data augmentation methods, as well as AI algorithms choice and cross-validation strategies, were highly heterogeneous. This heterogeneity in methodology resulted in unstable and potentially unreliable CAD performances, achieving accuracies ranging from 0.66 to 1.
To deepen our understanding of neurodegenerative diseases in general, and FTD in particular, standardized practices are needed when preprocessing and analyzing MRI data with DL, in order to make studies easier and more reproducible. Several initiatives have approached this problem. The Nipype (Neuroimaging in Python—Pipelines and Interfaces) package, an open-source Python project, provides functions where the interaction with neuroimaging software tools or algorithms can be performed in a single workflow, partly handling the heterogeneity of image processing [48]. NiftyTorch is a DL framework for neuroimaging in Python, giving users an interface for PyTorch modeling with the aim of providing a package to easily perform AI-based operations on neuroimaging data [49]. In the same fashion, MedicalTorch is an open-source framework for PyTorch, implementing an extensive set of loaders, pre-processors and datasets for medical imaging [50]. In 2016, efforts to increase data standardization and experiment reproducibility led to the proposal to the neuroimaging community of the Brain Imaging Data Structure (BIDS) as an organization format for clinical and imaging data [51]. Despite the scientific community’s acceptance of the BIDS standard, not all public neuroimaging datasets provide BIDS versions of their data.
Clinica, a set of automatic pipelines for the management of multimodal neuroimaging data, pursues the community effort of reproducibility and aims to make clinical research studies easier [52]. This project is developed and maintained from the Aramis Lab and consists of a set of automatic pipelines for neuroimaging data processing and analysis. The aim of Clinica is to make clinical neuroimaging studies easier and reproducible by providing standardized methods for data processing. Clinica includes converters of public neuroimaging datasets to BIDS, along with processing pipelines, and organization for processed files, statistical analyses, and Machine Learning algorithms.
One of the latest and most extensive initiatives for the standardization of AI applications to medical imaging is “project MONAI”, that originally started in 2020 by NVIDIA and King’s College London and brought to the MONAI framework [53]. It is an open-source PyTorch-based framework for DL in healthcare imaging. The aim of the project is developing and sharing best practices for AI in healthcare imaging, creating state-of-the-art, end-to-end training workflows for healthcare imaging and providing researchers with the optimized and standardized way to create and evaluate Deep Learning models. The MONAI framework provides workflows for using domain optimized networks, loss functions, metrics, and optimizers.
Considering the high heterogeneity in CAD tools development methodology, there is a need to foster standardized practices to spread the benefits of AI strategies adoption in the diagnosis pipeline. Here, we present a proof-of-concept of Clinica and MONAI application on the NIFD database to train and test a CAD tool on FTD data with the application of shared practices. Our aim was testing the efficacy of only using standardized frameworks for neuroimaging data preprocessing and DL modeling on medical imaging as main steps in CAD development. The achieved performance is comparable to that of other FTD classification systems, showing the appropriateness of this methodology. Explainable AI methods reveal that the model mimics human behavior when making its decision, mainly relying on morphological changes in hallmarking brain areas for FTD. The adoption of academy and industry-wide data standards coupled with standardized practices for neuroimaging data management will provide more reliable results upon the application of less biased procedures.

2. Materials and Methods

2.1. Neuroimaging in Frontotemporal Dementia Database

This study was performed on data from the NIFD database, hosted by the Laboratory of NeuroImaging from the University of Southern California. NIFD is the nickname for the frontotemporal lobar degeneration neuroimaging initiative (FTLDNI). FTLDNI was funded through the National Institute of Aging and started in 2010. The primary goals of FTLDNI were to identify neuroimaging modalities and methods of analysis for tracking frontotemporal lobar degeneration (FTLD) and to assess the value of imaging versus other biomarkers in diagnostic roles. The Principal Investigator of NIFD was Dr. Howard Rosen, MD, at the University of California, San Francisco. The data are the result of collaborative efforts at three sites in North America. For up-to-date information on participation and protocol, please visit http://memory.ucsf.edu/research/studies/nifd (accessed on 25 May 2022). NIFD includes data from 346 subjects followed over time including FTD and NC all with a careful assessment through interviews, physical examinations, cognitive testing and blood and/or CSF testing, along with brain MRI and PET acquisition. FTD patients included in this database are diagnosed with one of the following disease variants: behavioral variant, semantic variant, progressive non-fluent aphasia, progressive supranuclear palsy, or cortico-basal syndrome.

2.2. Clinica: An Open-Source Software Platform for Reproducible Clinical Neuroscience Studies

We used Clinica [52] functions to manage NIFD data. In particular, we applied the nifd-to-bids function to convert NIFD data to BIDS format, in order to be ready for processing. Next, we used the t1-linear pipeline to affinely align T1-weighted MR images to the MNI space. With this standardized preprocessing applied, we were ready for model training.

2.3. MONAI: Medical Open Network for Artificial Intelligence

We followed the MONAI workflow for 3D classification based on DenseNet. MONAI proposes to use the DenseNet121 architecture stored in PyTorch, a model from [54]. MONAI made importing and transforming the images easy, and provided useful functions to train and test the DL model. Such standardized and community-based practice makes data and model management more solid, increasing the robustness and reproducibility of CAD building practice.

2.4. Workflow Overview

The next sections describe the methodology for the proposed workflow, which is shown in Figure 1. NIFD data was downloaded from the IDA and filtered as specified in Section 2.5. Data preprocessing was performed with the standardized pipeline from Clinica, as specified in Section 2.6. Data augmentation was applied only on the train set and preceded model training and test phases, which were all performed in the MONAI framework. These two are described in Section 2.7 and Section 2.8, respectively. Finally, the behavior of resulting CAD is evaluated by extracting attention maps as explained in Section 2.9.

2.5. Data Collection

To perform the present study, we filtered the data to include only 3D T1-weighted Magnetization-Prepared Rapid Acquisition with Gradient Echo (MPRAGE) MRI scans at the first visit for cases and controls, as this was the most frequent acquisition modality among many others available in NIFD. Subjects were labeled as cases or controls following the diagnosis reported in the clinical data table. When the diagnosis was missing in the clinical data table, the “patient/control” label available in the MRI metadata was used. This resulted in a final dataset of 182 FTD and 130 NC.

2.6. Preprocessing Pipeline

Subjects were randomly assigned to train, validation, and test sets with a 70/30 proportion calculated on the group with the highest n (Table 1). This splitting proportion is most indicated for small sample sizes and it required ~40 subjects for validation and test [55]. Random sampling without replacement resulted in each participant being uniquely assigned to one of the three sets, avoiding data leakage.
The 3D T1 MPRAGE MRI scans underwent BIDS formatting with the application of the nifd-to-bids converter from Clinica [52]. Data preprocessing for the affine registration of T1w images to the MNI standard space was performed using the t1-linear pipeline of Clinica [52,56]. More precisely, bias field correction was applied using the N4ITK method [57]. Next, an affine registration was performed using the SyN algorithm [58] from ANTs [59] to align each image to the MNI space with the ICBM 2009c nonlinear symmetric template [60,61]. The registered images were further cropped to remove the background resulting in images of size 169 × 208 × 179, with 1 mm isotropic voxels.

2.7. Data Augmentation Pipeline

MRI scans from subjects in the train set underwent augmentation procedures to generate new observations and enlarge the training set. Data augmentation was performed using MONAI randomized data augmentation transforms. The specs of the applied transformations are described in Table 2. Each image was augmented 5 times.
This process enlarged the train set to 1170 images (715 FTD, 455 NC). In order to build the final train set, all the original images were kept for both groups and the augmented images were sampled to set the n of each group to 400. A summary for the composition of the final train set is shown in Table 3.

2.8. Deep Learning Pipeline

A train set of 440 images per group was used to train a classifier following the indications reported in MONAI [53]. A DenseNet121 was trained with default parameters for 3D images for 10 epochs and using a batch size of 2 (Figure 2). The Cross-Entropy loss was used for training with Adam optimizer, setting the learning rate to 1 × 10−5. DenseNet121 is a Dense Convolutional Network, where each layer is connected to every other layer and uses all preceding layers’ feature-maps as inputs [54]. When the images were imported in the PyTorch environment they underwent intensity scaling between 0 and 1, a channel was added to make the image in the channel-first format, then it was resized with scaling to 150 the longest dimension, keeping the aspect ratio of the initial image. The model with the best performance on the validation set was saved. Predictions on the test set were performed following MONAI indications and results were evaluated by computing the following evaluation metrics: accuracy, sensitivity, specificity, F1-score, and Area Under the Curve (AUC).

2.9. Explainable AI Using the Attention Map Method

In order to investigate where the DL model focused to make its prediction, we applied the Guided Gradient-weighted Class Activation Mapping (Guided Grad-CAM) algorithm using the M3d-CAM tool [62]. M3d-CAM simplifies the interpretability of PyTorch-based models by providing an easy-to-use library for generating attention maps. The application of the Guided Grad-CAM algorithm generated attention maps as new images to visualize important voxels, finally providing insights into model’s behavior [63]. In particular, the neuron importance weights ( α n c ) are computed as the global average pooling of the gradients via backpropagation, before the softmax layer of the DL network. The gradient of the score for the output is computed with respect to feature map activations of a convolutional layer, as shown in Equation (1).
α n c = 1 z i j k y c A i j k n  
where y is the model output before sigmoid function application, c is the class of interest, A is the feature map activation, i, j, and k are the width, height, and depth dimensions, respectively. To obtain a tensor heatmap of feature importance ( L G r a d C A M ) representing the activation map of the network Guided Grad-CAM executes a weighted combination of the obtained α n c with A n using the Rectified Linear Unit (ReLU) function, represented in Equation (2) [62].
L G r a d C A M = R e L U   ( n α n c A n )
The obtained tensor heatmap was then rendered as a pseudocolor image with grayscale colormapping. The areas contributing the most to the model’s output emerged with the greatest variation in grayscale.

3. Results

3.1. CAD Train and Test

A DenseNet121 architecture in PyTorch was trained to discriminate between FTD and NC T1 3D MRI following the MONAI framework guidelines. To ensure the model was not performing well only on the data it was trained on, we measured its performance on a separated validation set. In fact, the model was tested on the validation set after each epoch and the model with the best accuracy (0.92) was saved. After training, this saved model was tested on an independent test set to assess prediction performance. The model achieved 0.80 accuracy (95% confidence intervals: 0.64, 0.91), 1 sensitivity, 0.6 specificity, 0.83 F1-score, and an AUC of 0.86. Testing if the accuracy was higher than the no information rate achieved a p-value < 0.0001. The confusion matrix is reported in Figure 3, showing that the model only misclassified NC samples, achieving max sensitivity.

3.2. Comparison with Previous FTD Classification Approaches

We collected the performance metrics of the studies attempting to discriminate FTD from NC or AD to demonstrate that a reproducible DL powered CAD tool performs similarly (Table 4). Our application achieved 0.80 accuracy, which is in line with the results obtained by other research groups in the FTD vs. NC classification (Accuracymean = 0.84, sd = 0.08) while showing high reproducibility. Notably, only a few of the previous classification attempts used NIFD data, and their sample size, AI algorithms, and cross-validation strategies were highly heterogeneous. Most of the published DL applications to FTD were designed to discriminate FTD from AD, making it difficult to compare their performance with ours.

3.3. Attention Maps

Attention maps help visually explain DL models for image processing, showing which parts of the image contributed to the classification. We generated attention maps using the Guided Grad-CAM algorithm that provided an image with the same size as the test image, where relevant points have a highly perturbed value, thus emerging from the background and showing where the model focused to perform its prediction. Attention maps for the FTD and the NC subjects with the most accurate prediction are shown in Figure 3. Emerging areas concentrate around the ventricles, where the difference is clearly noticeable, as the FTD subject has expanded ventricles. It also seems that the model partly focuses on the skull. Figure 4 shows a single slice where the difference is evident, yet an animation of the full brain scans and attention maps is available in the Supplementary Materials.

4. Discussion

In this work, we showed the use of a standardized workflow to build a CAD tool for FTD. Data preprocessing was made easily reproducible by using the Clinica library [40], while an optimized state-of-the-art DL model was trained within the standardized MONAI framework [53]. The proposed workflow resulted in a CAD with max sensitivity, correctly identifying all FTD samples. Some NC samples were missed, leading to 0.8 accuracy, yet this result is in line with the previous FTD classification approaches based on MRI, whose accuracy in discriminating FTD from controls is around 0.8 [47,68]. Notwithstanding, most of the available papers on FTD classification with AI-based methods used Machine Learning on quantitative variables that, although from MRI brain scans (e.g., cortical thickness), have to be managed differently from 3D MRI data (Table 4). A few works were published where DL is used to detect FTD from 3D imaging, but their focus was on discriminating it from AD, making their results incomparable to ours (Table 4) [42,69]. There is only one published work really comparable to ours [42], but their trained DL model takes raw images as input and uses a non-standard architecture. On such a basis, we found this method built on standardized frameworks to be new in FTD research. Moreover, our work is one of few based on the NIFD database. We believe that our standardized methodology makes our work useful for benchmark comparison in FTD classification. Experiments based on neurodegenerative disease classification need large sample sizes and to the best of our knowledge NIFD offers the largest FTD cohort to date with neuroimaging and clinical data available. Finally, we found that several studies were underpowered, with few participants for each experimental group. Small sample sizes usually lead to unstable results and/or inflated performances of the models used for classification. This bias is worsened when coupled with misused Cross-Validation strategies such as k-fold Cross-Validation that are unreliable with ~20 subjects per group. A standardized workflow following data science best practices protects from data leakage or overfitting, ensuring reproducibility and reliability.
The performance of the CAD presented here could be further enhanced by training the model with a higher number of epochs, as other works show convergence after 100 epochs [42]. Additionally, tuning the model hyper-parameters such as learning rate and batch size could have yielded higher accuracy, along with choosing a different DL architecture, finally identifying the best performing combination. Moreover, we could have taken a data-centric approach to boost the effectiveness of the training phase, possibly improving test performance. As it has been shown, a data-centric approach counters the challenges of training with a small dataset and improves accuracy [70]. Nonetheless, our approach aimed at simplicity and reproducibility, showing that a CAD tool can be set up by using open-source and easy-to-use software platforms providing state-of-the-art methods for data preprocessing and analysis.
One of the biggest flaws in DL-based CAD tools is their lack of interpretability. In fact, DL neural networks are black boxes, meaning that highly complex data processing makes it unintelligible how the model comes to its final output (a class probability). To the extent of making black boxes more interpretable, a few methods for explaining model behavior have been developed, contributing to the realization of the explainable AI [12,71]. Neurodegenerative disease research is only recently approaching explainable AI [72,73,74,75] and there are only a few works available where Guided Grad-CAM has been used to generate attention maps for DL neural networks [76,77,78,79]. To the best of our knowledge, this is the first work where a DL model for FTD detection is studied with explainable AI methods. As reported in the results section, ventricular spaces were the most influential areas for the model output, and it has also been observed that the model was influenced by skull parts too. The AI presented in this work has a good performance, and attention maps showed that it seems to rely mostly on gross characteristics accounting for large differences in the images, as a human might do. To this extent, this model seems to be reliable as it shows to mimic human behavior when choosing, and this might make this CAD trustable by groups of physicians aiming to find aids from automated analysis methods when evaluating patients. We argue that the model might have misclassified the NC samples due to noise in the signal introduced by the skull parts. Evaluating the influence of skull parts in the model decision was out of the scope of this work; it would take specific experiments in order to determine how influential they were in the model decision.
This paper presents how the use of Clinica and MONAI for data preprocessing and analysis facilitates reproducibility in developing a CAD tool for FTD detection. We believe that having standardized, reproducible and trustable CAD tools would ease their inclusion in patient’s clinical management practices. In fact, although thousands of AI-based tools have been developed with the potential of smoothing disease detection or outcome prediction, their real practical application is lagging [11,80,81]. The stability of Clinica methods for data preprocessing well couples with the flexibility of the procedures proposed by MONAI, setting up a highly versatile system for CAD development. We believe that such practices should be applied on data adhering to the FAIR principles when developing an AI model to study a neurodegenerative disease. Additionally, we believe that explainable AI practices such as using and interpreting attention maps should be necessary steps in building a CAD tool, in order to bring the AI realm into standard clinical practice.

5. Limitations

This work has some limitations. First, we used only a subset of all the images available in NIFD, in essence the T1w MPRAGE MRI at T0 (patient’s first visit). NIFD includes many more MRI modalities and other timepoints. Using all images with appropriate data management strategies could improve classification performance. Second, our CAD was built on DenseNet121, the architecture that MONAI proposes for 3D classification, without hyper-parameter tuning. It is probable that a better performance could have been yielded with different parameters and it may also be that a different architecture among those available in MONAI could have performed better on our data, yet it was outside of the scope of this work to yield max performance on the test set, as performing as good as others was enough for this application. Third, images were resized when imported in the MONAI environment; thus, possibly reducing signal quality within the MRI brain scans and finally partially hindering the model discriminative performance. Nonetheless, attention maps evaluation revealed that the model focused on gross characteristics, probably disqualifying this limitation.

6. Conclusions

Although AI is now a leading technology in medical research, the real-life implementation of AI-based CAD tools in daily clinical practice is still facing obstacles. To be approved by regulators, AI-based decision support systems must be able to consistently reproduce their results on multiple sites or cohorts, while integrating with electronic health record systems. Similarly, the DL models powering the statistical engine of the CAD should be consistently updated over time to keep up with evolving clinical standards [11,80,81]. In particular, the high complexity of neurodegenerative diseases poses tough challenges for the development of reliable CAD tools, since complex diseases such as FTD, AD, and PD are characterized by a strong heterogeneity in their manifestation and underlying pathological mechanisms. Here, we presented a workflow for FTD detection based on a standardized preprocessing framework of 3D brain images, coupled with a reproducible protocol of data augmentation and Deep Learning model training and evaluation. Moreover, we used explainable AI methods to demonstrate how AI behavior can be understood by regulators and physicians. We found out that our standardized workflow for AI-based CAD tool development is comparable to other classification approaches in FTD, without compromising on reproducibility and interpretability of the DL model. Interestingly, we observed that methodological heterogeneity in FTD classification is not limited to development practices but also extends to data sources and cross-validation strategies with the latter being potentially harmful for generalizability of the results (Table 4).
Thus, we believe that health informaticians should develop AI-based CAD tools with pipeline standardization in mind, as these objectives cannot be achieved without it. In particular, we need standardized data management strategies during collection, preprocessing, and sharing, especially in case of a cross-site contribution to a database. The adherence to FAIR principles for shared data is pivotal to enhance their reusability by researchers worldwide. We believe that the widespread adoption of stronger standardization principles would foster the stability of the techniques and the reliability of findings in AI research. Within this setting, the application of explainable AI methods is required to overcome the issue posed by black box models, in particular about physicians’ trust in CAD. Increasing the understanding of AI behavior would weaken the hindering for CAD tools to be applied in real-world clinical settings, finally bringing us closer to a fruitful human–machine interaction in the biomedical field.

Supplementary Materials

The following supporting information can be downloaded at: www.mdpi.com/article/10.3390/life12070947/s1, Animation S1: full attention maps.

Author Contributions

Conceptualization, A.T. and C.F.; methodology, A.T. and C.F.; software, A.T. and C.F.; validation, A.T. and C.F.; formal analysis, A.T. and C.F.; investigation, A.T. and C.F.; resources, all authors; data curation, A.T. and C.F.; writing—original draft preparation, A.T. and C.F.; writing—review and editing, all authors; visualization, A.T. and C.F.; supervision, L.P. and C.C.; project administration, all authors; All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the Italian Ministry of Health, Ricerca Corrente.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study by the Frontotemporal Lobar Degeneration Neuroimaging Initiative.

Data Availability Statement

This study was performed using the NIFD. See the acknowledgements.

Acknowledgments

Data collection and sharing for this project was funded by the Frontotemporal Lobar Degeneration Neuroimaging Initiative (National Institutes of Health Grant R01 AG032306). The study is coordinated through the University of California, San Francisco, Memory and Aging Center. FTLDNI data are disseminated by the Laboratory for Neuro Imaging at the University of Southern California. Data used in preparation of this article were obtained from the Frontotemporal Lobar Degeneration Neuroimaging Initiative (FTLDNI) database, whose nickname is NIFD. The investigators at NIFD/FTLDNI contributed to the design and implementation of FTLDNI and/or provided data, but did not participate in analysis or writing of this report. The FTLDNI investigators included the following individuals: Howard Rosen; University of California, San Francisco (PI). Bradford C. Dickerson; Harvard Medical School and Massachusetts General Hospital. Kimoko Domoto-Reilly; University of Washington School of Medicine. David Knopman; Mayo Clinic, Rochester. Bradley F. Boeve; Mayo Clinic Rochester. Adam L. Boxer; University of California, San Francisco. John Kornak; University of California, San Francisco. Bruce L. Miller; University of California, San Francisco. William W. Seeley; University of California, San Francisco. Ma-ria-Luisa Gorno-Tempini; University of California, San Francisco. Scott McGinnis; University of California, San Francisco. Maria Luisa Mandelli; University of California, San Francisco.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Halalli, B.; Makandar, A. Computer Aided Diagnosis-Medical Image Analysis Techniques. In Breast Imaging; 2018; Volume 85, Available online: https://www.intechopen.com/chapters/56615 (accessed on 20 May 2022).
  2. Choi, J.-H.; Kang, B.J.; Baek, J.E.; Lee, H.S.; Kim, S.H. Application of Computer-Aided Diagnosis in Breast Ultrasound Interpretation: Improvements in Diagnostic Performance According to Reader Experience. Ultrasonography 2018, 37, 217–225. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Chan, H.-P.; Hadjiiski, L.M.; Samala, R.K. Computer-aided Diagnosis in the Era of Deep Learning. Med. Phys. 2020, 47, e218–e227. [Google Scholar] [CrossRef] [PubMed]
  4. Tacchella, A.; Romano, S.; Ferraldeschi, M.; Salvetti, M.; Zaccaria, A.; Crisanti, A.; Grassi, F. Collaboration between a Human Group and Artificial Intelligence Can Improve Prediction of Multiple Sclerosis Course: A Proof-of-Principle Study. F1000Research 2017, 6, 2172. [Google Scholar] [CrossRef] [PubMed]
  5. Jumper, J.; Evans, R.; Pritzel, A.; Green, T.; Figurnov, M.; Ronneberger, O.; Tunyasuvunakool, K.; Bates, R.; Žídek, A.; Potapenko, A.; et al. Highly Accurate Protein Structure Prediction with AlphaFold. Nature 2021, 596, 583–589. [Google Scholar] [CrossRef]
  6. Ramesh, A.; Dhariwal, P.; Nichol, A.; Chu, C.; Chen, M. Hierarchical Text-Conditional Image Generation with CLIP Latents. arXiv 2022, arXiv:2204.06125. [Google Scholar]
  7. Vinyals, O.; Babuschkin, I.; Czarnecki, W.M.; Mathieu, M.; Dudzik, A.; Chung, J.; Choi, D.H.; Powell, R.; Ewalds, T.; Georgiev, P.; et al. Grandmaster Level in StarCraft II Using Multi-Agent Reinforcement Learning. Nature 2019, 575, 350–354. [Google Scholar] [CrossRef]
  8. Fabrizio, C.; Termine, A.; Caltagirone, C.; Sancesario, G. Artificial Intelligence for Alzheimer’s Disease: Promise or Challenge? Diagnostics 2021, 11, 1473. [Google Scholar] [CrossRef]
  9. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  10. Healthcare Industry Trends 2022: Accessible Data, Pharma AI. Available online: https://cloud.google.com/blog/topics/healthcare-life-sciences/healthcare-industry-trends-2022-life-sciences-technology-predictions-data-ai-interoperability/ (accessed on 13 June 2022).
  11. Davenport, T.; Kalakota, R. The Potential for Artificial Intelligence in Healthcare. Future Healthc. J. 2019, 6, 94–98. [Google Scholar] [CrossRef] [Green Version]
  12. Gunning, D.; Stefik, M.; Choi, J.; Miller, T.; Stumpf, S.; Yang, G.-Z. XAI—Explainable Artificial Intelligence. Sci. Robot. 2019, 4, eaay7120. [Google Scholar] [CrossRef] [Green Version]
  13. Bang, J.; Spina, S.; Miller, B.L. Frontotemporal Dementia. Lancet 2015, 386, 1672–1682. [Google Scholar] [CrossRef] [Green Version]
  14. Vieira, R.T.; Caixeta, L.; Machado, S.; Silva, A.C.; Nardi, A.E.; Arias-Carrión, O.; Carta, M.G. Epidemiology of Early-Onset Dementia: A Review of the Literature. Clin. Pract. Epidemiol. Ment. Health 2013, 9, 88–95. [Google Scholar] [CrossRef]
  15. Brun, A.; Liu, X.; Erikson, C. Synapse Loss and Gliosis in the Molecular Layer of the Cerebral Cortex in Alzheimer’s Disease and in Frontal Lobe Degeneration. Neurodegeneration 1995, 4, 171–177. [Google Scholar] [CrossRef] [PubMed]
  16. Dugger, B.N.; Dickson, D.W. Pathology of Neurodegenerative Diseases. Cold Spring Harb. Perspect. Biol. 2017, 9, a028035. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Mackenzie, I.R.; Neumann, M. Molecular Neuropathology of Frontotemporal Dementia: Insights into Disease Mechanisms from Postmortem Studies. J. Neurochem. 2016, 138, 54–70. [Google Scholar] [CrossRef]
  18. Pijnenburg, Y.A.L.; Gillissen, F.; Jonker, C.; Scheltens, P. Initial Complaints in Frontotemporal Lobar Degeneration. Dement. Geriatr. Cogn. Disord. 2004, 17, 302–306. [Google Scholar] [CrossRef]
  19. Rosness, T.A.; Haugen, P.K.; Passant, U.; Engedal, K. Frontotemporal Dementia: A Clinically Complex Diagnosis. Int. J. Geriatr. Psychiatry 2008, 23, 837–842. [Google Scholar] [CrossRef]
  20. Gorno-Tempini, M.L.; Hillis, A.E.; Weintraub, S.; Kertesz, A.; Mendez, M.; Cappa, S.F.; Ogar, J.M.; Rohrer, J.D.; Black, S.; Boeve, B.F.; et al. Classification of Primary Progressive Aphasia and Its Variants. Neurology 2011, 76, 1006–1014. [Google Scholar] [CrossRef] [Green Version]
  21. Höglinger, G.U.; Respondek, G.; Stamelou, M.; Kurz, C.; Josephs, K.A.; Lang, A.E.; Mollenhauer, B.; Müller, U.; Nilsson, C.; Whitwell, J.L.; et al. Clinical Diagnosis of Progressive Supranuclear Palsy: The Movement Disorder Society Criteria. Mov. Disord. 2017, 32, 853–864. [Google Scholar] [CrossRef]
  22. Rascovsky, K.; Hodges, J.R.; Knopman, D.; Mendez, M.F.; Kramer, J.H.; Neuhaus, J.; van Swieten, J.C.; Seelaar, H.; Dopper, E.G.P.; Onyike, C.U.; et al. Sensitivity of Revised Diagnostic Criteria for the Behavioural Variant of Frontotemporal Dementia. Brain J. Neurol. 2011, 134, 2456–2477. [Google Scholar] [CrossRef]
  23. Benussi, A.; Di Lorenzo, F.; Dell’Era, V.; Cosseddu, M.; Alberici, A.; Caratozzolo, S.; Cotelli, M.S.; Micheli, A.; Rozzini, L.; Depari, A. Transcranial Magnetic Stimulation Distinguishes Alzheimer Disease from Frontotemporal Dementia. Neurology 2017, 89, 665–672. [Google Scholar] [CrossRef] [PubMed]
  24. Perri, R.; Koch, G.; Carlesimo, G.A.; Serra, L.; Fadda, L.; Pasqualetti, P.; Pettenati, C.; Caltagirone, C. Alzheimer’s Disease and Frontal Variant of Frontotemporal Dementia. J. Neurol. 2005, 252, 1238–1244. [Google Scholar] [CrossRef] [PubMed]
  25. Di Lorenzo, F.; Motta, C.; Caltagirone, C.; Koch, G.; Mercuri, N.B.; Martorana, A. Lacosamide in the Management of Behavioral Symptoms in Frontotemporal Dementia: A 2-Case Report. Alzheimer Dis. Assoc. Disord. 2018, 32, 364–365. [Google Scholar] [CrossRef] [PubMed]
  26. Mourik, J.C.; Rosso, S.M.; Niermeijer, M.F.; Duivenvoorden, H.J.; Van Swieten, J.C.; Tibben, A. Frontotemporal Dementia: Behavioral Symptoms and Caregiver Distress. Dement. Geriatr. Cogn. Disord. 2004, 18, 299–306. [Google Scholar] [CrossRef]
  27. Benussi, A.; Dell’Era, V.; Cosseddu, M.; Cantoni, V.; Cotelli, M.S.; Cotelli, M.; Manenti, R.; Benussi, L.; Brattini, C.; Alberici, A.; et al. Transcranial Stimulation in Frontotemporal Dementia: A Randomized, Double-Blind, Sham-Controlled Trial. Alzheimer’s Dement. 2020, 6. [Google Scholar] [CrossRef]
  28. Bonnì, S.; Koch, G.; Miniussi, C.; Bassi, M.S.; Caltagirone, C.; Gainotti, G. Role of the Anterior Temporal Lobes in Semantic Representations: Paradoxical Results of a CTBS Study. Neuropsychologia 2015, 76, 163–169. [Google Scholar] [CrossRef] [Green Version]
  29. Gerfo, E.L.; Oliveri, M.; Torriero, S.; Salerno, S.; Koch, G.; Caltagirone, C. The Influence of RTMS over Prefrontal and Motor Areas in a Morphological Task: Grammatical vs. Semantic Effects. Neuropsychologia 2008, 46, 764–770. [Google Scholar] [CrossRef]
  30. Jiskoot, L.C.; Panman, J.L.; Meeter, L.H.; Dopper, E.G.; Donker Kaat, L.; Franzen, S.; van der Ende, E.L.; van Minkelen, R.; Rombouts, S.A.; Papma, J.M. Longitudinal Multimodal MRI as Prognostic and Diagnostic Biomarker in Presymptomatic Familial Frontotemporal Dementia. Brain 2019, 142, 193–208. [Google Scholar] [CrossRef] [Green Version]
  31. Seeley, W.W.; Crawford, R.; Rascovsky, K.; Kramer, J.H.; Weiner, M.; Miller, B.L.; Gorno-Tempini, M.L. Frontal Paralimbic Network Atrophy in Very Mild Behavioral Variant Frontotemporal Dementia. Arch. Neurol. 2008, 65, 249–255. [Google Scholar] [CrossRef] [Green Version]
  32. Frings, L.; Yew, B.; Flanagan, E.; Lam, B.Y.; Hüll, M.; Huppertz, H.-J.; Hodges, J.R.; Hornberger, M. Longitudinal Grey and White Matter Changes in Frontotemporal Dementia and Alzheimer’s Disease. PLoS ONE 2014, 9, e90814. [Google Scholar]
  33. Whitwell, J.L.; Boeve, B.F.; Weigand, S.D.; Senjem, M.L.; Gunter, J.L.; Baker, M.C.; DeJesus-Hernandez, M.; Knopman, D.S.; Wszolek, Z.K.; Petersen, R.C. Brain Atrophy over Time in Genetic and Sporadic Frontotemporal Dementia: A Study of 198 Serial Magnetic Resonance Images. Eur. J. Neurol. 2015, 22, 745–752. [Google Scholar] [CrossRef] [PubMed]
  34. Knopman, D.S.; Jack, C.R.; Kramer, J.H.; Boeve, B.F.; Caselli, R.J.; Graff-Radford, N.R.; Mendez, M.F.; Miller, B.L.; Mercaldo, N.D. Brain and Ventricular Volumetric Changes in Frontotemporal Lobar Degeneration over 1 Year. Neurology 2009, 72, 1843–1849. [Google Scholar] [CrossRef] [PubMed]
  35. Manera, A.L.; Dadar, M.; Collins, D.L.; Ducharme, S.; Frontotemporal Lobar Degeneration Neuroimaging Initiative; Alzheimer’s Disease Neuroimaging Initiative. Ventricular Features as Reliable Differentiators between BvFTD and Other Dementias. NeuroImage Clin. 2022, 33, 102947. [Google Scholar] [CrossRef] [PubMed]
  36. Davatzikos, C.; Resnick, S.M.; Wu, X.; Parmpi, P.; Clark, C.M. Individual Patient Diagnosis of AD and FTD via High-Dimensional Pattern Classification of MRI. NeuroImage 2008, 41, 1220–1227. [Google Scholar] [CrossRef] [Green Version]
  37. Du, A.-T.; Schuff, N.; Kramer, J.H.; Rosen, H.J.; Gorno-Tempini, M.L.; Rankin, K.; Miller, B.L.; Weiner, M.W. Different Regional Patterns of Cortical Thinning in Alzheimer’s Disease and Frontotemporal Dementia. Brain 2007, 130, 1159–1166. [Google Scholar] [CrossRef]
  38. Spasov, S.; Passamonti, L.; Duggento, A.; Lio, P.; Toschi, N.; Alzheimer’s Disease Neuroimaging Initiative. A Parameter-Efficient Deep Learning Approach to Predict Conversion from Mild Cognitive Impairment to Alzheimer’s Disease. NeuroImage 2019, 189, 276–287. [Google Scholar] [CrossRef] [Green Version]
  39. Termine, A.; Fabrizio, C.; Strafella, C.; Caputo, V.; Petrosini, L.; Caltagirone, C.; Giardina, E.; Cascella, R. Multi-Layer Picture of Neurodegenerative Diseases: Lessons from the Use of Big Data through Artificial Intelligence. J. Pers. Med. 2021, 11, 280. [Google Scholar] [CrossRef]
  40. Crüwell, S.; van Doorn, J.; Etz, A.; Makel, M.C.; Moshontz, H.; Niebaum, J.C.; Orben, A.; Parsons, S.; Schulte-Mecklenbeck, M. Seven Easy Steps to Open Science. Zeitschrift Psychologie 2019, 227, 237–248. [Google Scholar] [CrossRef]
  41. Wilkinson, M.D.; Dumontier, M.; Aalbersberg, J.I.; Appleton, G.; Axton, M.; Baak, A.; Blomberg, N.; Boiten, J.-W.; da Silva Santos, L.B.; Bourne, P.E. The FAIR Guiding Principles for Scientific Data Management and Stewardship. Sci. Data 2016, 3, 160018. [Google Scholar] [CrossRef] [Green Version]
  42. Hu, J.; Qing, Z.; Liu, R.; Zhang, X.; Lv, P.; Wang, M.; Wang, Y.; He, K.; Gao, Y.; Zhang, B. Deep Learning-Based Classification and Voxel-Based Visualization of Frontotemporal Dementia and Alzheimer’s Disease. Front. Neurosci. 2020, 14, 626154. [Google Scholar] [CrossRef]
  43. Bron, E.E.; Smits, M.; Papma, J.M.; Steketee, R.M.E.; Meijboom, R.; de Groot, M.; van Swieten, J.C.; Niessen, W.J.; Klein, S. Multiparametric Computer-Aided Differential Diagnosis of Alzheimer’s Disease and Frontotemporal Dementia Using Structural and Advanced MRI. Eur. Radiol. 2017, 27, 3372–3382. [Google Scholar] [CrossRef] [Green Version]
  44. Zhang, Y.; Tartaglia, M.C.; Schuff, N.; Chiang, G.C.; Ching, C.; Rosen, H.J.; Gorno-Tempini, M.L.; Miller, B.L.; Weiner, M.W. MRI Signatures of Brain Macrostructural Atrophy and Microstructural Degradation in Frontotemporal Lobar Degeneration Subtypes. J. Alzheimer’s Dis. 2013, 33, 431–444. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Muñoz-Ruiz, M.Á.; Hartikainen, P.; Koikkalainen, J.; Wolz, R.; Julkunen, V.; Niskanen, E.; Herukka, S.-K.; Kivipelto, M.; Vanninen, R.; Rueckert, D.; et al. Structural MRI in Frontotemporal Dementia: Comparisons between Hippocampal Volumetry, Tensor-Based Morphometry and Voxel-Based Morphometry. PLoS ONE 2012, 7, e52531. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Dukart, J.; Mueller, K.; Horstmann, A.; Barthel, H.; Möller, H.E.; Villringer, A.; Sabri, O.; Schroeter, M.L. Combined Evaluation of FDG-PET and MRI Improves Detection and Differentiation of Dementia. PLoS ONE 2011, 6, e18111. [Google Scholar] [CrossRef] [PubMed]
  47. McCarthy, J.; Collins, D.L.; Ducharme, S. Morphometric MRI as a Diagnostic Biomarker of Frontotemporal Dementia: A Systematic Review to Determine Clinical Applicability. NeuroImage Clin. 2018, 20, 685–696. [Google Scholar] [CrossRef] [PubMed]
  48. Gorgolewski, K.; Burns, C.D.; Madison, C.; Clark, D.; Halchenko, Y.O.; Waskom, M.L.; Ghosh, S.S. Nipype: A Flexible, Lightweight and Extensible Neuroimaging Data Processing Framework in Python. Front. Neuroinform. 2011, 5, 13. [Google Scholar] [CrossRef] [Green Version]
  49. Subramanian, A.; Lan, H.; Govindarajan, S.; Viswanathan, L.; Choupan, J.; Sepehrband, F. NiftyTorch: A Deep Learning Framework for NeuroImaging. bioRxiv 2021. [Google Scholar] [CrossRef]
  50. Perone, C.S.; Cclauss; Saravia, E.; Ballester, P.L. MohitTare Perone/Medicaltorch: Release v0.2; Zenodo, 2018. Available online: https://zenodo.org/record/1495335 (accessed on 25 May 2022).
  51. Gorgolewski, K.; Alfaro-Almagro, F.; Auer, T.; Bellec, P.; Capotă, M.; Chakravarty, M.M.; Churchill, N.W.; Cohen, A.L.; Craddock, R.C.; Devenyi, G.A. BIDS Apps: Improving Ease of Use, Accessibility, and Reproducibility of Neuroimaging Data Analysis Methods. PLoS Comput. Biol. 2017, 13, e1005209. [Google Scholar] [CrossRef] [Green Version]
  52. Routier, A.; Burgos, N.; Díaz, M.; Bacci, M.; Bottani, S.; El-Rifai, O.; Fontanella, S.; Gori, P.; Guillon, J.; Guyot, A. Clinica: An Open-Source Software Platform for Reproducible Clinical Neuroscience Studies. Front. Neuroinform. 2021, 15, 689675. [Google Scholar] [CrossRef]
  53. MONAI Consortium MONAI: Medical Open Network for AI; Zenodo, 2022. Available online: https://zenodo.org/record/5728262 (accessed on 25 May 2022).
  54. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. arXiv 2016, arXiv:1608.06993. [Google Scholar]
  55. Hodnett, M.; Wiley, J.F. R Deep Learning Essentials: A Step-by-Step Guide to Building Deep Learning Models Using TensorFlow, Keras, and MXNet; Packt Publishing Ltd.: Birmingham, UK, 2018; ISBN 1-78899-780-8. [Google Scholar]
  56. Wen, J.; Thibeau-Sutre, E.; Diaz-Melo, M.; Samper-González, J.; Routier, A.; Bottani, S.; Dormont, D.; Durrleman, S.; Burgos, N.; Colliot, O. Convolutional Neural Networks for Classification of Alzheimer’s Disease: Overview and Reproducible Evaluation. Med. Image Anal. 2020, 63, 101694. [Google Scholar] [CrossRef] [PubMed]
  57. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 Bias Correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Avants, B.; Epstein, C.; Grossman, M.; Gee, J. Symmetric Diffeomorphic Image Registration with Cross-Correlation: Evaluating Automated Labeling of Elderly and Neurodegenerative Brain. Med. Image Anal. 2008, 12, 26–41. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Avants, B.; Tustison, N.J.; Stauffer, M.; Song, G.; Wu, B.; Gee, J.C. The Insight ToolKit Image Registration Framework. Front. Neuroinform. 2014, 8, 44. [Google Scholar] [CrossRef] [Green Version]
  60. Fonov, V.; Evans, A.C.; Botteron, K.; Almli, C.R.; McKinstry, R.C.; Collins, D.L. Unbiased Average Age-Appropriate Atlases for Pediatric Studies. NeuroImage 2011, 54, 313–327. [Google Scholar] [CrossRef] [Green Version]
  61. Fonov, V.; Evans, A.; McKinstry, R.; Almli, C.; Collins, D. Unbiased Nonlinear Average Age-Appropriate Brain Templates from Birth to Adulthood. NeuroImage 2009, 47, S102. [Google Scholar] [CrossRef]
  62. Gotkowski, K.; Gonzalez, C.; Bucher, A.; Mukhopadhyay, A. M3d-CAM: A PyTorch Library to Generate 3D Attention Maps for Medical Deep Learning. In Bildverarbeitung für die Medizin 2021; Palm, C., Deserno, T.M., Handels, H., Maier, A., Maier-Hein, K., Tolxdorff, T., Eds.; Informatik Aktuell; Springer: Wiesbaden, Germany, 2021; pp. 217–222. ISBN 978-3-658-33197-9. [Google Scholar]
  63. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. arXiv 2016, arXiv:1610.02391. [Google Scholar]
  64. Chagué, P.; Marro, B.; Fadili, S.; Houot, M.; Morin, A.; Samper-González, J.; Beunon, P.; Arrivé, L.; Dormont, D.; Dubois, B.; et al. Radiological Classification of Dementia from Anatomical MRI Assisted by Machine Learning-Derived Maps. J. Neuroradiol. 2021, 48, 412–418. [Google Scholar] [CrossRef]
  65. McMillan, C.T.; Avants, B.B.; Cook, P.; Ungar, L.; Trojanowski, J.Q.; Grossman, M. The Power of Neuroimaging Biomarkers for Screening Frontotemporal Dementia. Hum. Brain Mapp. 2014, 35, 4827–4840. [Google Scholar] [CrossRef] [Green Version]
  66. Lehmann, M.; Rohrer, J.D.; Clarkson, M.J.; Ridgway, G.R.; Scahill, R.I.; Modat, M.; Warren, J.D.; Ourselin, S.; Barnes, J.; Rossor, M.N.; et al. Reduced Cortical Thickness in the Posterior Cingulate Gyrus Is Characteristic of Both Typical and Atypical Alzheimer’s Disease. J. Alzheimer’s Dis. 2010, 20, 587–598. [Google Scholar] [CrossRef] [Green Version]
  67. Klöppel, S.; Stonnington, C.M.; Chu, C.; Draganski, B.; Scahill, R.I.; Rohrer, J.D.; Fox, N.C.; Jack, C.R.; Ashburner, J.; Frackowiak, R.S.J. Automatic Classification of MR Scans in Alzheimer’s Disease. Brain J. Neurol. 2008, 131, 681–689. [Google Scholar] [CrossRef] [PubMed]
  68. Kim, J.P.; Kim, J.; Park, Y.H.; Park, S.B.; Lee, J.S.; Yoo, S.; Kim, E.-J.; Kim, H.J.; Na, D.L.; Brown, J.A.; et al. Machine Learning Based Hierarchical Classification of Frontotemporal Dementia and Alzheimer’s Disease. NeuroImage Clin. 2019, 23, 101811. [Google Scholar] [CrossRef] [PubMed]
  69. Ma, D.; Lu, D.; Popuri, K.; Wang, L.; Beg, M.F.; Alzheimer’s Disease Neuroimaging Initiative. Differential Diagnosis of Frontotemporal Dementia, Alzheimer’s Disease, and Normal Aging Using a Multi-Scale Multi-Type Feature Generative Adversarial Deep Neural Network on Structural Magnetic Resonance Images. Front. Neurosci. 2020, 14, 853. [Google Scholar] [CrossRef] [PubMed]
  70. Motamedi, M.; Sakharnykh, N.; Kaldewey, T. A Data-Centric Approach for Training Deep Neural Networks with Less Data. arXiv 2021, arXiv:2110.03613. [Google Scholar]
  71. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A Survey of Methods for Explaining Black Box Models. ACM Comput. Surv. 2019, 51, 1–42. [Google Scholar] [CrossRef] [Green Version]
  72. El-Sappagh, S.; Alonso, J.M.; Islam, S.M.R.; Sultan, A.M.; Kwak, K.S. A Multilayer Multimodal Detection and Prediction Model Based on Explainable Artificial Intelligence for Alzheimer’s Disease. Sci. Rep. 2021, 11, 2660. [Google Scholar] [CrossRef]
  73. Essemlali, A.; St-Onge, E.; Descoteaux, M.; Jodoin, P.-M. Understanding Alzheimer Disease’s Structural Connectivity through Explainable AI. Proc. Mach. Learn. Res. 2020, 121, 217–229. [Google Scholar]
  74. Kamal, M.S.; Northcote, A.; Chowdhury, L.; Dey, N.; Crespo, R.G.; Herrera-Viedma, E. Alzheimer’s Patient Analysis Using Image and Gene Expression Data and Explainable-AI to Present Associated Genes. IEEE Trans. Instrum. Meas. 2021, 70. [Google Scholar] [CrossRef]
  75. Varzandian, A.; Razo, M.A.S.; Sanders, M.R.; Atmakuru, A.; Di Fatta, G. Classification-Biased Apparent Brain Age for the Prediction of Alzheimer’s Disease. Front. Neurosci. 2021, 15, 673120. [Google Scholar] [CrossRef]
  76. Iizuka, T.; Fukasawa, M.; Kameyama, M. Deep-Learning-Based Imaging-Classification Identified Cingulate Island Sign in Dementia with Lewy Bodies. Sci. Rep. 2019, 9, 8944. [Google Scholar] [CrossRef] [Green Version]
  77. Solano-Rojas, B.; Villalón-Fonseca, R. A Low-Cost Three-Dimensional DenseNet Neural Network for Alzheimer’s Disease Early Discovery. Sensors 2021, 21, 1302. [Google Scholar] [CrossRef] [PubMed]
  78. Tang, Z.; Chuang, K.V.; DeCarli, C.; Jin, L.-W.; Beckett, L.; Keiser, M.J.; Dugger, B.N. Interpretable Classification of Alzheimer’s Disease Pathologies with a Convolutional Neural Network Pipeline. Nat. Commun. 2019, 10, 2173. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  79. Zhang, X.; Han, L.; Zhu, W.; Sun, L.; Zhang, D. An Explainable 3D Residual Self-Attention Deep Neural Network for Joint Atrophy Localization and Alzheimer’s Disease Diagnosis Using Structural MRI. IEEE J. Biomed. Health Inform. 2021. [Google Scholar] [CrossRef] [PubMed]
  80. Jiang, F.; Jiang, Y.; Zhi, H.; Dong, Y.; Li, H.; Ma, S.; Wang, Y.; Dong, Q.; Shen, H.; Wang, Y. Artificial Intelligence in Healthcare: Past, Present and Future. Stroke Vasc. Neurol. 2017, 2, 230–243. [Google Scholar] [CrossRef]
  81. Kumar, Y.; Koul, A.; Singla, R.; Ijaz, M.F. Artificial Intelligence in Disease Diagnosis: A Systematic Literature Review, Synthesizing Framework and Future Research Agenda. J. Ambient Intell. Humaniz. Comput. 2022. [Google Scholar] [CrossRef]
Figure 1. Graphical representation of the main steps of the workflow.
Figure 1. Graphical representation of the main steps of the workflow.
Life 12 00947 g001
Figure 2. Schematic representation of DenseNet121. Its structure is similar to a classical convolutional neural network, yet DenseNet121 features dense blocks that concatenate outputs of multiple connected layers. In fact, within a dense block each layer is directly connected to every other layer in a feed-forward fashion.
Figure 2. Schematic representation of DenseNet121. Its structure is similar to a classical convolutional neural network, yet DenseNet121 features dense blocks that concatenate outputs of multiple connected layers. In fact, within a dense block each layer is directly connected to every other layer in a feed-forward fashion.
Life 12 00947 g002
Figure 3. (A) The confusion matrix indicates classification results for both classes. Rows indicate true labels and columns indicate predicted labels. (B) Receiver Operating Characteristic (ROC) curve of the DenseNet121 classifier obtained when predicting disease status (FTD/NC) using 3D T1w MRI. Area Under the Curve (AUC) was calculated as the definite integral between 0 and 1 on the x-axis and provides an aggregate measure of performance.
Figure 3. (A) The confusion matrix indicates classification results for both classes. Rows indicate true labels and columns indicate predicted labels. (B) Receiver Operating Characteristic (ROC) curve of the DenseNet121 classifier obtained when predicting disease status (FTD/NC) using 3D T1w MRI. Area Under the Curve (AUC) was calculated as the definite integral between 0 and 1 on the x-axis and provides an aggregate measure of performance.
Life 12 00947 g003
Figure 4. Coronal view of the brain for one FTD and one NC subject. The original brain scans used for testing are in the upper row, while the attention maps are in the lower row.
Figure 4. Coronal view of the brain for one FTD and one NC subject. The original brain scans used for testing are in the upper row, while the attention maps are in the lower row.
Life 12 00947 g004
Table 1. Data splitting before augmentation.
Table 1. Data splitting before augmentation.
GroupTrain (n)Validation (n)Test (n)
FTD1431920
NC911920
Table 2. Transformations applied to perform data augmentation.
Table 2. Transformations applied to perform data augmentation.
TransformationDescriptionProbability of ApplicationSpecs
TranslationTranslate voxels for every spatial dimension.1±2 voxels
RotationRandomly rotate the
input arrays.
1±5 degrees on the x-axis
Gaussian noiseAdd Gaussian noise to the image.0.5Mean = 0; standard deviation = 2.5% of the range of activation values in the image
Contrast
adjustment
Randomly updates each voxel intensity by gamma.1Gamma range = (0.0, 3.0)
Table 3. Train set after data augmentation.
Table 3. Train set after data augmentation.
GroupOriginal Images (n)Augmented Images (n)Total
FTD143257400
NC91309400
Total234566800
Table 4. FTD classification results.
Table 4. FTD classification results.
CitationComparisonSample SizeClassification MethodFeaturesMetric
Proposed applicationFTD vs. NC182 FTD
130 NC
HOTS DenseNet1213D T1 MRI scansAcc = 0.80
Hu et al., 2020 [42]FTD vs. NC552 FTD
354NC
HOTS CNNRaw 3D T1 MRI imagesAcc = 0.93
Bron et al., 2017 [43]FTD vs. NC33 FTD
34 NC
4-fold CV SVMWhole-brain VBM volume of GMAUC = 0.95
Zhang et al., 2013 [44]FTD vs. NC25 FTD
19 NC
4-fold CV SVMVBM GM volume on frontotemporal ROIAcc = 0.66
Muñoz-Ruiz et al., 2012 [45]FTD vs. NC37 FTD
26 NC
HOTS regressionVBM GM volumeAcc = 0.85
Dukart et al., 2011 [46]FTD vs. NC14 FTD
13 NC
LOOCV SVMROIs GMAcc = 0.85
Davatzikos et al., 2008 [36]FTD vs. NC12 FTD
12 NC
LOOCV SVMPCA on RAVENS GM and WM volumeAcc = 1
Du et al., 2007 [37]FTD vs. NC19 FTD
23 NC
LOOCV LRFrontal
volume
Acc = 0.89
Chagué et al., 2021 [64]FTD vs. Late Onset AD39 FTD
34 AD
10-fold CV SVMGM and WM
volumes
Acc = 0.72
Chagué et al., 2021 [64]FTD vs. Early Onset AD39 FTD
34 AD
10-fold CV
SVM
GM and WM volumesAcc = 0.80
Bron et al., 2017 [43]FTD vs. AD33 FTD
24 AD
4-fold CV SVMWhole-brain VBM volume of GMAUC = 0.78
McMillan et al., 2014 [65]FTD vs. AD72 FTD
21 AD
HOTS linear regressionGlobal ventricles volumeAUC = 0.83
Dukart et al., 2011 [46]FTD vs. AD14 FTD
21 AD
LOOCV SVMROIs GMAcc = 0.60
Lehmann et al., 2010 [66]FTD vs. AD23 FTD
17 AD
CV SVMWhole brain cortical thicknessAcc = 0.79
Davatzikos et al., 2008 [36]FTD vs. AD12 FTD
37 AD
LOOCV SVMPCA on
RAVENS GM and WM
volume
Acc = 0.84
Klöppel et al., 2008 [67]FTD vs. AD19 FTD
18 AD
LOOCV SVMGM volumeAcc = 0.89
Acc: Accuracy; AUC: Area Under the Curve; CV: Cross-Validation; GM: Grey Matter; HOTS: held-out test set; LOOCV: Leave-One-Out Cross-Validation; LR: Logistic Regression; SVM: Support Vector Machine; VBM: Voxel-Based Morphometry; WM: White Matter. In the case that a paper presents multiple classification results based on different feature sets (e.g., comparing classification performance on hippocampus volume vs. VBM-GM volume), only the best result was reported in this table. Moreover, only results obtained with brain morphometry data were considered. Accuracy was reported where available, otherwise AUC was reported. The table is arranged by Comparison (descending) and year of publication (descending).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Termine, A.; Fabrizio, C.; Caltagirone, C.; Petrosini, L.; on behalf of the Frontotemporal Lobar Degeneration Neuroimaging Initiative. A Reproducible Deep-Learning-Based Computer-Aided Diagnosis Tool for Frontotemporal Dementia Using MONAI and Clinica Frameworks. Life 2022, 12, 947. https://doi.org/10.3390/life12070947

AMA Style

Termine A, Fabrizio C, Caltagirone C, Petrosini L, on behalf of the Frontotemporal Lobar Degeneration Neuroimaging Initiative. A Reproducible Deep-Learning-Based Computer-Aided Diagnosis Tool for Frontotemporal Dementia Using MONAI and Clinica Frameworks. Life. 2022; 12(7):947. https://doi.org/10.3390/life12070947

Chicago/Turabian Style

Termine, Andrea, Carlo Fabrizio, Carlo Caltagirone, Laura Petrosini, and on behalf of the Frontotemporal Lobar Degeneration Neuroimaging Initiative. 2022. "A Reproducible Deep-Learning-Based Computer-Aided Diagnosis Tool for Frontotemporal Dementia Using MONAI and Clinica Frameworks" Life 12, no. 7: 947. https://doi.org/10.3390/life12070947

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop