You are currently viewing a new version of our website. To view the old version click .
Biophysica
  • Article
  • Open Access

12 November 2025

Evaluating the Effectiveness of Machine Learning for Alzheimer’s Disease Prediction Using Applied Explainability

,
and
1
School of Systems Biology, George Mason University, Fairfax, VA 22030, USA
2
Department of Biological Systems Engineering (BSE), Virginia Tech, Arlington, VA 22203, USA
3
Interdisciplinary Program in Neuroscience, George Mason University, Fairfax, VA 22030, USA
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Computational Biophysics

Abstract

Early and accurate diagnosis of Alzheimer’s disease (AD) is critical for patient outcomes yet presents a significant clinical challenge. This study evaluates the effectiveness of four machine learning models—Logistic Regression, Random Forest, Support Vector Machine, and a Feed-Forward Neural Network—for the five-class classification of AD stages. We systematically compare model performance under two conditions, one including cognitive assessment data and one without, to quantify the diagnostic value of these functional tests. To ensure transparency, we use SHapley Additive exPlanations (SHAPs) to interpret the model predictions. Results show that the inclusion of cognitive data is paramount for accuracy. The RF model performed best, achieving an accuracy of 84.4% with cognitive data included. Without this, performance for all models dropped significantly. SHAP analysis revealed that in the presence of cognitive data, models primarily rely on functional scores like the Clinical Dementia Rating—Sum of Boxes. In their absence, models correctly identify key biological markers, including PET (positron emission tomography) imaging of amyloid burden (FBB, AV45) and hippocampal atrophy, as the next-best predictors. This work underscores the indispensable role of cognitive assessments in AD classification and demonstrates that explainable AI can validate model behavior against clinical knowledge, fostering trust in computational diagnostic tools.

1. Introduction

Alzheimer’s disease (AD) is an irreversible, progressive brain disorder that severely impairs memory and cognitive functions, leading to significant challenges in daily activities. It is a leading cause of dementia and was associated with 5.8 million cases in the United States in 2020 alone, a number projected to more than double by the mid-century []. Genetic research over recent decades has identified several genes correlated with AD causality, enriching our understanding of its biological underpinnings [,,,].
The trajectory of Alzheimer’s disease encompasses seven stages: preclinical, asymptomatic cerebral amyloidosis, subjective cognitive decline, mild cognitive impairment (MCI), mild AD, moderate AD, and severe AD []. Our study focuses on a five-class classification task distinguishing cognitively normal (CN) individuals, those with subjective memory concerns (SMC), early mild cognitive impairment (EMCI), late mild cognitive impairment (LMCI) and those with AD.
Conventional diagnostic methods including clinical evaluations, cognitive assessments, magnetic resonance imaging (MRI), and positron emission tomography (PET) scans offer high accuracy but involve significant cost and invasiveness []. While these imaging modalities remain critical biomarkers for amyloid and tau pathology, integrating non-invasive measures such as demographic, genetic, and questionnaire-based data may enhance early detection accessibility. Recent surveys of AI and explainable AI in clinical decision support, including applications to neurodegenerative disorders, have identified best practices, common pitfalls, and emerging directions for ML and DNN models in healthcare [,,,,].
Previous Alzheimer’s disease prediction efforts using “black box” models such as extreme gradient boosting (XGBoost) and SVM have demonstrated strong classification performance but offered little insight into how inputs drive each prediction. For example, an XGBoost model achieved over 90% accuracy in early AD detection without any post hoc interpretation of feature contributions, and traditional SVM approaches similarly report high accuracy without transparent decision rationale [,]. SHAP overcomes these limitations by decomposing a model’s output into additive feature attributions, yielding both global importance rankings and patient-level explanations that clinicians can audit for trust and reliability []. Other research has also focused on early diagnosis by applying machine learning to various data modalities. For example, some studies have successfully used RF and Neural Networks (NNs) for early detection, employing methods like Gini importance to rank predictive variables []. While useful, Gini importance provides a global-level ranking and lacks the granular, instance-level insight that SHAP provides for auditing individual predictions.
We evaluated four predictive models: logistic regression (LR), random forest (RF), support vector machine (SVM), and feedforward neural network (FNN) on the ADNI merge dataset, which includes cognitive assessments, imaging biomarkers, genetic variants, and demographic measures (non-biological data points capturing contextual and functional information such as age, sex, education level, lifestyle factors, and self-reported cognitive scores). We then applied SHAP explainability techniques to each model to reveal both global feature importance and patient-level explanations, highlighting how cognitive assessments versus imaging features drive the predictions [,].
Early diagnosis is crucial for improving patient outcomes; by elucidating model behavior with XAI, we aim to foster clinician trust and pave the way for transparent, data-driven decision support in Alzheimer’s disease care. Launched in 2004, ADNI is a critical study aimed at identifying biomarkers for Alzheimer’s disease through clinical, genetic, and imaging data []. The ADNI merge dataset, a comprehensive amalgamation of ADNI1, ADNI-GO, and ADNI2 phases, includes detailed clinical outcomes, demographic data, genetic profiles, and critical biomarkers like amyloid and tau []. It is valuable for developing predictive models using advanced analytics such as machine learning and deep learning. This dataset comprises clinical, imaging, and genetic data and includes significant information collected through detailed cognitive assessments. These cognitive assessments cover various topics crucial for understanding the broader context of Alzheimer’s disease, including participants’ lifestyle factors, cognitive functions, and overall health status. While the inclusion of these cognitive assessments enriches the dataset, providing nuanced information about the progression and risk factors associated with Alzheimer’s disease, there is also a risk that they might lead to redundancy. This overlap could skew the predictions by weighing the models more heavily towards duplicated information from the datasets and the cognitive assessments.
By comparing models trained on the entire ADNI merge dataset to those trained exclusively on data subsets, excluding the cognitive assessments, this study aims to discern the value of these non-biological data points added. This comparison is essential for determining whether the cognitive assessments contribute meaningful insights that improve model performance or if they introduce noise and redundancy that could potentially skew the predictions. Additionally, the primary aim of this research is to compare the results between different predictive models, specifically machine learning models and deep neural networks, focusing on their explainability within the context of AD prediction. This multifaceted analysis will help optimize the dataset for future research and refine the models for higher accuracy and reliability in predicting Alzheimer’s disease progression and onset. Furthermore, by assessing the transparency and interpretability of these models, the study enhances understanding of how decisions are made, thus increasing trust and applicability in clinical settings. This approach underscores the importance of meticulously examining all data dimensions and their roles in enhancing the capabilities of predictive models in neurodegenerative diseases.
At baseline (n = 2399), the class distribution is imbalanced (Table 1), with LMCI the largest group (28.51%; n = 684) and SMC the smallest (14.46%; n = 347). It further includes demographic measures, clinical assessments, imaging biomarkers, genetic variants, and CSF markers (Table 2), offering a comprehensive set of predictors. The ‘Non-baseline duplicates’ column in Table 2 quantifies the number of variables (e.g., AV45) that exist in the raw dataset alongside a specific baseline-suffixed version (e.g., AV45_bl). As this study uses a strict baseline-only (cross-sectional) design, these 50 unsuffixed columns, which represent data from other time points, were removed during preprocessing to prevent temporal leakage. This ensured that only the baseline variables were used for analysis. To correct for the overrepresentation of cognitively normal cases and underrepresentation of Alzheimer’s disease cases, class weights inversely proportional to these baseline frequencies are applied during model training, ensuring that each diagnostic group contributes appropriately to the learning process.
Table 1. Cohort size and class distribution at baseline.
Table 2. Number of variables by category in the ADNI merge dataset.

2. Materials and Methods

2.1. Data Source and Cohort

Analyses use the Alzheimer’s Disease Neuroimaging Initiative (ADNI) merged table (ADNIMERGE) under a cross-sectional, baseline-only design. Records are restricted to the baseline visit (VISCODE = “bl”), and predictors are limited to variables with the baseline suffix “_bl”. Where baseline-suffixed and unsuffixed versions co-exist, unsuffixed duplicates are removed to avoid temporal leakage. Administrative and time-keeping fields (e.g., RID, PTID, site/protocol, EXAMDATE, Month/M, Month_bl/Years_bl, FSVERSION, IMAGEUID_bl) are removed. The outcome is baseline clinical diagnosis (DX_bl) encoded as a five-class label: CN, SMC, EMCI, LMCI, AD. Two analysis tables are produced: a with-cognition table that retains cognitive tests and a no-cognition table created by removing them. After preprocessing, each table contains 2399 rows; the with-cognition table has 35 columns, and the no-cognition table has 21 columns.

2.2. Study Design and Partitioning

All statistical analyses and model development were conducted using Python (version 3.9.9; Python Software Foundation, Wilmington, DE, USA). To ensure methodological rigor and prevent data leakage, the complete, preprocessed dataset is first partitioned into a stratified 80/20 train/test split (1919 training rows, 480 test rows). This partition is performed prior to any feature imputation, scaling, or selection. All subsequent preprocessing steps are encapsulated in scikit-learn pipelines that are fitted only on the 80% training data. The resulting transformations are then applied, without refitting, to the 20% hold-out test set to generate a valid performance estimate. To address class imbalance in the classical models, class weights are set inversely proportional to class frequency (the “balanced” scheme), where
w c =   N K   n c =   1 K   p c   .
Here,  w c  is the computed weight for a given class  c N  is the total number of training samples,  K  is the total number of classes (five in this study),  n c  is the number of samples belonging to class  c , and  p c  is the prevalence of class  c  in the training data. A fixed analysis seed of 42 is used for data splitting and model initialization to aid reproducibility.

2.3. Preprocessing Pipeline

A robust preprocessing workflow is defined using scikit-learn’s (version 1.5.2; Inria, Le Chesnay-Rocquencourt, France) Pipeline and ColumnTransformer objects. This approach ensures that all feature transformations are learned solely from the training partition. First, rows with more than 30% missing fields are removed. The remaining features are then separated into two distinct streams.
Numerical features are processed using a pipeline that first imputes missing values with a five-nearest-neighbor imputer (KNNImputer) and subsequently standardizes all features using StandardScaler (Z-score). Categorical features, which include demographic data (such as gender, marital status, education level, etc.), genetic status (APOE4), and certain CSF biomarkers (ABETA, TAU, PTAU), are processed with a separate pipeline. This categorical stream first imputes missing values with a constant “Missing” string (SimpleImputer) and then converts the features into a numerical format using OneHotEncoder, which is configured to ignore unknown categories encountered in the test data.

2.4. Feature Selection and Standardization

The ColumnTransformer (containing the numerical and categorical pipelines) is combined into a single Pipeline object with subsequent feature selection steps. Immediately following the preprocessing, a VarianceThreshold transformer is applied to remove any zero-variance features that may have been created during one-hot encoding. Following this, univariate ANOVA F-test screening (SelectKBest) is used to select the most informative features. The number of retained features (k) is tuned as a key hyperparameter during model training, with the search space set to 20, 50, or ‘all’ (all features remaining after the variance threshold). The F-statistics for a given feature  x j  is defined as:
F j =   SS between / K 1 SS within / N K ,   SS between =   c = 1 K n c x ¯ j ,   c   x ¯ j 2 ,   SS within =   c = 1 K n c   1 s j , c 2
Retained numeric predictors were z-scaled as described in preprocessing. This integrated pipeline ensures that feature selection is part of the cross-validation and training process, treating it as a component of the model itself to avoid data leakage.

2.5. Models and Training Procedures

Four model families are evaluated: multinomial LR, SVM, RF, and FNN. The placement of screening, scaling, nested cross-validation, and calibration is summarized in Figure 1.
Figure 1. Overall pipeline.

2.5.1. Logistic Regression (Multinomial, L2)

A multinomial, L2-regularized logistic regression fits after feature screening and z-scaling. The regularization strength C is tuned over four values (0.01, 0.1, 1, 10), and the maximum number of iterations is set to 1000. The selected C was chosen by an inner three-fold grid search within an outer five-fold scheme. For class  k , the linear score is  z k x = β k x ~ + b k  with standardized  x ~ , and the SoftMax likelihood is
p y = k x =   exp z k c = 1 K exp z k .
Training minimizes the class-weighted multinomial cross-entropy with L2 penalty:
min β k ,    b k i = 1 N k = 1 K 1 [ y i = k ] log p y = k x i w y i +   λ k = 1 K β k 2 2 ,                 λ =   1 / c ,
where  i  indexes  N  training samples,  k  indexes  K  classes,  1 ( · )  is the indicator function,  w y i  is the weight for class  y i β k  and  b k  are the weights and bias for class  k , and  λ  is the L2 regulation strength.
This specification provides a calibrated linear reference model with directionally interpretable coefficients and improved numerical stability in the presence of correlated predictors.

2.5.2. Support Vector Machine (One-Vs-Rest; RBF Kernels)

After screening and z-scaling, SVMs are trained in a one-vs-rest fashion using a radial-basis-function (RBF) kernel. The regularization parameter  C  is evaluated at three levels (0.1, 1, 10). The RBF with parameter  γ  follows the library’s scale and auto options. For each binary subproblem, the linear soft-margin primal is
min w , b , ξ 1 2 w   2 2 + C i = 1 N ξ i             s . t .             y i ( w x ~ i + b ) 1 ξ i   , ξ i 0 ,
and the kernel decision function is
f x =   i =   1 N α i y i K x ~ i , x ~ + b ,         K RBF x ,   x =   exp γ x x 2 2 ,
where  w  and  b  define the hyperplane,  ξ i  are slack variables,  C  is the regularization parameter,  α i  are the learned dual coefficients,  K ( · , · )  is the kernel function, and  γ  is the RBF kernel coefficient. Posterior class probabilities for SVMs are obtained by post hoc Platt (sigmoid) calibration as described below.

2.5.3. Random Forest (Gini Impurity; Leaf-Posterior Probabilities)

A random-forest classifier is tuned after feature screening. The number of trees varies between one hundred and two hundred. The maximum tree depth is either left unrestricted or capped at ten levels. At each split, the number of candidate features is set either to the square root of the total or to the base-2 logarithm of the total. The minimum number of samples required at a leaf node is tested at one, two, and four. Both bootstrap sampling and the no-bootstrap alternative are considered. Splits minimize weighted child impurity under the Gini index,
Gini S = 1   k = 1 K p k 2 ,                       Gini =   Gini S   v   left . right S v S Gini ( S v )   ,
where  S  is a node,  p k  is the proportion of class  k  samples at that node, the split gain  Gini  is the impurity of the parent  S  without the weighted impurity of its children ( S v ), and class probabilities are the leaf posteriors averaged across trees.

2.5.4. Feed-Forward Neural Network

A dense neural network, built with Keras (version 3.7.0; Google, Mountain View, CA, USA) on a TensorFlow (version 2.18.0; Google, Mountain View, CA, USA) backend, was tuned using Keras-Tuner (version 1.4.7; Google, Mountain View, CA, USA) RandomSearch. The number of hidden layers ranges from one to ten. The width of each hidden layer ranges from 32 to 256 units in steps of 32. Nonlinearities are either ReLU or tanh. Dropout is explored from 0.0 to 0.5 in increments of 0.1. The learning rate is searched between  10 4  and  10 2  on a log scale. Training uses Adam with early stopping (patience of five epochs), up to 500 epochs, and a batch size of 32. The tuner evaluates up to one hundred trials and executes each trial five times; the best configuration for each dataset is then saved and evaluated on the 20% test split. For summary estimates, those hyperparameters are frozen and assessed in an outer five-fold cross-validation. With hidden layers  h ( l ) = ϕ ( l ) W ( l ) h ( l 1 ) + b ( l ) , h ( 0 ) = x ~ , the SoftMax output is
p y = k   x ) =   exp z k c = 1 K exp z c ,                     z =   W l h l 1 +   b l
optimized via
L =   i = 1 N k = 1 K 1 y i = k log p y = k     x i ) ,
where  h ( l )  is the output of the  l -th hidden layer,  ϕ ( l )  is the activation function,  W ( l )  and  b ( l )  are the weights and bias, and  L  is the sparse categorical cross-entropy loss.

2.6. Validation and Calibrated Inference

For LR, SVM, and RF, model selection and performance estimation use nested cross-validation with an outer five-fold stratified split (with shuffling) and an inner three-fold grid search confined to the training portion of each outer fold. After tuning, the selected estimator is refit on the 80% training split and its probabilities are calibrated using three-fold Platt (sigmoid) calibration before evaluation on the 20% hold-out. For the FNN, the tuned model is evaluated on the hold-out and then assessed with an outer five-fold cross-validation using frozen hyperparameters. Primary metrics are Accuracy, weighted Precision, weighted Recall, weighted F1, and one-vs-rest ROC-AUC. Per-class reports, confusion matrices, and per-class ROC curves are generated for the hold-out.

2.7. Explainability

Model explainability was implemented using the SHAP library (version 0.43.0; Open-source), applying the optimal, model-specific explainer for each architecture. For the scikit-learn models, we first extract the final trained classifier and its corresponding processed data from the full pipeline. shap.LinearExplainer is used for Logistic Regression, and shap.TreeExplainer is used for Random Forest, disabling the additivity check to accommodate minor floating-point variations. For the SVM and FNN models, the model-agnostic shap.KernelExplainer is used, with its background dataset summarized using shap.kmeans ( k  = 10) to ensure computational feasibility. For an instance  x  and feature set  F , the Shapley value for the feature is
ϕ j =   S   F { j } S ! F S 1 ! F ! f x S     j f x S ,
where  ϕ j  is the Shapley value (contribution) of feature  j F  is the set of all features,  S  is a subset of features not including  j , and  f x ·  is the model’s prediction function for instance  x  given a feature subset approximated by the respective explainers. Up to 500 rows are subsampled per dataset for explanation, and both global importance bars (mean  ϕ j ) and class-wise beeswarm plots are produced.

2.8. Ethics and Data Use

Data used in the preparation of this article were obtained from the ADNI database (https://adni.loni.usc.edu; accessed on 20 September 2022). The ADNI was launched in 2003 as a public–private partnership, led by Principal Investigator Michael W. Weiner, MD. The primary goal of ADNI has been to test whether serial MRI, PET, other biological markers, and clinical and neuropsychological assessment can be combined to measure the progression of MCI and early AD.

3. Results

This study evaluates five diagnostic groups at baseline in ADNI: CN, SMC, EMCI, LMCI, and AD. The models were assessed under two feature regimes: models trained without cognitive assessments (“no-cognition”) and models trained with them (“with-cognition”). Four algorithms are evaluated consistently across both regimes: Logistic Regression (LR), Random Forest (RF), Support Vector Machine (SVM), and a Feed-Forward Neural Network (FNN). Model performance is summarized by a comparative plot of key metrics, while class-level behavior is detailed in confusion matrices. Crucially, model behavior is interrogated using SHAP to obtain global and class-conditioned attribution patterns, ensuring transparency.
The inclusion of cognitive assessments improves predictive performance across all models (Table 3). The Random Forest (RF) model was the top performer with or without cognitive assessment, achieving an overall accuracy of 84.4% when cognitive assessments were present. It was followed by the Support Vector Machine (SVM) at 79.7% and the Feed-Forward Neural Network (FNN) at 75.5%. The Logistic Regression (LR) model provided a strong baseline of 72.4% accuracy.
Table 3. Consolidated Model Performance on Held-Out Test Set.
In contrast, when cognitive assessments were excluded, model performance degraded, though all models still performed significantly better than chance. In this ‘no-cognition’ scenario, accuracies ranged from 67.4% (LR) to a high of 76.1% (RF).
As a linear baseline, the LR model’s performance clearly illustrates the difficulty of separating these clinical states, especially where class boundaries overlap. With cognitive data, the LR model achieved 72.4% accuracy. It was effective at identifying anchor classes, with a recall of 81.7% for AD and 92.6% for CN. However, its performance on ambiguous classes was poor; recall for the SMC class was only 22.9%, with the model misclassifying 51 of 70 SMC cases as CN (Figure 2A). Without cognitive data, its accuracy dropped to 67.4% (Figure 2B).
Figure 2. Confusion matrixes for the logistic regression model on the held-out test set with cognitive assessments (A) and without cognitive assessments (B).
The RF classifier emerged as the most powerful and well-balanced model. With cognitive data, it achieved 84.4% accuracy and demonstrated high, consistent F1-scores across all classes: AD (0.852), CN (0.832), EMCI (0.885), LMCI (0.888), and SMC (0.715). The confusion matrix shows a strong diagonal with high true-positive rates for all classes (Figure 3A). Even without cognitive assessments, the RF model led all others with 76.1% accuracy (Figure 3B).
Figure 3. Confusion matrixes for the random forest model on the held-out test set with cognitive assessments (A) and without cognitive assessments (B).
The SVM performed as a strong contender, though its performance revealed specific weaknesses in handling classes with subtle distinctions. The SVM achieved the second-highest accuracy at 79.7% with cognitive data. It performed well on AD (F1-score 0.881) and LMCI (F1-score 0.848) and showed high recall for the populous CN (79.6%) and LMCI (87.0%) classes. However, its recall for the SMC class was 57.1%, misclassifying 24 of 70 cases as CN (Figure 4A). Without cognitive data, its accuracy was 73.4% (Figure 4B).
Figure 4. Confusion matrixes for the support vector machine model on the held-out test set with cognitive assessments (A) and without cognitive assessments (B).
The FNN demonstrated strong but nuanced performance, highlighting the unique learning patterns of a deep learning architecture. With cognitive data, the FNN achieved 75.5% accuracy and produced the highest recall of any model for the AD class (85.4%). Its performance was hampered by a low F1-score for SMC (0.595), misclassifying 17 SMC cases as CN (Figure 5A). Without cognitive data, its accuracy was 69.1%. In this scenario, it had a uniquely high recall for CN (69.4%) but misclassified 25 LMCI cases as AD (Figure 5B).
Figure 5. Confusion matrixes for the feed-forward neural network model on the held-out test set with cognitive assessments (A) and without cognitive assessments (B).
SHAP analysis provides a coherent explanatory picture that differs logically and dramatically between the two feature regimes, revealing why the models make their decisions. When cognitive assessments are included, the feature importance landscape is unequivocally dominated by cognitive scores across all model types. For the top-performing RF model, the Clinical Dementia Rating Sum of Boxes (CDRSB) and Logical Memory Delayed Recall (LDELTOTAL) are overwhelmingly the most influential variables, with mean SHAP values far exceeding any other feature (Figure 6). The beeswarm plots confirm this at the instance level; for the AD class, high values of CDRSB (indicating greater impairment, colored red) consistently push the model’s prediction strongly positive, while low values (blue) push it negative (Figure 7). This provides a transparent and clinically intuitive explanation: the models learn to associate poor performance on key memory and functional tests directly with an AD-spectrum diagnosis.
Figure 6. Global feature importance for the Random Forest model with cognitive assessments.
Figure 7. SHAP beeswarm plot for the AD class using the Random Forest model with cognitive assessments.
When cognitive assessments are excluded, the models are forced to recalibrate their decision policies entirely, relying on a combination of imaging biomarkers and demographic features. In this scenario, the most influential predictors become markers of neurodegeneration. For the Random Forest model, PET biomarkers (FBB, AV45, FDG) and MRI volumetrics (Hippocampus, Entorhinal) rise to the top of the importance hierarchy, alongside the patient’s age (Figure 8). The SHAP beeswarm plot for the AD class (Figure 9) elegantly illustrates this shift in logic. It shows that low values (blue dots) for hippocampal volume push the prediction towards AD (positive SHAP values), while high values (red dots) push it away. This aligns perfectly with established neuropathology, where hippocampal atrophy is a well-known hallmark of Alzheimer’s disease.
Figure 8. Global feature importance for the Random Forest model without cognitive assessments.
Figure 9. SHAP beeswarm plot for the AD class using the Random Forest model without cognitive assessments.

4. Discussion

This study’s findings unequivocally demonstrate that cognitive assessment scores are the most powerful predictors for the machine learning-based classification of AD and its precursor stages. The dramatic degradation in performance observed when these functional measures were excluded, with the top-performing model’s accuracy falling from 84.4% to 76.1%, underscores the indispensable role of neuropsychological testing in accurately staging cognitive decline. While biological markers provide an essential foundation, the functional impairment captured by cognitive assessments is paramount for achieving diagnostic precision with computational models.
A central contribution of this work is the use of XAI to move beyond predictive accuracy and provide crucial transparency into the models’ decision-making processes. The application of SHAP confirmed that the models learned clinically intuitive and valid patterns from the data. When available, functional metrics such as the CDRSB and memory recall tests like LDELTOTAL overwhelmingly dominated the models’ logic, confirming that a patient’s ability to perform cognitive tasks is the most salient feature for classification. This contrasts with prior “black box” approaches that, despite high accuracy, offered no insight into their reasoning, limiting clinical trust and applicability.
In the absence of cognitive data, the models intelligently pivoted to a biological signature of AD. Delving deeper into the biomarker-driven model, the top features represent the core pathological tenets of the disease. PET tracers that quantify amyloid plaque burden (FBB, AV45) and cerebral glucose metabolism (FDG) rose to the top of the feature importance hierarchy. The models correctly learned to associate high amyloid burden and low glucose metabolism with an increased likelihood of an AD diagnosis. Similarly, structural MRI features indicating atrophy, such as a smaller Hippocampus and Entorhinal cortex volume, were also learned as significant predictors. This analysis confirms that the “no-cognition” model is not a random guesser but a rational agent that successfully learned the fundamental biological and structural signatures of AD from the data, validating its logic against established neuropathology.
The superior performance of the Random Forest classifier likely stems from its ensemble architecture, which is adept at capturing the complex, nonlinear interactions inherent in the data and is less prone to overfitting than a single deep network might be. A detailed analysis of the error structure across all models reveals the inherent challenges of applying discrete labels to a continuous disease process. The most persistent error was the misclassification of SMC as CN, evident in the low recall for the SMC class. This suggests that the captured features may be insufficient to reliably separate subjective complaints from objective cognitive performance. Furthermore, without cognitive data, all models struggled to differentiate EMCI from its adjacent classes, indicating that biological data alone may lack the resolution to parse subtle, transitional disease stages.
The robustness of these findings is supported by key methodological choices, particularly the use of class weights to mitigate the effects of the imbalanced class distribution present in the dataset. Without this correction, the models would have been heavily biased toward the more populous CN and LMCI classes, further obscuring the true performance on rarer diagnostic groups.
While this study provides valuable insights, its cross-sectional design, using only baseline data, is a key limitation. Future research should prioritize longitudinal modeling to capture disease progression and transitions between stages over time. Such models could offer predictive power not only for current diagnosis but also for a patient’s future trajectory. Furthermore, integrating additional data modalities, such as detailed CSF analytics, emerging blood-based biomarkers, or more advanced model architectures, could enhance accuracy, particularly in the ambiguous early stages of the disease. Ultimately, this work demonstrates that the synergy between high-performing models and robust XAI methods holds immense promise for developing the next generation of clinical decision-support tools that are not only accurate but also transparent and trustworthy.

5. Conclusions

By evaluating four machine learning architectures, this study highlights the inherent challenge of applying discrete classification labels to the AD continuum. Our findings confirm that cognitive assessment scores are the most powerful predictors for this task, with a Random Forest model achieving the highest accuracy at 84.4%. The central contribution of this work is demonstrating that an XAI approach can provide clinically valid insights into a model’s logic. SHAP analysis revealed that the models intelligently pivoted from relying on functional performance metrics to the core biological signatures of AD, such as amyloid burden and hippocampal atrophy, when cognitive data was absent. This synergy between predictive accuracy and validated, transparent reasoning is a crucial paradigm for developing the next generation of trustworthy clinical decision-support tools.

Author Contributions

Conceptualization, C.-H.H. and A.U.; methodology, C.-H.H. and F.A.B.; validation, C.-H.H.; formal analysis, C.-H.H.; resources, C.-H.H.; data curation, C.-H.H.; writing—original draft preparation, C.-H.H.; writing—review and editing, A.U. and F.A.B.; visualization, C.-H.H.; supervision, A.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The datasets analyzed during the current study are sourced from the ADNI and are available on the ADNI website for approved researchers. The processed dataset used in this study is not publicly available due to restrictions on sharing derived data but can be obtained from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADAlzheimer’s Disease
ADNIAlzheimer’s Disease Neuroimaging Initiative
CNCognitively Normal
CVCross-Validation
EMCIEarly Mild Cognitive Impairment
FNNFeed-Forward Neural Network
LMCILate Mild Cognitive Impairment
LRLogistic Regression
MCIMild Cognitive Impairment
MRIMagnetic Resonance Imaging
PETPositron Emission Tomography
RBFRadial-Basis-Function
RFRandom Forest
SHAPSHapley Additive exPlanations
SMCSubjective Memory Concerns
SVMSupport Vector Machine
XAIExplainable AI
XGBoosteXtreme Gradient Boosting

References

  1. Alzheimer’s Association. 2023 Alzheimer’s Disease Facts and Figures. Alzheimer’s Dement. 2023, 19, 1598–1695. [Google Scholar] [CrossRef] [PubMed]
  2. Bertram, L.; Tanzi, R.E. Alzheimer Disease Risk Genes: 29 and Counting. Nat. Rev. Neurol 2019, 15, 191–192. [Google Scholar] [CrossRef] [PubMed]
  3. Cruts, M.; Van Broeckhoven, C. Molecular Genetics of Alzheimer’s Disease. Ann. Med. 1998, 30, 560–565. [Google Scholar] [CrossRef] [PubMed]
  4. Karch, C.M.; Goate, A.M. Alzheimer’s Disease Risk Genes and Mechanisms of Disease Pathogenesis. Biol. Psychiatry 2015, 77, 43–51. [Google Scholar] [CrossRef] [PubMed]
  5. Szabo, M.P.; Mishra, S.; Knupp, A.; Young, J.E. The Role of Alzheimer’s Disease Risk Genes in Endolysosomal Pathways. Neurobiol. Dis. 2022, 162, 105576. [Google Scholar] [CrossRef] [PubMed]
  6. Jack, C.R.; Bennett, D.A.; Blennow, K.; Carrillo, M.C.; Dunn, B.; Haeberlein, S.B.; Holtzman, D.M.; Jagust, W.; Jessen, F.; Karlawish, J.; et al. NIA-AA Research Framework: Toward a Biological Definition of Alzheimer’s Disease. Alzheimer’s Dement. 2018, 14, 535–562. [Google Scholar] [CrossRef] [PubMed]
  7. Petersen, R.C.; Aisen, P.S.; Beckett, L.A.; Donohue, M.C.; Gamst, A.C.; Harvey, D.J.; Jack, C.R.; Jagust, W.J.; Shaw, L.M.; Toga, A.W.; et al. Alzheimer’s Disease Neuroimaging Initiative (ADNI): Clinical Characterization. Neurology 2010, 74, 201–209. [Google Scholar] [CrossRef] [PubMed]
  8. Zhao, Z.; Chuah, J.H.; Lai, K.W.; Chow, C.-O.; Gochoo, M.; Dhanalakshmi, S.; Wang, N.; Bao, W.; Wu, X. Conventional Machine Learning and Deep Learning in Alzheimer’s Disease Diagnosis Using Neuroimaging: A Review. Front. Comput. Neurosci. 2023, 17, 1038636. [Google Scholar] [CrossRef] [PubMed]
  9. Antoniadi, A.M.; Du, Y.; Guendouz, Y.; Wei, L.; Mazo, C.; Becker, B.A.; Mooney, C. Current Challenges and Future Opportunities for XAI in Machine Learning-Based Clinical Decision Support Systems: A Systematic Review. Appl. Sci. 2021, 11, 5088. [Google Scholar] [CrossRef]
  10. Batarseh, F.A.; Freeman, L.; Huang, C.-H. A Survey on Artificial Intelligence Assurance. J. Big Data 2021, 8, 60. [Google Scholar] [CrossRef]
  11. Yousefi, M.; Akhbari, M.; Mohamadi, Z.; Karami, S.; Dasoomi, H.; Atabi, A.; Sarkeshikian, S.A.; Abdoullahi Dehaki, M.; Bayati, H.; Mashayekhi, N.; et al. Machine Learning Based Algorithms for Virtual Early Detection and Screening of Neurodegenerative and Neurocognitive Disorders: A Systematic-Review. Front. Neurol. 2024, 15, 1413071. [Google Scholar] [CrossRef] [PubMed]
  12. Martin, S.A.; Townend, F.J.; Barkhof, F.; Cole, J.H. Interpretable Machine Learning for Dementia: A Systematic Review. Alzheimer’s Dement. 2023, 19, 2135–2149. [Google Scholar] [CrossRef] [PubMed]
  13. Nguyen, K.; Nguyen, M.; Dang, K.; Pham, B.; Huynh, V.; Vo, T.; Ngo, L.; Ha, H. Early Alzheimer’s Disease Diagnosis Using an XG-Boost Model Applied to MRI Images. Biomed. Res. Ther. 2023, 10, 5896–5911. [Google Scholar] [CrossRef]
  14. Bloch, L.; Friedrich, C.M. Machine Learning Workflow to Explain Black-Box Models for Early Alzheimer’s Disease Classification Evaluated for Multiple Datasets. SN Comput. Sci. 2022, 3, 509. [Google Scholar] [CrossRef]
  15. Dubey, Y.; Bhongade, A.; Palsodkar, P.; Fulzele, P. Efficient Explainable Models for Alzheimer’s Disease Classification with Feature Selection and Data Balancing Approach Using Ensemble Learning. Diagnostics 2024, 14, 2770. [Google Scholar] [CrossRef] [PubMed]
  16. Orovas, C.; Orovou, E.; Dagla, M.; Daponte, A.; Rigas, N.; Ougiaroglou, S.; Iatrakis, G.; Antoniou, E. Neural Networks for Early Diagnosis of Postpartum PTSD in Women after Cesarean Section. Appl. Sci. 2022, 12, 7492. [Google Scholar] [CrossRef]
  17. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 1135–1144. [Google Scholar]
  18. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning—Nature. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  19. Mueller, S.G.; Weiner, M.W.; Thal, L.J.; Petersen, R.C.; Jack, C.R.; Jagust, W.; Trojanowski, J.Q.; Toga, A.W.; Beckett, L. Ways toward an Early Diagnosis in Alzheimer’s Disease: The Alzheimer’s Disease Neuroimaging Initiative (ADNI). Alzheimer’s Dement. 2005, 1, 55–66. [Google Scholar] [CrossRef] [PubMed]
  20. Weiner, M.W.; Veitch, D.P.; Aisen, P.S.; Beckett, L.A.; Cairns, N.J.; Green, R.C.; Harvey, D.; Jack, C.R.; Jagust, W.; Morris, J.C.; et al. The Alzheimer’s Disease Neuroimaging Initiative 3: Continued Innovation for Clinical Trial Improvement. Alzheimer’s Dement. 2017, 13, 561–571. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.