Next Article in Journal
Counting Limb Length Ratios in Roux-en-Y Gastric Bypass: A Demonstration of Safety and Feasibility Using a 25-Patient Case Series in a High-Volume Academic Center
Previous Article in Journal
Intestinal Microbiota and Fecal Transplantation in Patients with Inflammatory Bowel Disease and Clostridioides difficile: An Updated Literature Review
Previous Article in Special Issue
Deep Brain Stimulation Combined with NMDA Antagonist Therapy in the Treatment of Alzheimer’s Disease: In Silico Trials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AI-Based Classification of Mild Cognitive Impairment and Cognitively Normal Patients

by
Rafail Christodoulou
1,
Giorgos Christofi
2,*,
Rafael Pitsillos
3,
Reina Ibrahim
4,
Platon Papageorgiou
5,
Sokratis G. Papageorgiou
6,
Evros Vassiliou
7 and
Michalis F. Georgiou
8
1
Department of Radiology, Stanford University School of Medicine, Stanford, CA 94305, USA
2
Faculty of Electrical Engineering, Mathematics and Computer Science, Delft University of Technology, 2628 Delft, The Netherlands
3
Department of Neurophysiology, The Cyprus Institute of Neurology and Genetics, Nicosia 2371, Cyprus
4
Faculty of Medicine, University of Balamand, Balamand 2807, Lebanon
5
2nd Department of Orthopaedic Surgery and Traumatology, Aghia Sophia Pediatric General Hospital, Thivon 3 Street, 15772 Athens, Greece
6
1st Department of Neurology, Medical School, National and Kapodistrian University of Athens, Eginition Hospital, 15772 Athens, Greece
7
Department of Biological Sciences, Union, NJ 07083, USA
8
Department of Radiology, Division of Nuclear Medicine, University of Miami, Miami, FL 33136, USA
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2025, 14(15), 5261; https://doi.org/10.3390/jcm14155261
Submission received: 23 June 2025 / Revised: 10 July 2025 / Accepted: 17 July 2025 / Published: 25 July 2025

Abstract

Background: Mild Cognitive Impairment (MCI) represents an intermediate stage between normal cognitive aging and Alzheimer’s Disease (AD). Early and accurate identification of MCI is crucial for implementing interventions that may delay or prevent further cognitive decline. This study aims to develop a machine learning-based model for differentiating between Cognitively Normal (CN) individuals and MCI patients using data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI). Methods: An ensemble classification approach was designed by integrating Extra Trees, Random Forest, and Light Gradient Boosting Machine (LightGBM) algorithms. Feature selection emphasized clinically relevant biomarkers, including Amyloid-β 42, phosphorylated tau, diastolic blood pressure, age, and gender. The dataset was split into training and held-out test sets. A probability thresholding strategy was employed to flag uncertain predictions for potential deferral, enhancing model reliability in borderline cases. Results: The final ensemble model achieved an accuracy of 83.2%, a recall of 80.2%, and a precision of 86.3% on the independent test set. The probability thresholding mechanism flagged 23.3% of cases as uncertain, allowing the system to abstain from low-confidence predictions. This strategy improved clinical interpretability and minimized the risk of misclassification in ambiguous cases. Conclusions: The proposed AI-driven ensemble model demonstrates strong performance in classifying MCI versus CN individuals using multimodal ADNI data. Incorporating a deferral mechanism through uncertainty estimation further enhances the model’s clinical utility. These findings support the integration of machine learning tools into early screening workflows for cognitive impairment.

1. Introduction

Mild cognitive Impairment (MCI) can be defined as the in-between stage of dementia and normal aging, reflecting an increased risk of Alzheimer’s Disease (AD). Mild Cognitive Impairment indicates a preclinical phase characterized by mild memory loss that exceeds what is expected from normal aging but does not yet meet the criteria for dementia [1]. The World Health Organization reported in 2021 that 57 million people suffer from dementia, a number expected to triple in the next 30 years [2]. A recent systematic review, which included 66 studies, demonstrated an overall MCI prevalence of around 15%. The same study showed a 10% increase in the prevalence of MCI as the age of individuals progresses from 60 to 80 years of life [3].
The criteria that have to be fulfilled for MCI characterization include complaints of memory impairment by the individual, with intact cognition and no interference of this current state with daily life activities [4]. Although not all MCI cases progress to AD, early identification and monitoring are critical for timely interventions. The required diagnostic evaluation for MCI can be roughly categorized into a subjective and an objective assessment, with the first referring to the history taking, focusing on cognitive function, and the second referring to a set of widely accepted neuropsychological tests; for instance, Mini Mental State Examination (MMSE) [5]. Further investigation for MCI includes cerebrospinal fluid (CSF) analysis, for detection of Aβ1–42 and phosphorylated tau (p-tau) increased levels, reflecting the presence of senile amyloid plaques and neurofibrillary tangles in the brain, and predicting the MCI conversion to Alzheimer’s Disease [6]. In addition to CSF analysis, neuroimaging contributes significantly to MCI detection, with MRI playing a pivotal role in identifying atrophy patterns in key brain regions, revealing candidate biomarkers [7]. These traditional diagnostic methods offer valuable insights but are resource-intensive, expensive, and subject to inter-clinician variability.
The necessity for developing more generalized approaches that provide better reproducibility led to the implication of artificial intelligence (AI) in neurodegenerative disease detection. Machine learning (ML) has evolved the imaging processing and interpretation in dementia research and clinical practice, by facilitating automation and bias reduction in the diagnostic procedure [8]. Regarding ML in disease detection, the models require labeled data for training to accurately identify patterns within novel input data [9]. Several studies utilizing ML models have been developed in recent years [10,11,12]. Deep learning (DL) technology, a subfield of ML, has also been widely explored for MRI and PET-based diagnosis and classification [13,14,15]. DL-based techniques identify patterns distributed across regions of interest, utilizing deep neural networks, without requiring already processed input [16]. The reliance of the DL models on large datasets and the computational complexity required for image analysis limit their applicability in smaller clinical settings. Instead, structured biomarker-based models offer a computationally efficient alternative.
Greater performance and higher levels of accuracy can be achieved by utilizing ensemble techniques instead of individual technologies, providing a fusion of the outcome prediction [17]. Various studies have developed and demonstrated ensemble models for the detection, classification, or prediction of Alzheimer’s Disease, increasing the robustness and accuracy of the established technologies [18,19,20]. These techniques showed great performance; nonetheless, the vast majority of the studies emphasize in neuroimaging application of ensemble models.
Our study aims to develop an ensemble machine learning (ML) model to accurately classify Cognitively Normal (CN) vs. Mild Cognitive Impairment (MCI), using biomarker data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database. Cross-validation was conducted for the selection of the best—performing models for the ensemble, and a probability-based classification threshold was introduced to enhance prediction confidence and reduce misclassification errors. This addition has clinical relevance by either lowering the threshold to support earlier detection or raising it to improve recruitment accuracy for clinical studies. Feature variance was estimated in order to identify the most predictive biomarkers for MCI classification, improving both interpretability and clinical applicability of the proposed model. Thereby, the main aims of this work were to design a robust ensemble ML model optimized via cross-validation for early MCI classification, identify the key biomarkers contributing to model predictions, and apply a probability-based threshold to improve classification reliability. Hence, the scope of the study focused on binary classification (MCI vs. CN) using biomarkers from the ADNI dataset, with an emphasis on interpretability and potential clinical translation.

2. Materials and Methods

2.1. Participants and Data Source

Data used in this study were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (http://adni.loni.usc.edu, accessed on 9 March 2025). The ADNI project was launched in 2003 as a public–private partnership with the primary goal of testing whether clinical, imaging, genetic, and biochemical biomarkers can be combined to measure the progression of Mild Cognitive Impairment (MCI) and early Alzheimer’s Disease (AD). All participants gave written informed consent at the time of enrollment for data collection and sharing. The study protocols and consent forms were approved by the institutional review boards (IRBs) of each participating institution.
For this analysis, we extracted demographic, clinical, and biochemical data from the ADNI repository. Only subjects classified as Cognitively Normal (CN) or any subtype of Mild Cognitive Impairment (MCI, EMCI, LMCI) were included. Subjects diagnosed with Alzheimer’s Disease were excluded to enable binary classification between CN and MCI. No neuroimaging or genetic modalities (e.g., APOE genotype) were used in model development.
Participants diagnosed with early cognitive impairment (EMCI), Mild Cognitive Impairment (MCI), and late MCI (LMCI) were grouped into a single MCI class. Subjects diagnosed with Alzheimer’s Disease (AD) were excluded to restrict the task to binary classification between Cognitively Normal (CN) and MCI individuals.
The dataset contains 3029 entries from the ADNI database, each representing a patient visit record. It includes demographic data (AGE, GENDER), vital signs (VSBPSYS, VSBPDIA, VSPULSE, and VSRESP), biochemical biomarkers (ABETA40, ABETA42, PTAU, TAU, GLUCOSE, PROTEIN, CTWHITE, CTRED, PROTEIN, GLUCOSE, and GENOTYPE), and administrative fields (SUBJECT_ID, NVISIT, VISIT, and SUBJECT_DATE). There is also the RESEARCH_GROUP column, which is for diagnostic labels (CN, MCI, EMCI, LMCI, and AD). Amyloid-β 40 (ABETA40) was excluded due to excessive missingness with over 2100 absent values. Samples with any remaining missing data were then removed, resulting in a filtered dataset containing 1313 MCI cases and 986 CN cases. To balance the classes, 327 MCI samples were randomly removed, producing a final dataset of 1972 subjects that was evenly distributed between CN and MCI groups, 986 subjects per class.
The number of classes after each stage is shown in Table 1.

2.2. Feature Selection and Preprocessing

Features retained for analyses included demographic variables such as age and gender, vital signs, systolic blood pressure (VSBPSYS), diastolic blood pressure (VSBPDIA), pulse rate (VSPULSE), and respiratory rate (VSRESP). Moreover, biochemical measurements such as amyloid-β 42 (ABETA42), tau, phosphorylated tau (ptau), CSF protein and glucose, and CSF red blood cells (CTRED) and white blood cells (CTWHITE). Administrative fields such as subject ID, visit number, and visit date were excluded. The APOE genotype variable was also removed to reduce model cost and complexity, as it is one of the most expensive biomarkers.
Gender was encoded as a binary feature, meaning that male subjects were encoded as 1, female subjects as 2, and all continuous variables were scaled to the [0,1] interval using the Min–Max Scaler function from the scikit learn Python (version 3.12.2) library.
Feature variance was calculated to identify the most informative inputs and was used as the selection criterion. This approach assumes that features with the highest variability are more informative for classification tasks. The five features with the highest variance were gender (0.250), age (0.036), diastolic blood pressure (0.020), ABETA42 (0.019), and ptau (0.018). These five features were selected for model development.
A complete summary of variance across features is shown in Table 2.

2.3. Model Development

A total of eleven supervised machine learning classifiers were trained using the five selected features: ptgender, age, diastolic blood pressure, ABETA42, and ptau. These models included both traditional and advanced algorithms, such as Naive Bayes, Logistic Regression, K-Nearest Neighbors, Support Vector Machines, Random Forest, Extra Trees, XGBoost, CatBoost, LightGBM, Histogram-Based Gradient Boosting, and a Multi-layer Perceptron. Each model was trained using a stratified 70/30 split and optimized through 5-fold cross-validation with grid search over relevant hyperparameters.
Following cross-validation, the three highest-performing models were Extra Trees, LightGBM, and Random Forest and were selected for final evaluation and integration into an ensemble. These models were chosen due to their superior accuracy when evaluated on the validation set. Three models were chosen (as opposed to two) since this reduces the number of uncertain predictions due to ties in the majority voting process. Notably, they were all tree-based classifiers known for strong performance on tabular data. The ensemble strategy involved majority voting with a probability threshold to improve reliability by abstaining on uncertain predictions.

3. Results

3.1. Cross-Validation Performance

All models were implemented in Python using the scikit learn library and evaluated on the test set to ensure robustness, accuracy, and generalization. To identify the most promising models, all 11 classifiers were evaluated using stratified five-fold cross-validation on the training set on 70% of the full dataset. Hyperparameter tuning was performed via grid search, and performance was assessed using average validation accuracy across folds.
The results are summarized in Table 3.
These results demonstrate that tree-based ensemble methods consistently outperformed traditional classifiers such as SVM, Naive Bayes, and Logistic Regression. While the Multi-layer Perceptron achieved relatively high best accuracy, its mean accuracy and variance were less stable compared to the top models.
The best performing models based on mean cross-validation accuracy were Extra Trees with an accuracy of 73.2%, Random Forest with an accuracy of 72.3%, and LightGBM with an accuracy of 71.6%. Their respective optimal hyperparameters were: for Extra Trees, max_depth: 18 and n_estimators: 200; for Random Forest, max_depth: None, min_samples_split: 2, and n_estimators: 200; and for LightGBM, learning_rate: 0.1, min_samples_leaf: 2, n_estimators: 100, and num_leaves: 100. These models were selected for subsequent testing on the hold-out validation set.

3.2. Evaluation on Hold Out Test Set

To assess model generalizability on unseen data, the three best-performing classifiers from cross-validation, Extra Trees, LightGBM, and Random Forest, were evaluated on the 30% hold-out test set that was not used during training or model selection.
As shown in Table 4, all models performed similarly on the test set, with Extra Trees achieving the highest test accuracy of 0.728 while both LightGBM and Random Forest have accuracies of 0.720. Extra Trees achieved the highest overall performance, with an accuracy of 0.728, an F1 score of 0.736, a precision of 0.757, and a recall of 0.716. The Random Forest model had comparable performance with LightGBM but was a bit worse than Extra Trees.
To improve classification reliability, a majority voting ensemble strategy was implemented. A sample was classified only if at least two of the three base models predicted the same label with an average class probability exceeding a predefined threshold pp. By varying p from 0.5 to 1.0, the ensemble allowed flexible control over the balance between precision and recall. As shown in Figure 1, increasing the threshold resulted in higher precision and overall reliability, as the model abstained from making predictions on uncertain cases. However, this came at the cost of reduced recall, since more true positives were left unclassified at higher thresholds.
At a threshold of p equal to 0.90, the model achieved its most effective balance between performance and safety, with a precision of 86.3%, a recall of 80.2%, an F1 score of 83.1%, and an abstention rate of 43.6%. Importantly, 90.5% of all predictions were either correct or safely withheld due to low confidence, as shown in Table 5. This approach is particularly beneficial in clinical settings where the consequences of incorrect classification can be significant. By abstaining from ambiguous cases, the model minimizes the risk of false positives or false negatives, allowing those uncertain samples to be flagged for further clinical evaluation. Such a framework supports safer decision-making and can serve as an effective triage tool in diagnostic workflows.
The final ensemble model, integrating the Extra Trees, LightGBM, and Random Forest classifiers, demonstrated improved overall performance by leveraging agreement among high-performing tree-based algorithms. The model only takes into consideration predictions with a probability greater than or equal to p, and classifies them only if there is a majority label among those.
This conservative, abstention-aware approach helps mitigate the risk of misclassification, particularly in ambiguous or borderline cases. By deferring low-confidence predictions, the system aligns with clinical needs where incorrect classifications may lead to unnecessary interventions or missed diagnoses. These results underscore the ensemble’s potential as a decision support tool that balances precision with safety in diagnostic workflows. Importantly, the confidence threshold p is not fixed, and it can be dynamically adjusted depending on the clinical use case. For instance, in high-risk populations or screening scenarios where sensitivity is critical, a lower threshold may be preferred to reduce missed diagnoses. Conversely, in settings where false positives must be minimized, a higher threshold may be more appropriate. This tunability allows clinicians or healthcare institutions to adapt the model’s behavior to their specific risk tolerance and diagnostic strategy, enhancing its real-world utility.

3.3. Comparison with Additional Classifiers

In addition to the ensemble approach, a total of eleven individual machine learning classifiers were trained and evaluated, including tree-based models, boosting methods, and traditional algorithms such as SVM, Naive Bayes, Logistic Regression, and K-Nearest Neighbors, as shown in Table 2. While several models demonstrated competitive accuracy, particularly Extra Trees with an accuracy of 72.47% and LightGBM with an accuracy of 71.96% on the hold-out test set, the ensemble model outperformed all individual classifiers in terms of overall balance across metrics. It achieved a higher F1 score of 78.0% and improved precision by 79.6% while maintaining a solid recall of 76.5%, with the added benefit of a built-in abstention mechanism.
Unlike standalone classifiers, the ensemble approach was able to identify and withhold uncertain predictions, reducing the risk of false positives and negatives in ambiguous cases. This added layer of reliability is particularly valuable in clinical applications, where incorrect classifications can have serious consequences. By combining the strengths of multiple high-performing models and deferring uncertain decisions, the ensemble strategy offers a robust, safety-oriented framework for real-world diagnostic support.
Overall, the ensemble model demonstrated the strongest classification performance while at the same time offering a threshold-based mechanism to minimize confident misclassification in borderline cases and suggest to the physicians that more tests are needed.

4. Discussion

This study demonstrates the efficacy of an ensemble AI-based biomarker-driven approach in classifying CN vs. MCI patients. A total of 1972 entries from ADNI were used, equally allocated to 986 MCI patients and 986 CN cases. A significant feature that led to the exclusion of many cases was ABETA40. Multiple variables were subsequently examined to identify the most appropriate features that could be utilized for our ensemble model. Although apolipoprotein E ε4 (APOEε4) allele, located in chromosome 19, has a very strong correlation with disease risk and onset age of “sporadic” Alzheimer’s Disease [21], it was excluded from our model’s features, as the complexity regarding disease causation and prediction risk [22], as well as its high cost, disqualify it as a viable option. Various other factors were examined, ultimately resulting in the five features with the highest variance that indicate significance in the prediction: gender, age, diastolic blood pressure, ABETA42, and ptau. These features have been extensively investigated by scholars, characterizing their contribution to MCI and hence AD development. Women have a higher risk of developing AD, as demonstrated by Azad et al. in a 2007 study [23]. Terry et al, in 1983 identified an increased number of amyloid plaques in individuals above 75 years old [24], proposing an association between AD and advanced age. The detection of ABETA42 in plasma and CSF is shown to be an accurate marker for differentiating AD from other forms of dementia [25], and detection of high levels of Aβ42 and hyperphosphorylated tau protein in CSF is associated with increased load of AD histopathological hallmarks in the brain [26]. Ruiz et al. in 2013 correlated high DBP with Aβ40 levels, thereby linking it to an increased incidence of AD [27]. Consequently, unlike deep learning models requiring large imaging datasets, our model relies on structured biomarker data, making it computationally efficient and accessible for clinical use.

4.1. Comparison with Existing Approaches

Our ensemble model is composed of three machine learning technologies, which demonstrated the highest performance when cross-validation was conducted on all eleven classifiers that were proposed. The evaluation of the classifiers showed a 73.91% best accuracy of Extra Trees, 72.17% of Random Forest, and 71.59% accuracy of LightGBM, outperforming traditional models that are most frequently implemented in AI-based disease prediction, like SVM and Logistic Regression. To enhance the classification reliability among CN and MCI patients, a majority voting ensemble was employed, requiring agreement among base models and a minimum confidence threshold (p). Hence, by increasing the p value, precision and reliability improve, as well as uncertainty reduces. The equilibrium between performance and safety was observed when p was set at 0.94. Previous models demonstrated classification accuracies ranging from 70 to 80% for CN vs. MCI detection using MRI features. Some characteristic studies conducted in the last five years are [19,28,29]. Therefore, our biomarker-driven model achieves a comparable accuracy of 78.2% without requiring costly neuroimaging, making it a practical alternative for resource-limited settings.
The studies with the most significant contribution σ to the field have also been described in a valuable meta-analysis performed by Battineni and colleagues in 2024, where 24 studies were included, demonstrating the most powerful tools for detecting AD progression [30]. More recently, Rehman Ur-Zia et al. conducted an informative systematic review on AD prediction using neuroimaging modalities as input for deep learning technology, a subfield of machine learning that utilizes artificial neural networks, able to predict patterns after training with raw, unprocessed data [31]. Both studies highlight the significance and applicability of our work in the context of artificial intelligence in Alzheimer’s Disease prodromal stage classification.

4.2. Clinical Implications

A key contribution of our study is the clinical relevance of the proposed model. By integrating uncertainty awareness, this model provides clinicians with interpretable risk assessments, allowing them to prioritize uncertain cases for further evaluation. The implementation of probability-based classification thresholds reduces the risk of misclassification, making the model more reliable for early-stage MCI detection. Two hypothetical scenarios can enhance the clinical relevance of this model, demonstrating its implications in both early detection clinic settings and in research centers. In the first case, a patient who has ambiguous cognitive alterations, mainly focusing on mild memory impairment, a strong family history, and a slight reduction in cognitive scales, arrives in the clinic, and our ensemble model is applied while the MRI is pending. The importance of confidence threshold (p) adjustment is that in early detection dementia clinics, the reduction in missed diagnoses is vital for minimizing the burden of dementia; hence, lowering the threshold achieves complete sensitivity. Alternatively, in a research study regarding clinical trials, which requires MCI patients’ recruitment, increasing the confidence threshold (p) leads to an elevated specificity in the sample, hence reducing the falsely positive cases. This can also be applied in high-cost intervention settings where avoiding overdiagnosis is crucial. In clinical practice, cases with prediction probabilities near the threshold—flagged as uncertain—could prompt additional assessments, such as follow-up imaging, cognitive testing, and closer clinical monitoring. Integrating this uncertainty-aware approach into diagnostic workflows may support more cautious decision-making and help prioritize resources for borderline cases, aligning with current practices that emphasize risk assessment and individualized follow-up.

4.3. Limitations and Future Work

Despite the promising results, certain limitations should be acknowledged. The model lacks external validation, and future studies should test it on an independent dataset to assess its generalizability across different populations. For instance, the ADNI database lacks in incorporating younger individuals, as the age of participants ranges between 55 and 90 years old. This could be limiting in a study examining early stages of dementia, as described in [32]. Additionally, expanding feature selection to include MRI or PET scan biomarkers could further enhance classification performance by incorporating imaging-based insights into cognitive decline. A limitation of our work regarding the selection of the sample has to do with the removal of cases when data was missing, instead of applying imputation techniques. This cleaned our data and maintained the integrity of feature relationships without introducing imputation-related assumptions. However, this technique introduces, to some degree, a selection bias. Various studies have utilized fusion AI-based models to identify patterns and detect dementias at early stages, indicating the significance of multimodal techniques [33,34]. Regarding sensitivity and specificity of the model, false-positive results might emerge and be identified as MCI cases when the confidence threshold is low, and false-negative results, respectively, when the threshold is high. This could be potentially resolved by integrating an expanded feature set. Finally, for broader clinical translation, integrating the model into electronic health record (EHR) systems could enable real-time patient screening and decision support, facilitating early detection and timely intervention in clinical settings.

5. Conclusions

This study presents an AI-based classification model for early MCI detection, utilizing a combination of Extra Trees, Random Forest, and LightGBM classifiers with a probability-based thresholding mechanism. The adjustable threshold offers a key advantage over similar techniques, enabling its use across a wide range of settings from routine practice in community healthcare centers to large research hospitals, such as academic medical centers, while maintaining high accuracy. The overall accuracy and precision achieved by the model were 83.2% and 86.3%, respectively, demonstrating its potential for AI-assisted diagnostics. The application of this model aims to facilitate an earlier and more consistent MCI classification, based on routine metrics and biomarkers. Notably, certain limitations related to the subjects’ demographics, including age and ethnicity, slightly restrict the widespread applicability of the study. To build upon these results and further enhance the model’s clinical relevance, future research should focus on external validation—by employing independent cohorts from diverse geographic regions—and multimodal feature integration—by combining imaging, genetic, and clinical data—in order to enhance clinical applicability and reproducibility of the model.

Author Contributions

Conceptualization, R.C. and M.F.G.; investigation, R.C., R.P. and R.I.; methodology, R.C., G.C. and R.P.; project administration, M.F.G.; software, G.C.; supervision, S.G.P., E.V. and M.F.G.; visualization, P.P., S.G.P. and M.F.G.; writing—original draft, R.C., G.C., R.P., R.I. and P.P.; writing—review and editing, S.G.P., E.V. and M.F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. And the APC was funded by Delft University of Technology under its partnership with MDPI.

Institutional Review Board Statement

Ethical review and approval were waived for this study. The study used publicly available, de-identified data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI), and did not involve direct interaction with human participants.

Informed Consent Statement

Patient consent was waived. This study used de-identified data from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database and did not involve direct contact with human subjects.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data is available from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) at http://adni.loni.usc.edu.

Acknowledgments

The authors thank the Alzheimer’s Disease Neuroimaging Initiative (ADNI) for providing access to the dataset used in this study. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADAlzheimer’s Disease
ADNIAlzheimer’s Disease Neuroimaging Initiative
AIArtificial Intelligence
AUCArea Under the Curve
CNCognitively Normal
CSFCerebrospinal Fluid
DLDeep Learning
HERElectronic Health Record
EMCIEarly Mild Cognitive Impairment
FTPFlortaucipir
GBGradient Boosting
IA-ElcImmunoassay-Electrochemiluminescence
IP-MS-WashUImmunoprecipitation–Mass Spectrometry at Washington University
LMCILate Mild Cognitive Impairment
LLMLarge Language Model
MCIMild Cognitive Impairment
MLMachine Learning
MRIMagnetic Resonance Imaging
PETPositron Emission Tomography
ROIRegion of Interest
SPMStatistical Parametric Mapping
SVMSupport Vector Machine
SUVRStandardized Uptake Value Ratio

References

  1. Anderson, N.D. State of the science on mild cognitive impairment (MCI). CNS Spectr. 2019, 24, 78–87. [Google Scholar] [CrossRef] [PubMed]
  2. World Health Organization (WHO). Dementia. 2025. Available online: https://www.who.int/news-room/fact-sheets/detail/dementia (accessed on 30 April 2025).
  3. Bai, W.; Chen, P.; Cai, H.; Zhang, Q.; Su, Z.; Cheung, T.; Jackson, T.; Sha, S.; Xiang, Y.-T. Worldwide prevalence of mild cognitive impairment among community dwellers aged 50 years and older: A meta-analysis and systematic review of epidemiology studies. Age Ageing 2022, 51, afac173. [Google Scholar] [CrossRef] [PubMed]
  4. Petersen, R.C.; Smith, G.E.; Waring, S.C.; Ivnik, R.J.; Tangalos, E.G.; Kokmen, E. Mild cognitive impairment: Clinical characterization and outcome. Arch. Neurol. 1999, 56, 303–308. [Google Scholar] [CrossRef] [PubMed]
  5. Hugo, J.; Ganguli, M. Dementia and cognitive impairment: Epidemiology, diagnosis, and treatment. Clin. Geriatr. Med. 2014, 30, 421–442. [Google Scholar] [CrossRef] [PubMed]
  6. Lehmann, M.; Koedam, E.L.; Barnes, J.; Bartlett, J.W.; Barkhof, F.; Wattjes, M.P.; Schott, J.M.; Scheltens, P.; Fox, N.C. Visual ratings of atrophy in MCI: Prediction of conversion and relationship with CSF biomarkers. Neurobiol. Aging 2013, 34, 73–82. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, X.; Huang, W.; Su, L.; Xing, Y.; Jessen, F.; Sun, Y.; Shu, N.; Han, Y. Neuroimaging advances regarding subjective cognitive decline in preclinical Alzheimer’s disease. Mol. Neurodegener. 2020, 15, 55. [Google Scholar] [CrossRef] [PubMed]
  8. Borchert, R.J.; Azevedo, T.; Badhwar, A.; Bernal, J.; Betts, M.; Bruffaerts, R.; Burkhart, M.C.; Dewachter, I.; Gellersen, H.M.; Low, A.; et al. Artificial intelligence for diagnostic and prognostic neuroimaging in dementia: A systematic review. Alzheimer’s Dement. J. Alzheimer’s Assoc. 2023, 19, 5885–5904. [Google Scholar] [CrossRef] [PubMed]
  9. Jiang, T.; Gradus, J.L.; Rosellini, A.J. Supervised Machine Learning: A Brief Primer. Behav. Ther. 2020, 51, 675–687. [Google Scholar] [CrossRef] [PubMed]
  10. Lama, R.K.; Kwon, G.R. Diagnosis of Alzheimer’s Disease Using Brain Network. Front. Neurosci. 2021, 15, 605115. [Google Scholar] [CrossRef] [PubMed]
  11. Battineni, G.; Hossain, M.A.; Chintalapudi, N.; Traini, E.; Dhulipalla, V.R.; Ramasamy, M.; Amenta, F. Improved Alzheimer’s Disease Detection by MRI Using Multimodal Machine Learning Algorithms. Diagnostics 2021, 11, 2103. [Google Scholar] [CrossRef] [PubMed]
  12. Bao, Y.W.; Wang, Z.J.; Shea, Y.F.; Chiu, P.K.-C.; Kwan, J.S.; Chan, F.H.-W.; Mak, H.K.-F. Combined Quantitative amyloid-β PET and Structural MRI Features Improve Alzheimer’s Disease Classification in Random Forest Model—A Multicenter Study. Acad. Radiol. 2024, 31, 5154–5163. [Google Scholar] [CrossRef] [PubMed]
  13. El-Latif, A.A.A.; Chelloug, S.A.; Alabdulhafith, M.; Hammad, M. Accurate Detection of Alzheimer’s Disease Using Lightweight Deep Learning Model on MRI Data. Diagnostics 2023, 13, 1216. [Google Scholar] [CrossRef] [PubMed]
  14. Faisal, F.U.R.; Kwon, G.R. Automated Detection of Alzheimer’s Disease and Mild Cognitive Impairment Using Whole Brain MRI. IEEE Access 2022, 10, 65055–65066. [Google Scholar] [CrossRef]
  15. Basher, A.; Kim, B.C.; Lee, K.H.; Jung, H.Y. Volumetric Feature-Based Alzheimer’s Disease Diagnosis from sMRI Data Using a Convolutional Neural Network and a Deep Neural Network. IEEE Access 2021, 9, 29870–29882. [Google Scholar] [CrossRef]
  16. Chen, X.; Wang, X.; Zhang, K.; Fung, K.-M.; Thai, T.C.; Moore, K.; Mannel, R.S.; Liu, H.; Zheng, B.; Qiu, Y. Recent advances and clinical applications of deep learning in medical image analysis. Med. Image Anal. 2022, 79, 102444. [Google Scholar] [CrossRef] [PubMed]
  17. Guan, T.; Shang, L.; Yang, P.; Tan, Z.; Liu, Y.; Dong, C.; Li, X.; Hu, Z.; Su, H.; Zhang, Y. Joint ensemble learning-based risk prediction of Alzheimer’s disease among mild cognitive impairment patients. J. Prev. Alzheimer’s Dis. 2025, 12, 100083. [Google Scholar] [CrossRef] [PubMed]
  18. Mujahid, M.; Rehman, A.; Alam, T.; Alamri, F.S.; Fati, S.M.; Saba, T. An Efficient Ensemble Approach for Alzheimer’s Disease Detection Using an Adaptive Synthetic Technique and Deep Learning. Diagnostics 2023, 13, 2489. [Google Scholar] [CrossRef] [PubMed]
  19. Zhang, P.; Lin, S.; Qiao, J.; Tu, Y. Diagnosis of Alzheimer’s Disease with Ensemble Learning Classifier and 3D Convolutional Neural Network. Sensors 2021, 21, 7634. [Google Scholar] [CrossRef] [PubMed]
  20. Khan, Y.F.; Kaushik, B.; Chowdhary, C.L.; Srivastava, G. Ensemble Model for Diagnostic Classification of Alzheimer’s Disease Based on Brain Anatomical Magnetic Resonance Imaging. Diagnostics 2022, 12, 3193. [Google Scholar] [CrossRef] [PubMed]
  21. Corder, E.H.; Saunders, A.M.; Strittmatter, W.J.; Schmechel, D.E.; Gaskell, P.C.; Small, G.W.; Roses, A.D.; Haines, J.L.; Pericak-Vance, M.A. Gene dose of apolipoprotein E type 4 allele and the risk of Alzheimer’s disease in late onset families. Science 1993, 261, 921–923. [Google Scholar] [CrossRef] [PubMed]
  22. Caselli, R.J.; Beach, T.G.; Knopman, D.S.; Graff-Radford, N.R. Alzheimer Disease: Scientific Breakthroughs and Translational Challenges. Mayo Clin. Proc. 2017, 92, 978–994. [Google Scholar] [CrossRef] [PubMed]
  23. Azad, N.A.; Al Bugami, M.; Loy-English, I. Gender differences in dementia risk factors. Gend. Med. 2007, 4, 120–129. [Google Scholar] [CrossRef] [PubMed]
  24. Terry, R.D.; Katzman, R. Senile dementia of the Alzheimer type. Ann. Neurol. 1983, 14, 497–506. [Google Scholar] [CrossRef] [PubMed]
  25. Kokkinou, M.; Beishon, L.C.; Smailagic, N.; Noel-Storr, A.H.; Hyde, C.; Ukoumunne, O.; Worrall, R.E.; Hayen, A.; Desai, M.; Ashok, A.H.; et al. Plasma and cerebrospinal fluid ABeta42 for the differential diagnosis of Alzheimer’s disease dementia in participants diagnosed with any dementia subtype in a specialist care setting. Cochrane Database Syst. Rev. 2021, 2, CD010945. [Google Scholar] [CrossRef] [PubMed]
  26. Tapiola, T.; Alafuzoff, I.; Herukka, S.K.; Parkkinen, L.; Hartikainen, P.; Soininen, H.; Pirttilä, T. Cerebrospinal fluid {beta}-amyloid 42 and tau proteins as biomarkers of Alzheimer-type pathologic changes in the brain. Arch. Neurol. 2009, 66, 382–389. [Google Scholar] [CrossRef] [PubMed]
  27. Ruiz, A.; Pesini, P.; Espinosa, A.; Pérez-Grijalba, V.; Valero, S.; Sotolongo-Grau, O.; Alegret, M.; Monleón, I.; Lafuente, A.; Buendía, M.; et al. Blood amyloid beta levels in healthy, mild cognitive impairment and Alzheimer’s disease individuals: Replication of diastolic blood pressure correlations and analysis of critical covariates. PLoS ONE 2013, 8, e81334. [Google Scholar] [CrossRef] [PubMed]
  28. Jimenez-Mesa, C.; Illan, I.A.; Martin-Martin, A.; Castillo-Barnes, D.; Martinez-Murcia, F.J.; Ramirez, J.; Gorriz, J.M. Optimized one vs one approach in multiclass classification for early alzheimer’s disease and mild cognitive impairment diagnosis. IEEE Access 2020, 8, 96981–96993. [Google Scholar] [CrossRef]
  29. Gamal, A.; Elattar, M.; Selim, S. Automatic Early Diagnosis of Alzheimer’s Disease Using 3D Deep Ensemble Approach. IEEE Access 2022, 10, 115974–115987. [Google Scholar] [CrossRef]
  30. Battineni, G.; Chintalapudi, N.; Amenta, F. Machine learning driven by magnetic resonance imaging for the classification of Alzheimer disease progression: Systematic review and meta-analysis. JMIR Aging 2024, 7, e59370. [Google Scholar] [CrossRef] [PubMed]
  31. Zia-Ur-Rehman, Z.; Awang, M.K.; Ali, G.; Faheem, M. Recent advancements in neuroimaging-based Alzheimer’s disease prediction using deep learning approaches in e-health: A systematic review. Health Sci. Rep. 2025, 8, e70802. [Google Scholar] [CrossRef] [PubMed]
  32. Weiner, M.W.; Veitch, D.P.; Aisen, P.S.; Beckett, L.A.; Cairns, N.J.; Cedarbaum, J.; Donohue, M.C.; Green, R.C.; Harvey, D.; Jack, C.R.; et al. Impact of the Alzheimer’s Disease Neuroimaging Initiative, 2004 to 2014. Alzheimer’s Dement. J. Alzheimer’s Assoc. 2015, 11, 865–884. [Google Scholar] [CrossRef] [PubMed]
  33. Gupta, Y.; Kim, J.I.; Kim, B.C.; Kwon, G.R. Classification and Graphical Analysis of Alzheimer’s Disease and Its Prodromal Stage Using Multimodal Features From Structural, Diffusion, and Functional Neuroimaging Data and the APOE Genotype. Front. Aging Neurosci. 2020, 12, 238. [Google Scholar] [CrossRef] [PubMed]
  34. Wu, J.; Dong, Q.; Zhang, J.; Su, Y.; Wu, T.; Caselli, R.J.; Reiman, E.M.; Ye, J.; Lepore, N.; Chen, K.; et al. Federated Morphometry Feature Selection for Hippocampal Morphometry Associated Beta-Amyloid and Tau Pathology. Front. Neurosci. 2021, 15, 762458. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Model performance metrics as a function of voting probability threshold p (line graph showing accuracy, precision, recall, and F1 score across varying p values).
Figure 1. Model performance metrics as a function of voting probability threshold p (line graph showing accuracy, precision, recall, and F1 score across varying p values).
Jcm 14 05261 g001
Table 1. Number of subjects per diagnostic category at each preprocessing stage.
Table 1. Number of subjects per diagnostic category at each preprocessing stage.
Raw Data Grouped—Unbalanced Grouped—Balanced
Research GroupCountResearch GroupCountResearch GroupCount
CN1125MCI1313MCI986
MCI750CN986CN986
EMCI475
AD397
LMCI282
Table 2. Feature variance values.
Table 2. Feature variance values.
FeatureVariance
PTGENDER0.2497
AGE0.0357
VSBPDIA0.0202
ABETA420.0191
PTAU0.0176
TAU0.0175
VSBPSYS0.0163
VSRESP0.0099
VSPULSE0.0047
CTRED0.0067
CTWHITE0.0035
PROTEIN0.0030
GLUCOSE0.0014
Table 3. Cross-validation accuracy of all classifiers.
Table 3. Cross-validation accuracy of all classifiers.
ModelBest AccuracyMean AccuracyStandard Deviation
Extra Trees0.73190.72150.0058
Random Forest0.72320.71090.0063
LightGBM0.71590.70040.0085
CatBoost0.71450.70200.0073
Histogram GB0.71450.69220.0087
XGBoost0.71230.69640.0084
SVM0.69860.64420.0436
Multi-layer Perceptron0.69350.67040.0134
Logistic Regression0.66880.66730.0006
K-Nearest Neighbors0.66880.65810.0061
Naive Bayes0.66450.66450.0000
Table 4. Model performance metrics on the test set.
Table 4. Model performance metrics on the test set.
MetricExtra TreesLightGBMRandom Forest
True Positive224213217
True Negative207213209
False Positive728379
False Negative898387
Accuracy0.7280.7200.720
Precision0.7570.7200.733
Recall0.7160.7200.714
F1 Score0.7360.7200.723
Table 5. Model performance for given thresholds.
Table 5. Model performance for given thresholds.
Metricp = 0.65p = 0.80p = 0.85p = 0.90p = 0.95
Uncertain50148195258344
True Positive204174157138102
True Negative204176162140107
False Positive6544342213
False Negative6950443426
Accuracy0.7530.7880.8040.8320.843
Precision0.7580.7980.8220.8630.887
Recall0.7470.7770.7810.8020.797
F1 Score0.7530.7870.8010.8310.840
Correct or Uncertain77.3%84.1%86.8%90.5%93.4%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Christodoulou, R.; Christofi, G.; Pitsillos, R.; Ibrahim, R.; Papageorgiou, P.; Papageorgiou, S.G.; Vassiliou, E.; Georgiou, M.F. AI-Based Classification of Mild Cognitive Impairment and Cognitively Normal Patients. J. Clin. Med. 2025, 14, 5261. https://doi.org/10.3390/jcm14155261

AMA Style

Christodoulou R, Christofi G, Pitsillos R, Ibrahim R, Papageorgiou P, Papageorgiou SG, Vassiliou E, Georgiou MF. AI-Based Classification of Mild Cognitive Impairment and Cognitively Normal Patients. Journal of Clinical Medicine. 2025; 14(15):5261. https://doi.org/10.3390/jcm14155261

Chicago/Turabian Style

Christodoulou, Rafail, Giorgos Christofi, Rafael Pitsillos, Reina Ibrahim, Platon Papageorgiou, Sokratis G. Papageorgiou, Evros Vassiliou, and Michalis F. Georgiou. 2025. "AI-Based Classification of Mild Cognitive Impairment and Cognitively Normal Patients" Journal of Clinical Medicine 14, no. 15: 5261. https://doi.org/10.3390/jcm14155261

APA Style

Christodoulou, R., Christofi, G., Pitsillos, R., Ibrahim, R., Papageorgiou, P., Papageorgiou, S. G., Vassiliou, E., & Georgiou, M. F. (2025). AI-Based Classification of Mild Cognitive Impairment and Cognitively Normal Patients. Journal of Clinical Medicine, 14(15), 5261. https://doi.org/10.3390/jcm14155261

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop