Next Article in Journal
Effects of Thymoquinone on Cell Proliferation, Oxidative Damage, and Toll-like Signaling Pathway Genes in H1650 Lung Adenocarcinoma Cell Line
Previous Article in Journal
Effect of Continuous Positive Airway Pressure Treatment on Hearing and Inner Ear Function in Patients with Obstructive Sleep Apnoea—Original Research
Previous Article in Special Issue
Prolotherapy as a Regenerative Treatment in the Management of Chronic Low Back Pain: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Classification of Cognitive Status in Community-Dwelling Sarcopenic Women: A SHAP-Based Analysis of Physical Activity and Anthropometric Factors

1
Department of Management Information Systems, Faculty of Economics and Administrative Sciences, Sivas Cumhuriyet University, Sivas 58140, Türkiye
2
Department of Biostatistics, Faculty of Medicine, Malatya Turgut Ozal University, Malatya 44210, Türkiye
3
Department of Computer Science, Lakehead University, Thunder Bay, ON P7B 5E1, Canada
4
Department of Sport Management, Faculty of Sport Sciences, Inonu University, Malatya 44280, Türkiye
5
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
6
Department of Ocean Operations and Civil Engineering, Norwegian University of Science and Technology (NTNU), 6025 Alesund, Norway
*
Authors to whom correspondence should be addressed.
Medicina 2025, 61(10), 1834; https://doi.org/10.3390/medicina61101834
Submission received: 8 September 2025 / Revised: 5 October 2025 / Accepted: 10 October 2025 / Published: 14 October 2025

Abstract

Background and Objectives: Sarcopenia, characterized by progressive loss of skeletal muscle mass and function, has increasingly been recognized not only as a physical health concern but also as a potential risk factor for cognitive decline. This study investigates the application of machine learning algorithms to classify cognitive status based on Mini-Mental State Examination (MMSE) scores in community-dwelling sarcopenic women. Materials and Methods: A dataset of 67 participants was analyzed, with MMSE scores categorized into severe (≤17) and mild (>17) cognitive impairment. Eight classification models—MLP, CatBoost, LightGBM, XGBoost, Random Forest (RF), Gradient Boosting (GB), Logistic Regression (LR), and AdaBoost—were evaluated using a repeated holdout strategy over 100 iterations. Hyperparameter optimization was performed via Bayesian optimization, and model performance was assessed using metrics including weighted F1-score (w_f1), accuracy, precision, recall, PR-AUC, and ROC-AUC. Results: Among the models, CatBoost achieved the highest w_f1 (87.05 ± 2.85%) and ROC-AUC (90 ± 5.65%), while AdaBoost and GB showed superior PR-AUC scores (92.49% and 91.88%, respectively), indicating strong performance in handling class imbalance and threshold sensitivity. SHAP (SHapley Additive exPlanations) analysis revealed that moderate physical activity (moderatePA minutes), walking days, and sitting time were among the most influential features, with higher physical activity associated with reduced risk of cognitive impairment. Anthropometric factors such as age, BMI, and weight also contributed significantly. Conclusions: The results highlight the effectiveness of boosting-based models in capturing complex patterns in clinical data and provide interpretable evidence supporting the role of modifiable lifestyle factors in cognitive health. These findings suggest that machine learning, combined with explainable AI, can enhance risk assessment and inform targeted interventions for cognitive decline in older women.

1. Introduction

Cognitive impairment represents a critical public health challenge as the global population ages, with profound implications for healthcare systems and individual quality of life. Age-related cognitive decline, affecting millions of older adults worldwide, is driven by biological processes including mitochondrial dysfunction, cellular senescence, and chronic inflammation, as well as neuropathological lesions characteristic of dementia [1,2]. The Mini-Mental State Examination (MMSE) remains the most widely used screening tool for cognitive assessment, providing standardized evaluation across orientation, memory, attention, language, and visuospatial domains, particularly valuable for community-dwelling older populations [1,2].
Emerging evidence reveals complex interrelationships between physical health, body composition, and cognitive function. Sarcopenia—progressive loss of skeletal muscle mass and function—has been increasingly recognized as a potential risk factor for cognitive decline, with cognitive impairment prevalence 30–50% higher among sarcopenic versus non-sarcopenic older adults [3,4,5]. The ‘sarcopenia-cognitive impairment’ hypothesis proposes shared biological mechanisms including chronic inflammation, oxidative stress, mitochondrial dysfunction, and dysregulation of neurotrophic factors. Physical activity has emerged as a promising modifiable factor, with moderate-intensity exercise demonstrating protective effects against both conditions through enhanced cerebral blood flow, neurogenesis, reduced inflammation, and improved insulin sensitivity [6,7].
Despite growing recognition of these relationships, critical knowledge gaps remain. First, the specific patterns and intensities of physical activity that most strongly influence cognitive status in older women—a population particularly vulnerable to both sarcopenia and cognitive decline—have not been adequately characterized. Second, the complex, potentially nonlinear interactions between physical activity patterns, sedentary behavior, and anthropometric variables require analytical approaches beyond traditional statistical methods. Third, while machine learning techniques offer powerful tools for identifying patterns in multidimensional datasets, their “black box” nature has limited clinical interpretability and implementation [8,9].
Recent advances in explainable artificial intelligence (XAI), particularly SHapley Additive exPlanations (SHAP), provide model-agnostic methods for interpreting machine learning predictions by quantifying each feature’s contribution to model output [10]. This approach enables transparent identification of modifiable factors most strongly associated with cognitive outcomes, addressing the interpretability challenge while maintaining analytical sophistication.
The present study addresses these gaps by applying multiple machine learning algorithms with SHAP-based interpretation to classify cognitive status in community-dwelling older women. Specifically, we examine: (1) which aspects of physical activity and sedentary behavior most strongly predict cognitive function; (2) how these factors interact with anthropometric variables; and (3) whether the protective association between moderate physical activity and sarcopenia extends to cognitive health outcomes. Our approach advances beyond previous research by capturing the complex interplay between physical activity, body composition, and cognitive function through a comprehensive, interpretable machine learning framework. We employ Bayesian hyperparameter optimization to enhance model performance and utilize SHAP values for clinically actionable risk assessment. These findings have important implications for developing targeted, evidence-based interventions to preserve cognitive health in older women, supporting physical activity assessment and promotion as integral components of cognitive health strategies for aging populations.

2. Materials and Methods

2.1. Population, Dataset, and Ethical Procedures

The current study, using publicly available data from a cross-sectional study using a quantitative descriptive approach, included data from community-dwelling older women aged ≥60 years with sarcopenia [11]. The Inonu University Health Sciences Non-Interventional Clinical Research Ethics Committee approved this study (approval number: 2025/7305, 11 March 2025). Inclusion criteria were female, the ability to walk independently, the absence of unstable cardiovascular conditions, acute infections, back pain, or tumors, and no unintentional weight loss of more than 3 kg in the last three months. Participants completed a cognitive assessment questionnaire on the first day, and PA and sarcopenia results were measured the following day. The MMSE test was administered to identify cognitive impairment in older women. Data from women with a score of 12 or less were excluded from the analysis. The short version of the International Physical Activity Questionnaire (IPAQ) was used to assess participants’ physical activity levels [11].
This study analyzed physical activity and anthropometric factors to classify cognitive status in community-dwelling older women based on MMSE scores. Below is a concise summary of the key risk factors and types of cognitive impairment considered:
  • Physical Activity Factors
    Moderate physical activity duration (moderatePA minutes): Weekly duration of moderate-intensity physical activity (in minutes).
    Walking days (walk days): Number of days per week spent walking.
    Sitting time (sitting time minutes, 7 days): Total weekly sitting time (in minutes), as an indicator of sedentary behavior.
  • Anthropometric Factors
    Age: Age of participants (in years).
    Body Mass Index (BMI): Calculated as weight divided by the square of height (kg/m2).
    Weight: Body weight of participants (in kg).
    Height: Height of participants (in cm).
  • Types of Cognitive Impairment
    Severe cognitive impairment: Individuals with MMSE scores ≤ 17.
    Mild cognitive impairment: Individuals with MMSE scores > 17.

2.2. Classification Models

Classification is a supervised machine learning task that aims to assign observations to one of several predefined categorical classes. In this study, various algorithms were employed to solve the classification problem. The multilayer perceptron (MLP) neural network stands out for its ability to learn complex nonlinear relationships [11]. Gradient boosting-based algorithms such as CatBoost [12], LightGBM [13], and XGBoost [8] provide high performance, especially when working with categorical variables and large datasets. Random Forest (RF) [14], and Gradient Boosting (GB) [15] algorithms rely on ensembles of decision trees to balance bias and variance. Logistic regression (LR) [16] is a widely used baseline method for modeling binary outcomes. AdaBoost [17] aims to improve prediction accuracy by sequentially boosting weak classifiers. In this study, these models were comparatively evaluated, and classification performance was assessed using various metrics. All computational analyses in this research were conducted using Python 3.10 within the Spyder 6.0.7 platform. This integrated development environment provides an interactive framework for data analysis and model construction.

2.3. Performance Metrics

Performance metrics are quantitative measures used to evaluate how well a machine learning model performs on a given task. They are crucial for assessing not only the overall predictive capability of the model but also for understanding its strengths and limitations across different aspects of classification performance. In this study, we utilized weighted F1-score (w_f1), accuracy (acc), precision (pre), recall (rec), precision-recall curve score (pr_auc), and roc_auc score to comprehensively assess the classification models, capturing both threshold-dependent and threshold-independent performance characteristics. Detailed calculations and explanations of the metrics used can be found in the relevant references [18,19,20,21].

2.4. Bayesian Optimization

Bayesian optimization is a sequential model-based optimization technique that efficiently finds the global optimum of expensive, black-box functions by building a probabilistic surrogate (typically a Gaussian process) to model the objective and selecting new points via an acquisition function that balances exploration and exploitation. Unlike grid or random search, it strategically uses prior evaluations to predict promising regions, often requiring far fewer evaluations to locate optima. This makes it particularly effective for hyperparameter tuning in machine learning. For a detailed explanation of the method, please refer to the cited article [22].

2.5. SHapley Additive exPlanations

SHAP is a unified framework that explains machine learning model predictions by assigning each feature a Shapley value, derived from cooperative game theory. These values quantify how much each feature contributes to pushing a prediction away from the dataset’s average output. SHAP ensures properties like local accuracy (the contributions sum to the model output), consistency, and missingness. It offers both model-agnostic (e.g., Kernel SHAP) and model-specific (e.g., TreeSHAP) algorithms for efficient computation. For a detailed explanation of the method, please refer to the cited article [23,24].

3. Results

As illustrated in Figure 1, the machine learning experimental workflow of this study comprises several sequential steps. First, the class label in the dataset (MMSE) was categorized, followed by data normalization. The dataset was then split into three subsets: training, testing, and validation. Hyperparameter optimization was performed using the training and validation sets through a Bayesian optimization approach, and the resulting optimal hyperparameters were employed to train the model on the training set. Subsequently, performance metric scores and SHAP values were calculated on the test set using the trained model. Since a repeated holdout strategy was adopted, this entire process was executed 100 times, after which the average metric scores and model explanations were reported.

3.1. Data Preparation

In the initial phase of dataset preparation, the MMSE score, originally provided as a continuous variable, was categorized. In this process, guided by findings from previous studies in the literature, samples with an MMSE score of 17 or below were labeled as severe cognitive impairment, while those with a score above 17 were labeled as mild cognitive impairment. The MMSE cut-off of 17 was selected based on Whelan et al. (2009), who identified this threshold as optimal for distinguishing severe cognitive impairment [25]. Following the categorization, data normalization was performed to ensure that all features contributed equally to the model, thereby promoting fairer and more accurate training. For this purpose, the MinMaxScaler function from the sklearn library in Python version 3.10 was employed. At the end of this stage, the resulting dataset comprised 67 samples, with 28 (42.8%) belonging to severe cognitive impairment and 39 (57.2%) to mild cognitive impairment. Although the class sizes were not perfectly balanced, the relatively small difference led to the expectation that this level of imbalance would not pose significant issues during model training. Indeed, prior research suggests that class distributions within the 60–40% range are generally considered to represent mild imbalance, which in most cases does not substantially affect model performance [26].

3.2. Hyperparameter Optimization

After preparing the dataset for training, the hyperparameter optimization phase was initiated. In this step, 10% of the samples from the entire dataset were randomly selected to form the validation set, while 20% were randomly selected to constitute the test set. The remaining samples were allocated to the training set. To ensure that the class distribution was preserved across the training, testing, and validation sets, the train_test_split function from the sklearn library was utilized with the stratify parameter. As a result, the training set contained 46 samples, the test set 14 samples, and the validation set 7 samples. During the model optimization phase, several specific library versions were employed to ensure reproducibility and stability of the results. Bayesian optimization was performed using the scikit-optimize (skopt) library version 0.9.0. For developing the XGBoost model, xgboost version 2.0.3 was used, while the LightGBM and CatBoost models were implemented with lightgbm version 4.6.0 and catboost version 1.2.5, respectively. Additionally, other classification models were developed using scikit-learn (sklearn) version 1.5.0. The optimized hyperparameter values for each model are presented in Table 1.
As presented in Table 1, different hyperparameters of the proposed models were optimized. Since the hyperparameter optimization process was repeated in each holdout iteration, each model has 100 distinct sets of optimal hyperparameters. Therefore, these optimal hyperparameters are not included in Table 1. For each model, the remaining hyperparameters not shown in Table 1 were kept at their default values as specified in the respective libraries used during model development. During the hyperparameter optimization phase, the w_f1 computed on the validation dataset was used as the performance metric. In the skopt library, the acquisition function parameter was set to Expected Improvement, and the Bayesian optimization process was executed for 25 calls.

3.3. Model Training and Evaluation

Upon completion of hyperparameter optimization, each model was trained with the identified optimal hyperparameters during every repeated holdout iteration. Subsequently, model performances were evaluated on the testing dataset employing a range of metrics. Given that performance metrics were computed separately for each holdout, this process yielded 100 distinct scores per metric for each model. Table 2 summarizes these results by presenting the mean and the corresponding standard deviation (std) for each metric.
The table presents the comparative performance of eight different machine learning models (MLP, CatBoost, LightGBM, XGBoost, RF, GB, LR, and AdaBoost) in classifying MMSE status, evaluated over 100 repeated holdout experiments. Each cell reports the mean ± std for various metrics: w_f1, acc, pre, rec, pr_auc, and auc_roc, all expressed as percentages. According to the results in this table, CatBoost, GB, and AdaBoost consistently outperformed other models across most metrics. CatBoost achieved the highest mean w_f1 (87.05 ± 2.85%), closely followed by AdaBoost (86.43 ± 2.15%) and GB (86.35 ± 2.14%). Similarly, these three models also reported the highest accuracies. All models maintained a reasonable balance between precision and recall. Notably, GB and AdaBoost achieved high precision (88.58% and 88.32%, respectively) while maintaining high recall (both 86.42%), indicating stable sensitivity and specificity. AdaBoost and GB reported the highest pr_auc (92.49 ± 5.83% and 91.88 ± 5.18%, respectively), suggesting superior ability to differentiate between classes under varying decision thresholds. CatBoost followed with 89.6 ± 8.48%. Regarding roc_auc, CatBoost led with 90 ± 5.65%, slightly ahead of AdaBoost and GB. The std indicate the models’ robustness under repeated sampling. CatBoost, GB, and AdaBoost exhibited low variability across most metrics (std < 3% for w_f1 and acc), emphasizing their stable generalization. In contrast, RF showed relatively higher variability (e.g., w_f1 std of 6.68%), implying greater sensitivity to training data splits. The MLP, LightGBM, XGBoost, and LR demonstrated lower overall performance, with w_f1 ranging from ~77% to ~80%, though still maintaining acceptable classification levels. Interestingly, XGBoost’s pr_auc (85.93 ± 10.79%) was somewhat higher than its roc_auc (81.88 ± 9.4%), suggesting potential calibration issues or threshold sensitivity. In summary, ensemble-based tree models, particularly CatBoost, GB, and AdaBoost, provided superior and more consistent performance for MMSE status classification in this study. These findings highlight the effectiveness of boosting approaches in handling structured clinical data with possible nonlinear patterns.

3.4. Model Explanations

The ability to interpret the predictions of machine learning models, particularly those of complex algorithms often referred to as “black-box” models, is becoming increasingly important in both academic research and practical applications. Understanding why a model makes a particular prediction is critical for enhancing the model’s trustworthiness, identifying and mitigating potential biases, diagnosing the root causes of incorrect predictions to improve performance, and deriving business or scientific insights from the findings. Therefore, in this study, the algorithms proposed for MMSE prediction were explained using the SHAP method after the training phase. To facilitate model interpretation, version 0.46.0 of the official SHAP library in Python was utilized. For each method, SHAP values were computed by applying the trained model from each holdout iteration to the corresponding test dataset. These SHAP values from all holdouts were then concatenated to obtain the aggregated SHAP values across all iterations. Finally, violin plots illustrating these aggregated SHAP values are presented in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9.
The figures display SHAP summary violin plots for each of the eight machine learning models employed to classify MMSE status, providing insight into the contribution and impact of individual features on model predictions. The SHAP values were computed on testing sets from 100 repeated holdout experiments, ensuring robust estimation of feature importance and the consistency of their effects. According to this figure, several features consistently emerged as influential across models, including moderatePA minutes, walk days, Sitting time minutes (7 days), age, height, BMI, and weight. This aligns with known clinical and epidemiological associations between physical activity, anthropometrics, and cognitive function. Notably, moderatePA minutes and walk days were recurrently among the top-ranked features, especially prominent in AdaBoost, LightGBM, and XGBoost, underscoring the relevance of moderate physical activity frequency and walking behavior in differentiating MMSE categories.
Models such as CatBoost, GB, and AdaBoost, which demonstrated superior predictive metrics (e.g., CatBoost with a w_f1 of 87.05% and AdaBoost with the highest pr_auc of 92.49%), also exhibited clear, well-separated SHAP distributions for key features. The CatBoost plot, sitting time, walk days, and moderatePA minutes show distinct SHAP value spreads, reflecting stable, meaningful contributions to model output. Similarly, GB and AdaBoost plots reveal broader SHAP value distributions for top features, indicating that variations in these variables have substantial and consistent impacts on the predicted probability of MMSE impairment.
The color gradients in the violin plots (from blue indicating low feature values to pink indicating high feature values) illustrate how increases in features such as moderatePA minutes and walk days generally lead to positive SHAP values in several models (e.g., XGBoost, AdaBoost), suggesting a protective role (lower risk of cognitive impairment). Conversely, longer sitting time minutes tends to shift SHAP values negatively in models like GB and CatBoost, highlighting its potential adverse influence on MMSE classification.
The LR plot shows more symmetric and uniform SHAP distributions across features, consistent with its simpler linear nature and comparatively lower predictive performance (w_f1 = 77.45%). In contrast, tree-based ensemble models captured more complex, heterogeneous relationships, evident from the wider spread of SHAP values. The RF model exhibited less pronounced SHAP value separation, aligning with its higher standard deviation in performance metrics, suggesting greater sensitivity to data variability. The SHAP analysis validates the interpretability and reliability of the ensemble boosting models (CatBoost, GB, AdaBoost) not only through quantitative performance metrics but also by demonstrating clear, biologically plausible patterns of feature importance. These plots collectively reinforce the central role of moderate-intensity physical activity, walking frequency, and sedentary behavior in influencing cognitive outcomes, providing valuable interpretive support for the clinical implications of the machine learning findings.

4. Discussion

4.1. Interpretation of Machine Learning Performance in MMSE Classification

The present study investigated the application of various machine learning algorithms for classifying cognitive status based on MMSE scores in community-dwelling older women. Our findings demonstrate that ensemble-based tree models, particularly CatBoost, GB, and AdaBoost, consistently outperformed other algorithms across multiple performance metrics. CatBoost achieved the highest weighted F1-score (87.05 ± 2.85%), while AdaBoost and GB reported superior precision-recall area under the curve (pr_auc) values of 92.49 ± 5.83% and 91.88 ± 5.18%, respectively. These results align with previous research indicating that gradient boosting algorithms often excel in handling structured clinical data with complex, nonlinear relationships [8,13].
The superior performance of CatBoost is particularly noteworthy given its robust handling of categorical variables without requiring extensive preprocessing, which is advantageous in clinical datasets where categorical features are common. The relatively low standard deviations observed for CatBoost, GB, and AdaBoost across repeated holdout experiments (w_f1 std < 3%) indicate their stability and consistent generalization capabilities, which are crucial for clinical applications where model reliability is paramount. In contrast, the Random Forest model exhibited higher variability (w_f1 std of 6.68%), suggesting greater sensitivity to specific data partitions—a finding consistent with previous comparative studies of tree-based ensemble methods [27,28].
Our results, showing the effectiveness of boosting algorithms in cognitive status classification, support the growing body of literature demonstrating the utility of these methods in healthcare applications. Recent studies have similarly found gradient boosting approaches to outperform traditional statistical methods in predicting various health outcomes, including cognitive decline. The consistent superiority of CatBoost in our study may be partially attributed to its innovative ordered boosting technique and effective handling of categorical features, which appears particularly well-suited for the mixed data types present in geriatric clinical assessments [9,27].

4.2. Clinical Significance of Physical Activity and Sedentary Behavior

The SHAP analysis revealed that physical activity metrics, particularly moderate physical activity minutes and walking days, emerged as consistently influential features across multiple high-performing models. This provides machine learning-based evidence supporting the well-established association between physical activity and cognitive health in older adults. Notably, higher values of moderate PA minutes and walking days generally corresponded to positive SHAP values in several models, suggesting an association with lower likelihood of severe cognitive impairment.
These results are consistent with previous findings [7], reporting that moderate physical activity is linked to reduced odds of sarcopenia in community-dwelling older women. The present findings extend this relationship to cognitive function, suggesting a potential interconnected pathway between sarcopenia, physical activity, and cognitive health in elderly women. This triad of relationships aligns with the “sarcopenia–cognitive impairment” hypothesis, which proposes shared biological mechanisms such as inflammation, oxidative stress, and neurotrophic factor dysregulation [3,4].
Conversely, longer sitting time minutes consistently demonstrated negative SHAP values in models like GB and CatBoost, indicating a potential adverse association with cognitive status. This is particularly significant, as it provides quantitative evidence for the detrimental impact of sedentary behavior on cognition, independent of physical activity levels. These results support epidemiological studies showing that prolonged sedentary time is associated with greater risk of cognitive decline, even among individuals who meet physical activity guidelines [29,30].
The observed negative association between prolonged sitting time and cognitive outcomes may not only reflect physical inactivity but also underlying motivational or affective dimensions. Apathy, defined as a multidimensional construct encompassing diminished goal-directed behavior, cognition, and emotional engagement, has been increasingly recognized as a key neuropsychiatric correlate of neurodegenerative conditions [31,32]. Reduced spontaneous movement or extended sedentary behavior may therefore represent behavioral manifestations of motivational decline rather than purely lifestyle choices. This interpretation suggests that interventions targeting cognitive health in older adults should consider integrating psychologically oriented strategies—such as behavioral activation or motivational coaching—alongside physical activity promotion to optimize outcomes.
The consistency of these findings across multiple high-performing machine learning models strengthens the evidence for physical activity and sedentary behavior as modifiable correlates of cognitive impairment in older women. This has important clinical implications, suggesting that interventions designed to promote moderate physical activity and reduce sedentary time may be associated with more favorable cognitive outcomes [33].

4.3. Anthropometric Factors and Cognitive Health

In addition to physical activity metrics, anthropometric variables including age, height, BMI, and weight consistently emerged as influential features in the SHAP analysis. The prominence of age as a predictor aligns with the well-established relationship between advancing age and increased risk of cognitive impairment. However, the significant contribution of BMI and weight to model predictions suggests a more complex relationship between body composition and cognitive health than previously recognized [6].
These findings resonate with the sarcopenia literature, as sarcopenia is often characterized by both loss of muscle mass and increased fat infiltration, which can affect BMI measurements in nonlinear ways. The fact that these anthropometric variables maintained consistent importance across models suggests potential underlying physiological mechanisms linking body composition to cognitive function. This supports the growing recognition of the “obesity-cognition paradox” in older adults, where both underweight and obesity have been associated with increased risk of cognitive decline [5,34].
The integration of physical activity metrics with anthropometric variables in predictive models represents a significant advancement over previous studies that typically examined these factors in isolation. Our machine learning approach has effectively captured the complex interplay between these factors, providing a more holistic understanding of their combined influence on cognitive status. This multidimensional perspective is crucial for developing comprehensive prevention and intervention strategies for cognitive impairment in older women [35].
While BMI emerged as one of the influential predictors in the SHAP analyses, its relationship with cognitive health is unlikely to be linear or uniform across individuals, particularly in older adults. BMI is a crude anthropometric metric that does not differentiate between adipose tissue and lean mass, and thus may obscure clinically relevant phenomena such as sarcopenic obesity—a condition characterized by excess body fat alongside reduced muscle strength and mass. Prior research suggests that higher BMI may exhibit both protective and detrimental effects depending on body composition, inflammatory status, and cardiometabolic profile, which may explain the modest yet complex contribution of BMI to cognitive performance observed in our models. Rather than functioning as an isolated risk factor, BMI likely interacts with muscle mass, physical function, and metabolic resilience to influence neurocognitive outcomes. Future studies employing dual-energy X-ray absorptiometry (DEXA) or bioimpedance-based body composition assessments would provide more precise insights into the mechanistic role of fat-to-lean mass distribution in cognitive aging.
Beyond the conventional interpretation of adiposity as a negative health determinant, recent systematic reviews have highlighted a potential “obesity paradox” in aging populations, whereby higher body mass may confer protective effects under certain physiological conditions. For instance, individuals with greater metabolic reserves or preserved muscle mass may be more resilient to neurodegenerative processes despite elevated BMI. Conversely, excess adiposity coupled with high inflammatory burden or reduced physical function may exacerbate cognitive deterioration. These divergent trajectories underscore the need to move beyond BMI as a unidimensional proxy and instead adopt body composition-based phenotyping approaches. Integrating measures such as waist-to-hip ratio, appendicular lean mass, or inflammatory biomarkers could refine risk stratification models and provide a more coherent understanding of adiposity-related cognitive resilience versus vulnerability.

4.4. Methodological Considerations and Model Interpretability

Although the methodological framework of this study incorporated widely used machine learning algorithms and SHAP-based interpretability techniques, these tools were not applied in a generic or algorithm-agnostic manner. Instead, several design decisions were deliberately tailored to the constraints of the dataset, including the choice of lightweight models to accommodate the limited sample size, the use of stratified repeated holdout validation to preserve class balance, and the prioritization of interpretable features to ensure clinical relevance. Rather than aiming for maximal predictive accuracy through complex architectures, the analytical strategy intentionally emphasized transparency, generalizability, and clinical plausibility—critical factors when modeling cognitive outcomes in small-scale geriatric populations.
The application of SHAP values for model interpretation represents a significant strength of this study, addressing the critical need for transparency in clinical machine learning applications. Unlike traditional feature importance measures that only indicate the relative contribution of features without directionality, SHAP values provide both the magnitude and direction of each feature’s influence on individual predictions. This enhanced interpretability is particularly valuable in clinical contexts where understanding the specific impact of risk factors on outcomes is essential for developing targeted interventions [10,36].
Our findings demonstrate how SHAP analysis can bridge the gap between high-performing ‘black-box’ models and clinically actionable insights. For instance, the color gradients in our SHAP violin plots clearly illustrate how increases in moderate physical activity generally lead to protective effects against severe cognitive impairment, while prolonged sitting time exerts detrimental effects. This level of interpretability transforms complex ensemble models from opaque predictors into tools that can inform clinical decision-making and patient counseling [37].
The consistent patterns observed across multiple high-performing models (CatBoost, GB, and AdaBoost) further strengthen the reliability of our feature importance findings. When multiple algorithms with different underlying mechanisms identify the same features as influential, it provides robust evidence for the genuine relevance of those features to the outcome of interest. This convergence of evidence from different modeling approaches represents a sophisticated validation method that goes beyond traditional statistical significance testing [38].

4.5. Limitations and Future Research Directions

Despite the promising findings, several limitations should be acknowledged. The relatively small sample size (67 participants) constrains the generalizability of our results, though the repeated holdout approach helped maximize data utility. The cross-sectional design prevents causal inference; therefore, all findings should be interpreted as associations rather than causal effects. The categorization of MMSE scores into two groups (≤17 vs. >17), while widely used in the literature to distinguish severe from mild cognitive impairment, reduces the granularity of cognitive function assessment. Additionally, our models did not incorporate potential mediating or moderating variables that could explain underlying mechanisms. Finally, focusing exclusively on community-dwelling older women limits generalizability to men or institutionalized populations, particularly given known sex differences in sarcopenia prevalence and cognitive aging trajectories.
Based on these limitations, we propose three prioritized research directions. First and most urgently, prospective longitudinal studies with larger sample sizes (300–500 participants) are needed to establish temporal and potentially causal relationships between physical activity patterns, body composition, and cognitive trajectories. Such studies should include repeated measures over 3–5 years follow-up, recruit diverse populations across both sexes and ethnic groups, and examine dose–response relationships between physical activity parameters and cognitive outcomes. Second, future research should incorporate biological markers and advanced imaging to elucidate underlying mechanisms, including inflammatory markers (IL-6, TNF-α, CRP), neurotrophic factors (BDNF, IGF-1), neuroimaging (structural MRI, functional MRI, PET scans) to assess brain structure and cerebral blood flow, and genetic data (e.g., APOE genotype) to examine gene–environment interactions. Third, to capture the full spectrum of cognitive function, future machine learning studies should treat MMSE as a continuous outcome or implement multi-class classification approaches, incorporate domain-specific cognitive assessments beyond MMSE (memory, executive function, processing speed), and validate models across multiple independent cohorts. These three priorities address the fundamental limitations of cross-sectional design, lack of mechanistic understanding, and binary classification, respectively, while providing a clear roadmap for advancing this field toward evidence-based interventions for cognitive health in aging populations.
While the present study offers valuable insights into cognitive health determinants among older women, the exclusive focus on a single sex limits the generalizability of the findings to the broader aging population. Evidence suggests that cognitive aging trajectories may differ between men and women, particularly in domains related to executive functioning, emotion recognition, and neuropsychiatric symptoms [39,40]. Biological factors such as hormone regulation, inflammatory profiles, and body composition, as well as sociocultural differences in health behaviors, may contribute to sex-specific mechanisms of cognitive decline. Therefore, future research should aim to include both sexes in order to elucidate potential interaction effects between sex, lifestyle factors, and cognitive performance, thereby enabling the development of more tailored intervention strategies.
Although the machine learning models yielded satisfactory performance in distinguishing cognitive impairment levels based on MMSE classification, the relatively small sample size (n = 67) warrants careful interpretation of the findings. Working with limited data inherently increases the risk of model overfitting, as algorithms may inadvertently learn patterns specific to the current sample rather than generalizable relationships applicable to broader populations. Repeated holdout validation was employed to partially mitigate this risk; however, this approach cannot fully substitute for external validation across independent cohorts. Moreover, more robust resampling techniques—such as bootstrapped confidence intervals or permutation-based stability assessments—were not feasible given the dataset constraints. Future studies should therefore aim to recruit larger and more heterogeneous populations, ideally including longitudinal follow-up to assess temporal consistency in prediction accuracy. Incorporating external replication datasets or bootstrapped resampling pipelines would further strengthen the reliability and clinical applicability of machine learning-based risk stratification in cognitive health research.

4.6. Clinical and Public Health Implications

The findings of this study have significant implications for clinical practice and public health interventions targeting cognitive health in older women. The consistent identification of moderate physical activity and walking frequency as protective factors suggests that relatively simple, accessible interventions could have meaningful impacts on cognitive outcomes. Clinicians should consider incorporating assessments of physical activity patterns and sedentary behavior into routine cognitive evaluations of older patients, and provide tailored recommendations for increasing moderate activity and reducing sitting time.
From a public health perspective, these findings support the development of community-based programs that promote physical activity specifically designed for older women. Given the emphasis on moderate-intensity activity in our results, interventions need not focus exclusively on high-intensity exercise, which may be inaccessible or unsafe for many older adults. Instead, programs emphasizing regular walking and other moderate activities could provide cognitive benefits while being more widely adoptable.
The integration of machine learning approaches into geriatric assessment represents a promising direction for personalized medicine. By identifying individual risk profiles based on physical activity patterns, anthropometric measures, and other factors, clinicians could develop targeted prevention strategies for cognitive decline. Our study demonstrates that ensemble boosting methods, particularly when combined with SHAP-based interpretation, can provide both accurate predictions and clinically interpretable insights—addressing a critical challenge in the adoption of artificial intelligence in healthcare settings.

5. Conclusions

This study demonstrates the superior performance of ensemble boosting algorithms, particularly CatBoost, in classifying cognitive status based on MMSE scores in older women. More importantly, the SHAP-based interpretation of these models revealed consistent patterns linking moderate physical activity, walking frequency, and reduced sedentary time with better cognitive outcomes, while highlighting the importance of anthropometric factors. These findings provide machine learning-validated evidence supporting the integration of physical activity assessment and promotion into cognitive health strategies for older women.
The convergence of evidence from multiple high-performing machine learning models strengthens the credibility of our findings and suggests robust relationships between physical activity patterns, body composition, and cognitive status. While acknowledging the limitations of our cross-sectional design and modest sample size, these results contribute to the growing body of evidence supporting physical activity as a modifiable protective factor against cognitive decline in older adults.
Future research should build upon these findings through longitudinal studies with larger, more diverse samples, incorporation of biological markers to elucidate underlying mechanisms, and development of targeted interventions based on machine learning-derived risk profiles. As the global population ages and the burden of cognitive impairment grows, the integration of advanced analytics with clinical gerontology represents a promising frontier for preserving cognitive health in later life.

Author Contributions

Conceptualization, Y.G. and F.H.Y.; data curation, Y.G., F.H.Y. and Y.A.; formal analysis, Y.G. and F.H.Y.; investigation, Y.A., S.A.A., A.A.A. and M.A.; methodology, Y.G. and F.H.Y.; resources, Y.G., F.H.Y., Y.A., S.A.A., A.A.A. and M.A.; software, F.H.Y.; validation, Y.G., F.H.Y., Y.A., S.A.A., A.A.A. and M.A.; writing—original draft, Y.G., F.H.Y., Y.A., S.A.A., A.A.A. and M.A.; writing—review and editing, Y.G., F.H.Y., Y.A., S.A.A., A.A.A. and M.A. All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R716), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

The studies involving humans were approved by the İnonu University Health Sciences Non-Interventional Clinical Research Ethics Committee (approval number: 2025/7305, date: 11 March 2025). The studies were conducted in accordance with the local legislation and institutional requirements.

Informed Consent Statement

Since this study using open access is a retrospective study with minimal risk and potential risks for research subjects are not greater than for individuals who do not participate in the study, exemption from the requirement to obtain prior consent does not affect the rights and interests of research subjects.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

This study was supported by the Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2025R716), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gonzales, M.; Garbarino, V.; Pollet, E.; Palavicini, J.; Kellogg, D.; Kraig, E.; Orr, M. Biological aging processes underlying cognitive decline and neurodegenerative disease. J. Clin. Investig. 2022, 132, e158453. [Google Scholar] [CrossRef]
  2. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. “Mini-mental state”: A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef]
  3. Zhu, H.; Li, H.D.; Feng, B.L.; Zhang, L.; Zheng, Z.X.; Zhang, Y.; Wang, D.J.; Xiong, Z.; Kang, J.F.; Jin, J.C.; et al. Association between sarcopenia and cognitive impairment in community-dwelling population. Chin. Med. J. 2020, 134, 725–727. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, X.; Zhang, W.; Wang, C.; Tao, W.; Dou, Q.; Yang, Y. Sarcopenia as a predictor of hospitalization among older people: A systematic review and meta-analysis. BMC Geriatr. 2018, 18, 188. [Google Scholar] [CrossRef]
  5. Cruz-Jentoft, A.J.; Bahat, G.; Bauer, J.; Boirie, Y.; Bruyère, O.; Cederholm, T.; Cooper, C.; Landi, F.; Rolland, Y.; Sayer, A.A.; et al. Sarcopenia: Revised European consensus on definition and diagnosis. Age Ageing 2019, 48, 16–31. [Google Scholar] [CrossRef] [PubMed]
  6. Livingston, G.; Huntley, J.; Sommerlad, A.; Ames, D.; Ballard, C.; Banerjee, S.; Brayne, C.; Burns, A.; Cohen-Mansfield, J.; Cooper, C.; et al. Dementia prevention, intervention, and care: 2020 report of the Lancet Commission. Lancet 2020, 396, 413–446. [Google Scholar] [CrossRef] [PubMed]
  7. Marini, J.A.G.; Abdalla, P.P.; Bohn, L.; Mota, J.; Duncan, M.; Dos Santos, A.P.; Machado, D.R.L. Moderate Physical Activity Reduces the Odds of Sarcopenia in Community-dwelling Older Women: A Cross-sectional Study. Curr. Aging Sci. 2023, 16, 219–226. [Google Scholar] [CrossRef]
  8. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  9. Nallapu, B.T.; Petersen, K.K.; Qian, T.; Demirsoy, I.; Ghanbarian, E.; Davatzikos, C.; Lipton, R.B.; Ezzati, A. A Machine Learning Approach to Predict Cognitive Decline in Alzheimer Disease Clinical Trials. Neurology 2025, 104, e213490. [Google Scholar] [CrossRef]
  10. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4768–4777. [Google Scholar]
  11. Gardner, M.W.; Dorling, S.R. Artificial neural networks (the multilayer perceptron)—A review of applications in the atmospheric sciences. Atmos. Environ. 1998, 32, 2627–2636. [Google Scholar] [CrossRef]
  12. Hancock, J.T.; Khoshgoftaar, T.M. CatBoost for big data: An interdisciplinary review. J. Big Data 2020, 7, 94. [Google Scholar] [CrossRef] [PubMed]
  13. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A highly efficient gradient boosting decision tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  14. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  15. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  16. Hosmer Jr, D.W.; Lemeshow, S.; Sturdivant, R.X. Applied Logistic Regression; John Wiley & Sons: New York, NY, USA, 2013. [Google Scholar]
  17. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  18. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar] [CrossRef]
  19. Saito, T.; Rehmsmeier, M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef]
  20. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  21. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  22. Yagin, F.H.; Alkhateeb, A.; Raza, A.; Samee, N.A.; Mahmoud, N.F.; Colak, C.; Yagin, B. An explainable artificial intelligence model proposed for the prediction of myalgic encephalomyelitis/chronic fatigue syndrome and the identification of distinctive metabolites. Diagnostics 2023, 13, 3495. [Google Scholar] [CrossRef] [PubMed]
  23. Arslan, A.K.; Yagin, F.H.; Algarni, A.; Karaaslan, E.; Al-Hashem, F.; Ardigò, L.P. Enhancing type 2 diabetes mellitus prediction by integrating metabolomics and tree-based boosting approaches. Front. Endocrinol. 2024, 15, 1444282. [Google Scholar] [CrossRef] [PubMed]
  24. Choi, S.H.; Lee, J.M. Explainable fault diagnosis model using stacked autoencoder and kernel SHAP. In Proceedings of the 2022 IEEE International Symposium on Advanced Control of Industrial Processes (AdCONIP), Vancouver, BC, Canada, 7–9 August 2022; pp. 182–187. [Google Scholar]
  25. Whelan, P.J.; Oleszek, J.; Macdonald, A.; Gaughran, F. The utility of the Mini-Mental State Examination in guiding assessment of capacity to consent to research. Int. Psychogeriatr. 2009, 21, 338–344. [Google Scholar] [CrossRef] [PubMed]
  26. Buda, M.; Maki, A.; Mazurowski, M.A. A systematic study of the class imbalance problem in convolutional neural networks. Neural Netw. 2018, 106, 249–259. [Google Scholar] [CrossRef]
  27. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 2–8 December 2018; pp. 6639–6649. [Google Scholar]
  28. Caruana, R.; Niculescu-Mizil, A.; Crew, G.; Ksikes, A. Ensemble selection from libraries of models. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004; p. 18. [Google Scholar]
  29. Falck, R.; Davis, J.; Liu-Ambrose, T. What is the association between sedentary behaviour and cognitive function? A systematic review. Br. J. Sports Med. 2016, 51, 800–811. [Google Scholar] [CrossRef]
  30. Dillon, K.; Morava, A.; Prapavessis, H.; Grigsby-Duffy, L.; Novic, A.; Gardiner, P. Total Sedentary Time and Cognitive Function in Middle-Aged and Older Adults: A Systematic Review and Meta-analysis. Sports Med.-Open 2022, 8, 127. [Google Scholar] [CrossRef]
  31. Yi, H.-J.; Tan, C.-H.; Hong, W.-P.; Yu, R.-L. Development and validation of the geriatric apathy scale: Examining multi-dimensional apathy profiles in a neurodegenerative population with cultural considerations. Asian J. Psychiatry 2024, 93, 103924. [Google Scholar] [CrossRef] [PubMed]
  32. Montoya--Murillo, G.; Ibarretxe--Bilbao, N.; Peña, J.; Ojeda, N. The impact of apathy on cognitive performance in the elderly. Int. J. Geriatr. Psychiatry 2019, 34, 657–665. [Google Scholar] [CrossRef] [PubMed]
  33. Marino, F.; Deal, J.; Gross, A.; An, Y.; Tian, Q.; Simonsick, E.; Ferrucci, L.; Resnick, S.; Schrack, J.; Wanigatunga, A. Directionality between cognitive function and daily physical activity patterns. Alzheimer’s Dement. Transl. Res. Clin. Interv. 2025, 11, e70068. [Google Scholar] [CrossRef] [PubMed]
  34. Zhai, W.; Zhang, G.; Wei, C.; Zhao, M.; Sun, L. The obesity paradox in cognitive decline: Impact of BMI dynamics and APOE genotypes across various cognitive status. Diabetes Obes. Metab. 2025, 27, 3967–3983. [Google Scholar] [CrossRef]
  35. Brodaty, H.; Chau, T.; Heffernan, M.; Ginige, J.; Andrews, G.; Millard, M.; Sachdev, P.; Anstey, K.; Lautenschlager, N.; McNeil, J.; et al. An online multidomain lifestyle intervention to prevent cognitive decline in at-risk older adults: A randomized controlled trial. Nat. Med. 2025, 31, 565–573. [Google Scholar] [CrossRef]
  36. Hu, L.; Chen, S.; Fu, Y.; Gao, Z.; Long, H.; Ren, H.-W.; Zuo, Y.; Wang, J.; Li, H.; Xu, Q.; et al. Risk Factors Associated with Clinical Outcomes in 323 COVID-19 Hospitalized Patients in Wuhan, China. Clin. Infect. Dis. 2020, 71, 2089–2098. [Google Scholar] [CrossRef]
  37. Zhao, X.; Sun, B.; Chu, X.; Wu, D.; Jiang, G.; Zhou, H.; Cai, J. A decision-making approach under uncertainty based on ensemble learning model with multimodal data and its application in medical diagnosis. Expert Syst. Appl. 2025, 265, 125983. [Google Scholar] [CrossRef]
  38. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
  39. Avila, J.F.; Vonk, J.M.; Verney, S.P.; Witkiewitz, K.; Rentería, M.A.; Schupf, N.; Mayeux, R.; Manly, J.J. Sex/gender differences in cognitive trajectories vary as a function of race/ethnicity. Alzheimer’s Dement. 2019, 15, 1516–1523. [Google Scholar] [CrossRef] [PubMed]
  40. Pua, S.-Y.; Yu, R.-L. Effects of executive function on age-related emotion recognition decline varied by sex. Soc. Sci. Med. 2024, 361, 117392. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Step-by-Step Process of the Machine Learning Experiments Conducted for Data Analysis and Model Evaluation.
Figure 1. Step-by-Step Process of the Machine Learning Experiments Conducted for Data Analysis and Model Evaluation.
Medicina 61 01834 g001
Figure 2. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for AdaBoost Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 2. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for AdaBoost Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g002
Figure 3. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for CatBoost Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 3. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for CatBoost Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g003
Figure 4. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for Gradient Boosting Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 4. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for Gradient Boosting Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g004
Figure 5. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for LightGBM Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 5. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for LightGBM Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g005
Figure 6. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for Logistic Regression Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 6. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for Logistic Regression Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g006
Figure 7. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for MLP Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 7. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for MLP Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g007
Figure 8. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for Random Forest Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 8. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for Random Forest Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g008
Figure 9. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for XGBoost Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Figure 9. SHAP summary violin plots indicating feature importance across different machine learning models for MMSE classification for XGBoost Classifier. Plots are generated using SHAP values on test datasets from 100 repeated holdout experiments.
Medicina 61 01834 g009
Table 1. Overview of Optimized Hyperparameters for Each Model.
Table 1. Overview of Optimized Hyperparameters for Each Model.
Model NameHyperparameter NameHyperparameter TypeHyperparameter Space
MLPhidden_layer_sizesCategorical50 to 2000 (step = 50)
AlphaRealHigh = 10−1, Low = 10−4
Learning_rateRealHigh = 10−1, Low = 10−4
CatBoostLearning_rateRealHigh = 0.3, Low = 10−3
IterationsIntegerHigh = 1500, Low = 100
DepthIntegerHigh = 10, Low = 4
l2_leaf_regRealHigh = 10, Low = 10−3
loss_functionCategorical‘Logloss’, ‘CrossEntropy’
LightGBMMax_depthIntegerHigh = 10, Low = 3
Learning_rateRealHigh = 0.3, Low = 10−2
n_estimatorsIntegerHigh = 500, Low = 50
SubsampleRealHigh = 1.0, Low = 0.5
colsample_bytreeRealHigh = 1.0, Low = 0.5
min_child_samplesIntegerHigh = 100, Low = 20
reg_alphaRealHigh = 10, Low = 10−3
reg_lambdaRealHigh = 10, Low = 10−3
XGBoostMax_depthIntegerHigh = 10, Low = 3
learning_rateRealHigh = 0.3, Low = 0.01
n_estimatorsIntegerHigh = 1500, Low = 50
SubsampleRealHigh = 1, Low = 0.5
colsample_bytreeRealHigh = 1, Low = 0.5
GammaRealHigh = 10, Low = 0
min_child_weightIntegerHigh = 10, Low = 1
RFn_estimatorsIntegerHigh = 1500, Low = 10
max_depthIntegerHigh = 20, Low = 2
min_samples_splitRealHigh = 0.99, Low = 0.01
max_featuresRealHigh = 0.5, Low = 0.001
GBn_estimatorsIntegerHigh = 1500, Low = 50
learning_rateRealHigh = 0.3, Low = 0.001
max_depthIntegerHigh = 30, Low = 1
min_samples_splitIntegerHigh = 20, Low = 2
min_samples_leafIntegerHigh = 20, Low = 1
LRCRealHigh = 100, Low = 10−6
SolverCategorical‘lbfgs’, ‘liblinear’
AdaBoostn_estimatorsIntegerHigh = 10, Low = 200
learning_rateRealHigh = 0.01, Low = 2
Table 2. Mean performance scores and standard deviations obtained over 100 repeated holdout experiments for machine learning models used to classify MMSE status.
Table 2. Mean performance scores and standard deviations obtained over 100 repeated holdout experiments for machine learning models used to classify MMSE status.
Model Namew_f1 (%)std (%)acc (%)std (%)pre (%)std (%)rec (%)std (%)pr_auc (%)std (%)auc_roc (%)std (%)
MLP80.043.4980.713.2783.233.9180.713.2776.229.8976.466.52
CatBoost87.052.8587.142.8688.33.0587.142.8689.68.48905.65
LightGBM79.742.68802.8680.963.83802.8681.379.7781.047.52
XGBoost79.692.7802.8680.873.86802.8685.9310.7981.889.4
RF81.846.6882.146.5984.996.4482.146.5988.349.0785.947.94
GB86.352.1486.422.1488.582.386.422.1491.885.1888.966.18
LR77.453.6977.863.8578.914.8677.863.8580.218.5178.965.39
AdaBoost86.432.1586.422.1488.322.5186.422.1492.495.8389.586.86
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gormez, Y.; Yagin, F.H.; Aygun, Y.; Alzakari, S.A.; Alhussan, A.A.; Aghaei, M. Machine Learning Classification of Cognitive Status in Community-Dwelling Sarcopenic Women: A SHAP-Based Analysis of Physical Activity and Anthropometric Factors. Medicina 2025, 61, 1834. https://doi.org/10.3390/medicina61101834

AMA Style

Gormez Y, Yagin FH, Aygun Y, Alzakari SA, Alhussan AA, Aghaei M. Machine Learning Classification of Cognitive Status in Community-Dwelling Sarcopenic Women: A SHAP-Based Analysis of Physical Activity and Anthropometric Factors. Medicina. 2025; 61(10):1834. https://doi.org/10.3390/medicina61101834

Chicago/Turabian Style

Gormez, Yasin, Fatma Hilal Yagin, Yalin Aygun, Sarah A. Alzakari, Amel Ali Alhussan, and Mohammadreza Aghaei. 2025. "Machine Learning Classification of Cognitive Status in Community-Dwelling Sarcopenic Women: A SHAP-Based Analysis of Physical Activity and Anthropometric Factors" Medicina 61, no. 10: 1834. https://doi.org/10.3390/medicina61101834

APA Style

Gormez, Y., Yagin, F. H., Aygun, Y., Alzakari, S. A., Alhussan, A. A., & Aghaei, M. (2025). Machine Learning Classification of Cognitive Status in Community-Dwelling Sarcopenic Women: A SHAP-Based Analysis of Physical Activity and Anthropometric Factors. Medicina, 61(10), 1834. https://doi.org/10.3390/medicina61101834

Article Metrics

Back to TopTop