Next Article in Journal
Graph Sampling-Based Multi-Stream Enhancement Network for Visible-Infrared Person Re-Identification
Next Article in Special Issue
Advancing Breast Cancer Diagnosis through Breast Mass Images, Machine Learning, and Regression Models
Previous Article in Journal
Accurate Visual Simultaneous Localization and Mapping (SLAM) against Around View Monitor (AVM) Distortion Error Using Weighted Generalized Iterative Closest Point (GICP)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Explainable Risk Prediction of Post-Stroke Adverse Mental Outcomes Using Machine Learning Techniques in a Population of 1780 Patients

1
Management Information Department, Office of Clinical Epidemiology, Analytics and kNowledge (OCEAN), Tan Tock Seng Hospital, Singapore 308433, Singapore
2
School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore 639798, Singapore
3
Rehabilitation Research Institute of Singapore, Nanyang Technological University, Singapore 308232, Singapore
4
National Heart Centre Singapore, Singapore 169609, Singapore
5
Duke-NUS Medical School, Singapore 169857, Singapore
6
Department of General Surgery, Vascular Surgery Service, Tan Tock Seng Hospital, Singapore 308433, Singapore
7
Department of Psychiatry, Tan Tock Seng Hospital, Singapore 308433, Singapore
8
Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore 308232, Singapore
9
School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, QLD 4305, Australia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(18), 7946; https://doi.org/10.3390/s23187946
Submission received: 15 August 2023 / Revised: 13 September 2023 / Accepted: 15 September 2023 / Published: 17 September 2023
(This article belongs to the Special Issue AI-Based Automated Recognition and Detection in Healthcare)

Abstract

:
Post-stroke depression and anxiety, collectively known as post-stroke adverse mental outcome (PSAMO) are common sequelae of stroke. About 30% of stroke survivors develop depression and about 20% develop anxiety. Stroke survivors with PSAMO have poorer health outcomes with higher mortality and greater functional disability. In this study, we aimed to develop a machine learning (ML) model to predict the risk of PSAMO. We retrospectively studied 1780 patients with stroke who were divided into PSAMO vs. no PSAMO groups based on results of validated depression and anxiety questionnaires. The features collected included demographic and sociological data, quality of life scores, stroke-related information, medical and medication history, and comorbidities. Recursive feature elimination was used to select features to input in parallel to eight ML algorithms to train and test the model. Bayesian optimization was used for hyperparameter tuning. Shapley additive explanations (SHAP), an explainable AI (XAI) method, was applied to interpret the model. The best performing ML algorithm was gradient-boosted tree, which attained 74.7% binary classification accuracy. Feature importance calculated by SHAP produced a list of ranked important features that contributed to the prediction, which were consistent with findings of prior clinical studies. Some of these factors were modifiable, and potentially amenable to intervention at early stages of stroke to reduce the incidence of PSAMO.

1. Introduction

1.1. Background

Stroke is one of the main contributors to morbidity and mortality in developed countries [1]. In 2019, Singapore’s age-specific crude incidence rate of stroke was 257.6 per 100,000 population; stroke was the fourth leading cause of death and the leading cause of disability [2]. Post-stroke depression (PSD) and anxiety (PSA), collectively known as post-stroke adverse mental outcome (PSAMO),are common sequelae of stroke. About 30% of stroke survivors develop clinical symptoms of depression at some point following stroke [3,4]; and about 20% develop anxiety [4,5]. Stroke survivors with PSAMO have poorer health outcomes, including higher mortality [6,7] and greater functional disability [8,9].
Contributors to PSA include post-stroke fatigue [10,11], sleep disturbance [10], and psychosocial factors like nonmarital status, lack of social support, [11] living alone [12], family history of depression [13], and severity of stroke [13]. There is considerable overlap in the factors that contribute to PSD and PSA, e.g., left hemisphere lesions and cognitive impairment [14]. Indeed, patients may experience symptoms of anxiety after PSD, and vice versa [15,16,17]. Early intervention and treatment can play important roles in managing PSD and PSA. Studies have shown that antidepressants and psychotherapy in the early stages of stroke can be helpful to manage PSAMO [12,18,19].

1.2. Literature Review

Wang et al. also studied the use of ML algorithms to predict PSA, using a sample size of 395 cases and predictors such as demographics and lab results. However, risk factors that attributed to PSA were identified using conventional statistical significance that was calculated from multivariate logistic regression and not explained by the ML model. Ryu et al. and Fast et al. conducted studies in the area of PSD, and utilized a large variety of features for modeling. However, the limitation of both studies was their relatively small sample sizes (65 and 307 respectively), which impaired the generalizability of the models. Among the studies mentioned above, only Fast et al. developed a model that was explainable from which risk factors attributed to PSD could be evaluated. The other studies did not use any explainable artificial intelligence (XAI) methods.
In summary, there are limited studies that have utilized ML algorithms to develop a predictive model. Regarding the studies that have used ML algorithms, PSD and PSA have been studied separately and not as PSAMO as a whole. Most of the studies have used a small sample size and the developed model has not been explainable. Thus, we have not been able to identify the important features that the model used to arrive at such a conclusion. A summary of the studies reviewed can be found below (Table 1).

1.3. Motivation and Proposed Method

We were motivated to develop a predictive model to automatically identify stroke patients at risk of PSAMO for early intervention. Compared to manual screening for PSD and PSA via questionnaire administration [20,21], which is onerous and prone to human biases, such a model could facilitate efficient screening. Machine learning (ML) have been shown to be superior to classical statistics in developing prediction models [22,23]. As such, we tested several ML algorithms, including the following: logistic regression, decision tree, gradient-boosted tree, random forest, XGBoost, CatBoost, AdaBoost, and LightGBM. Compared with recently published ML models that have either studied only PSD or PSA [24,25,26], our model attained higher accuracy for the combined PSAMO diagnosis (Table 1).
Table 1. Summary of studies for automated prediction of post-stroke adverse mental outcome.
Table 1. Summary of studies for automated prediction of post-stroke adverse mental outcome.
AuthorDatasetFeaturesOutcomeTechniquesBest
Performance
Ryu et al. [24], 202231 PSD and 34
non-PSD cases
Medical history, demographics,
neurological, cognitive, and functional test data
PSDSVM, KNN, RFSVM: AUC 0.711; Acc 0.70; Sens 0.742; Spec 0.517
Fast et al. [25], 202349 PSD and 258 non-PSD casesDemographics, clinical, serological, and MRI dataPSD *GBT, SVMGBT: Balanced Acc 0.63; AUC 0.70
Wang et al. [26], 2021395 cases Demographics, lab results, vascular risk factors PSARF, DT, SVM, stochastic gradient descent, multi-layer perceptronRF: 18.625 Euclidean distance between anxiety scores
Current study285 PSAMO and 1495 no PSAMO casesDemographics, stroke-related data, surgical and medical history, etc. PSAMO *Logistic regression, DT, GBT, RF, XGBoost, CatBoost, AdaBoost, LightGBMGBT: AUC 0.620; Acc 0.747; F1-score 0.341
Acc, accuracy; AUC, area under the curve; DT, decision tree; GBT, gradient-boosted tree; KNN, k-nearest neighbor; RF, random forest; Sens, sensitivity; Spec, specificity; SVM, support vector machine. * Developed models were explainable.

1.4. Main Contributions

To the best of our knowledge, our study is the first ML model for predicting the risk of PSAMO, instead of PSA and PSD separately. Our model was trained on a 1780-subject dataset, the largest to date, which represented a broader population and enhanced the generalizability of our results. Like [26], our model incorporated explainable artificial intelligence (XAI) that was able to highlight the most discriminative features used by the model for PSAMO risk prediction, which would be useful for targeted intervention of identified patients with high-risk features.

2. Methods

2.1. Data Collection and Study Design

The study population comprised 1780 patients who had been admitted for ischemic or hemorrhagic stroke to a tertiary care hospital, each of whom had completed an anxiety or depression screening assessment within 7–37 days from the time of stroke. The time window facilitated the inclusion of patients who had been stabilized after acute stroke management, and it was consistent with the 30-day window of prediction in the literature for identifying PSAMO during stroke recovery [27,28,29,30]. Baseline characteristics, including demographics, social information (e.g., occupation, educational level, etc.), quality of life scores, stroke-related information, medical history and comorbidities, medication history, and history of psychiatric conditions and interventions, were collected that constituted potential features to be input into the model (see Table S1). Patients who died before 37 days were excluded. The retrospective analysis of the data had been approved by the hospital ethics review board. A summarized workflow of the proposed model can be found below (Figure 1).

2.2. Identification of PSAMO

We used the Hospital Anxiety and Depression Scale (HADS) [31] and the Patient Health Questionnaire (PHQ) [32] to diagnose PSA, PSD, and PSAMO among our study participants. HADS is a fourteen-item questionnaire; anxiety and depression have seven items each, with each item scored from 0 to 3. PSA and PSD were diagnosed based on scores of 7 or greater on the HADS-anxiety and HADS-depression scales, respectively (82% sensitivity and 78% specificity [33]), and PSAMO was diagnosed based on scores of 10 or greater on the HADS-total scale [34]. The PHQ has two versions. PHQ-9 is a nine-item questionnaire with a score range from 0 to 27; a score of 8 or greater indicated PSD [32] (88% sensitivity and 86% specificity for major depression [35]). PHQ-2 is an abbreviated version containing only the first two items of PHQ-9 [36]; a score of 3 or greater denoted PSD (83% sensitivity and 92% specificity for major depression [36]). In the presence of the abovementioned scenarios, the stroke patient was defined as having PSAMO; and in the absence, no PSAMO.

2.3. Statistical Analysis

Features with continuous values were first tested for normality using the Shapiro–Wilk test. If the distribution was normal, the Student’s t-test was employed to test for significance between the PSAMO and no PSAMO groups. The values were reported as mean ± standard deviation. If the distribution was non-normal, the Mann–Whitney U test was employed to compare the groups and their values were reported as medians (interquartile range). Categorical features were tested for significance using a chi-square test and were reported as counts (n) and percentages (%). The results of the statistical analysis helped to provide a preliminary understanding of the data collected between the 2 cohorts (PSAMO and no PSAMO).

2.4. Data Preprocessing and Engineering

Due to the retrospective nature of our study, features with more than 25% missing data were excluded from modeling. Missing data in the remaining features were assumed to be missing at random and were imputed using multiple imputation-chained equations [37]. For training, the test data split was 70:30. The training set was used to train and validate the model using 10-fold cross-validation; the test set was set aside as unseen data to be used to evaluate model performance after model development. For the training set, continuous features were standardized, and the calculated mean and standard deviation were applied to and transformed on the test set. Categorical features were also one-hot encoded, with similar treatments on both the training and test sets. After preprocessing the dataset, 46 features were obtained that were used for modeling. The list of features is found in Table S1.

2.5. Model Development

Figure 2 is the flow diagram of the steps taken for model development.

2.5.1. Recursive Feature Elimination (RFE)

We used recursive feature elimination (RFE) as a feature selector to choose the most discriminative features from a given dataset for downstream modeling. To find the optimal number of the most discriminative features, RFE iteratively eliminates less important features based on their impact on the model’s performance using the same ML algorithm that is deployed to train the model later. A basic default model of each algorithm was used during RFE to potentiality reduce overfitting of the training data. Taking support vector machine (SVM) as an example, a basic default SVM model can be used to perform RFE for feature selection. Then, the data with the selected features can be further trained using the SVM algorithm with its hyperparameter tuned using Bayesian optimization as explained in Section 2.5.3 and Section 2.5.4 below.

2.5.2. Synthetic Minority Oversampling Technique (SMOTE)

The prevalence of PSAMO in our study population was mildly imbalanced at 17%. Class imbalance can exacerbate the classification bias towards the majority class, especially in ML modeling of high-dimensional problems [38]. To address the class imbalance, we applied the synthetic minority oversampling technique (SMOTE) [39] to the dataset to increase the proportion of the minority class to 50% by synthetically creating samples. The SMOTE algorithm would randomly select a minority class sample, identify its nearest few other samples using the k-nearest neighbor technique, and then synthesize samples by interpolating feature values between the selected sample and its neighbors.

2.5.3. Application of Machine Learning Algorithms

We deployed eight ML algorithms to train the model. The logistic regression ML algorithm with the least absolute shrinkage and selection operator [40] is a regularization technique for mitigating overfitting that estimates the probability of the outcome belonging to either class, akin to conventional statistical logistic regression. The other seven algorithms were tree-based algorithms. The decision tree [41] iteratively splits the data into new trees depending on the values of the features, arriving eventually at the final classifications, the leaves. In a random forest classifier [42], multiple decision tree algorithms are combined to make better aggregate predictions via an ensemble learning method called bagging [43], in which new trees are trained in parallel on subsets of the training data that have been selected randomly with replacement. This reduces overfitting and enhances model generalizability. The remaining five ML algorithms employ another ensemble learning method called boosting [44], in which training is iterated sequentially, with each new tree focusing on and learning from errors made by the preceding tree. Gradient-boosted tree [45] trains on the errors of previous iterations of trees. XGBoost incorporates a regularization element in the trees to prevent overfitting [46]. AdaBoost assigns higher weights to data points that have been wrongly classified, giving them higher importance in the next iteration [47]. CatBoost constrains model complexity by growing only symmetric trees [48]. LightGBM, a more efficient boosting algorithm, uses a gradient-based one-side sampling technique to select the optimal set of data to train the next iteration of trees, reducing computation time while maintaining model performance [49].

2.5.4. Bayesian Optimization and Cross-Validation

During the training of each ML algorithm, the hyperparameters were tuned using Bayesian optimization [50,51]. Bayesian optimization evaluates past iterations and leverages on the results to iteratively explore the best possible parameter space to maximize model performance. The process stops upon reaching the maximum number of iterations or when early stopping criteria have been met. Each set of hyperparameters is evaluated using repeated stratified 10-fold cross-validation with five repeats.

2.6. Model Evaluation

We evaluated the model using standard performance metrics: accuracy and F1-score (the geometric mean of specificity and sensitivity). In addition, we performed a receiver operating characteristic analysis, reporting the area under the curve (AUC) as an index of the degree of discrimination between the two groups. Youden’s J adjustment [52] was performed to calculate the optimal binary classification threshold value, i.e., the Youden index, which is defined as the largest distance from the AUC curve to the line of no discrimination, for each of the ML algorithms:
Y o u d e n   I n d e x = M a x c u t o f f ( S e n s i t i v i t y c u t o f f + S p e c i f i c i t y c u t o f f 1 )

2.7. Model Explanation

We applied Shapley additive explanations (SHAP) [53], a model-agnostic XAI method that uses corporative game theory to quantitate the predictive value of every feature, on the model with the best-performing ML algorithm. SHAP estimates the average marginal contribution of each feature and then ranks the features based on their respective contributions to the prediction, which is a measure of feature importance. SHAP is an XAI method that offers a global explanation, which is useful in our context to understand how different features impact the model’s predictions on average and the overall behavior of a model across the entire dataset. This helps to identify key factors that attribute to PSAMO, which is one of the research objectives of this study, and intervention plans can be developed around such factors.

2.8. Packages Used

The statistical analyses and modeling in this study were implemented in Python 3 programming language [39,46,48,49,50,53,54,55,56,57,58,59]. The packages used and their functions are listed in Appendix A.

3. Results

Descriptive statistics of the study population and comparisons between the PSAMO vs. no PSAMO groups are found in Table S2.
The training set which consists of 69 features was fitted into RFE for feature selection. During RFE, the optimal number of the most discriminative features were selected for each ML algorithm. Figure 3 plots the average accuracy of the individual algorithms against the number of features. The optimal number of discriminative features differed among the algorithms. For instance, the decision tree algorithm attained its peak mean test accuracy using 13 features. As such, the decision tree algorithm would be trained and tuned using only the 13 features selected by RFE. The hyperparameters of the decision tree algorithms were tuned using Bayesian optimization, with each set of hyperparameters evaluated using repeated stratified 10-fold cross-validation with five repeats. The best set of hyperparameters, i.e., the decision tree algorithm with the best performance, was then evaluated on the test set, which was held out and not used during the training.
Table 2 shows the performance of all ML algorithms with the best set of hyperparameters that were used and trained on. The training set is reported with a mean and 95% confidence interval of the results obtained from the repeated stratified 10-fold cross-validation with five repeats. The test set is reported after Youden’s J adjustment.
After training all algorithms, the best-performing model for the prediction of PSAMO was the gradient-boosted tree with accuracy, F1-score, and AUC values of 0.747, 0.341, and 0.620, respectively; the latter after adjustment of the cut-off value using the Youden Index (Table 2).

SHAP Explanation of the Gradient-Boosted Tree Model

SHAP was applied to the gradient-boosted tree model. Figure 4 depicts the top ten most important features calculated using the SHAP algorithm.

4. Discussion

4.1. Model Performance

The developed ML model enabled automated classification of PSAMO vs. no PSAMO in stroke patients with 74.7% accuracy based on baseline features (as referenced against conventional manually administered HADS and PHQ tests), allowing for reproducible and expeditious risk predictions. An early prediction model would enable doctors to systematically triage stroke patients at risk of PSAMO for further confirmatory assessments and to institute early intervention where applicable to prevent or ameliorate adverse mental health outcomes. Of note, our ML model could predict the risk of PSAMO, i.e., PSD and PSA collectively, and was more accurate than similar models in the literature (Table 1). Finally, even though we had labeled our dataset based on the results of manually administered PHQ and HADS, these tests were nonetheless subject to human errors and biases [20,21]. In particular, their reliance on the Likert scale for scoring exposed them to measurement errors and tendency biases that could be sensitive to different modes of administering the tests [60].
The best performing model in our study is the gradient-boosted tree (GBT). The advantage of using GBT is its ability to model nonlinear relationships with complex interactions between features. GBT is also known to be more robust to outliers as compared to some other algorithms like linear regression. GBT has its limitations, for example, the potential of overfitting, especially if the depth of the tree is not well chosen as a hyperparameter. This limitation is mitigated by using Bayesian optimization to identify the optimal hyperparameters. Another limitation that arises from using GBT is the lack of interpretability due to its ensemble nature as compared to linear regression, where the model can be interpreted by its calculated coefficient. This is also mitigated by using SHAP as a model explanation algorithm to identify the important features.
There are limitations to conduct a one-to-one comparison against previous studies as our study is the first to study PSAMO, while previous studies have examined PSD or PSA separately [24,25,26]. Other studies have also utilized different evaluation metrics and ML algorithms that were not used in this study [24,26]. The closest study that could be referenced is Fast et al.’s [25] study. Fast et al.’s best performing model is also GBT. Both studies showed similar AUROC performances (0.7 ± 0.1 vs. 0.620). However, our study has a greater sample size (1780 vs. 307), which is more representative of the population, and therefore it improves the generalizability of our findings. The features that were used were also slightly different, with only lab markers and demographics as the only common features that were used in both studies. There were also similar findings in the feature importance in both studies, which is further elaborated below.

4.2. Explainable Features

The most discriminative features in the dataset as assessed by SHAP (Figure 4) were consistent with the findings of prior clinical studies [61,62,63,64,65,66,67,68,69]. While many factors like recurrent stroke, the most important feature that predicted PSAMO in our study, were non-modifiable, some risk factors might be amenable to interventions. Lower educational level and household income were risk factors for PSAMO in our study, as well as in others [61,62,63,64]. This is also coherent with Fast et al.’s study [25], which also used SHAP to interpret its GBT model and its top important features are years of education and sex. Higher education enables patients to have clearer insights into their disease, participate more actively in their own care, and better manage their emotions, all of which may reduce the risk of PSAMO [63,65]. Lower income has been found to be associated with less patient participation in after-stroke care [66], possibly due to a lack of access to appropriate care. Based on our model findings, doctors can initiate intervention plans that address socioeconomic risk factors through social aid, improved access to care services, and personalized disease education. Being single was also a risk factor for PSAMO (Figure 4), which could be attributed to lower social and emotional support [67,68,69]. Enhancing community emotional support can reduce depressive symptoms [67], and encouraging patients to stay connected with a social network can help promote positive support-seeking behavior.

4.3. Study Advantages and Limitations

Our study has the following advantages and limitations.

4.3.1. Advantages

  • To the best of our knowledge, this is the first ML model that has been designed to predict the risk of PSAMO, a composite of PSD and PSA.
  • The model predicted the risk of PSAMO with good accuracy (i.e., 74.7%).
  • Trained on the largest PSAMO dataset to date, our model results are less susceptible to the influence of outliers, and therefore are representative of the broader stroke population.
  • XAI-enabled model interpretability allows doctors to develop intervention plans for important risk factors for PSAMO.

4.3.2. Limitations

  • This is a cross-sectional study and the observed associations cannot infer causality. Future expansion of this study to longitudinal data may offer stronger insights.
  • Deep learning methods such as neural networks can be investigated in the future, which may produce better results.

5. Conclusions

We have demonstrated the feasibility of using ML algorithms to predict PSAMO. Our ML model predicted PSAMO reproducibly with good accuracy and at a low cost, without the need for onerous manual diagnostic tests. Our study used a large dataset to train the model, which enhanced the generalizability of the results. Moreover, the interpretation of the model was facilitated using SHAP, which identified important risk predictors of PSAMO that were consistent with published clinical studies. This provided indirect validation of our model, which would help boost confidence among potential clinician users. The added insight into the key risk factors of PSAMO offers opportunities for early intervention. In future works, we propose to study more complex algorithms, such as neural networks or other deep learning methods, which may drive improvements in classification performance.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s23187946/s1, Table S1: List of features collected and used for modelling; Table S2: Descriptive Statistics of the study cohort.

Author Contributions

Conceptualization, C.W.O. and E.Y.K.N.; Methodology and formal analysis, C.W.O., L.G.C. and U.R.A.; Data curation, M.H.S.N. and L.G.C.; Writing—original draft preparation, C.W.O.; Writing—review and editing, E.Y.K.N., L.G.C., R.-S.T. and U.R.A.; supervision, E.Y.K.N., L.G.C. and U.R.A.; project administration, C.W.O., E.Y.K.N., M.H.S.N., L.G.C. and Y.M.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study is approved by the Ethics Committee of National Healthcare Group, Domain Specific Review Board (DSRB) (DSRB Ref No. 2022/00289).

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are not available due to privacy restrictions.

Acknowledgments

The authors would like to thank Tan Tock Seng Hospital, Singapore for providing for allowing this study to be done.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. List of Python packages used.
Table A1. List of Python packages used.
Package NameFunctions
PandasData pre-processing and engineering
Numpy
ScipyConducting statistical tests and descriptive statistics
Statsmodels
fancyimputeImplementation of Multiple Imputation-Chained Equations (MICE) imputation
ImblearnImplementation of SMOTE
scikit-learnTraining and cross validation of Machine Learning Models such as “Logistic Regression”, “Support Vector Machine”, “Decision Tree”, “Random Forest” and “AdaBoost”
XgboostTraining of “XGBoost” algorithm
CatboostTraining of “CatBoost” algorithm
LightgbmTraining of “LightGBM” algorithm
bayes_optBayesian Optimization for hyperparameter tuning
SHAPSHAP values
MatplotlibDisplay of graphs and charts

References

  1. The Top 10 Causes of Death. Available online: https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death (accessed on 4 August 2023).
  2. Sun, Y.; Lee, S.H.; Heng, B.H.; Chin, V.S. 5-Year Survival and Rehospitalization Due to Stroke Recurrence among Patients with Hemorrhagic or Ischemic Strokes in Singapore. BMC Neurol. 2013, 13, 133. [Google Scholar] [CrossRef]
  3. Ellis, C.; Zhao, Y.; Egede, L.E. Depression and Increased Risk of Death in Adults with Stroke. J. Psychosom. Res. 2010, 68, 545–551. [Google Scholar] [CrossRef] [PubMed]
  4. Schöttke, H.; Giabbiconi, C.-M. Post-Stroke Depression and Post-Stroke Anxiety: Prevalence and Predictors. Int. Psychogeriatr. 2015, 27, 1805–1812. [Google Scholar] [CrossRef]
  5. Burton, C.A.C.; Murray, J.; Holmes, J.; Astin, F.; Greenwood, D.; Knapp, P. Frequency of Anxiety after Stroke: A Systematic Review and Meta-Analysis of Observational Studies. Int. J. Stroke 2013, 8, 545–559. [Google Scholar] [CrossRef]
  6. De Mello, R.F.; Santos, I.d.S.; Alencar, A.P.; Benseñor, I.M.; Lotufo, P.A.; Goulart, A.C. Major Depression as a Predictor of Poor Long-Term Survival in a Brazilian Stroke Cohort (Study of Stroke Mortality and Morbidity in Adults) EMMA Study. J. Stroke Cerebrovasc. Dis. 2016, 25, 618–625. [Google Scholar] [CrossRef] [PubMed]
  7. Cai, W.; Mueller, C.; Li, Y.-J.; Shen, W.-D.; Stewart, R. Post Stroke Depression and Risk of Stroke Recurrence and Mortality: A Systematic Review and Meta-Analysis. Ageing Res. Rev. 2019, 50, 102–109. [Google Scholar] [CrossRef]
  8. Astuti, P.; Kusnanto, K.; Dwi Novitasari, F. Depression and Functional Disability in Stroke Patients. J. Public Health Res. 2020, 9, 1835. [Google Scholar] [CrossRef] [PubMed]
  9. Lee, E.-H.; Kim, J.-W.; Kang, H.-J.; Kim, S.-W.; Kim, J.-T.; Park, M.-S.; Cho, K.-H.; Kim, J.-M. Association between Anxiety and Functional Outcomes in Patients with Stroke: A 1-Year Longitudinal Study. Psychiatry Investig. 2019, 16, 919–925. [Google Scholar] [CrossRef] [PubMed]
  10. Wright, F.; Wu, S.; Chun, H.-Y.Y.; Mead, G. Factors Associated with Poststroke Anxiety: A Systematic Review and Meta-Analysis. Stroke Res. Treat. 2017, 2017, e2124743. [Google Scholar] [CrossRef]
  11. Sanner Beauchamp, J.E.; Casameni Montiel, T.; Cai, C.; Tallavajhula, S.; Hinojosa, E.; Okpala, M.N.; Vahidy, F.S.; Savitz, S.I.; Sharrief, A.Z. A Retrospective Study to Identify Novel Factors Associated with Post-Stroke Anxiety. J. Stroke Cerebrovasc. Dis. 2020, 29, 104582. [Google Scholar] [CrossRef]
  12. Fang, Y.; Mpofu, E.; Athanasou, J. Reducing Depressive or Anxiety Symptoms in Post-Stroke Patients: Pilot Trial of a Constructive Integrative Psychosocial Intervention. Int. J. Health Sci. 2017, 11, 53–58. [Google Scholar]
  13. Shi, Y.; Yang, D.; Zeng, Y.; Wu, W. Risk Factors for Post-Stroke Depression: A Meta-Analysis. Front. Aging Neurosci. 2017, 9, 218. [Google Scholar] [CrossRef]
  14. Li, X.; Wang, X. Relationships between Stroke, Depression, Generalized Anxiety Disorder and Physical Disability: Some Evidence from the Canadian Community Health Survey-Mental Health. Psychiatry Res. 2020, 290, 113074. [Google Scholar] [CrossRef]
  15. Castillo, C.S.; Schultz, S.K.; Robinson, R.G. Clinical Correlates of Early-Onset and Late-Onset Poststroke Generalized Anxiety. Am. J. Psychiatry 1995, 152, 1174–1179. [Google Scholar] [CrossRef]
  16. Starkstein, S.E.; Cohen, B.S.; Fedoroff, P.; Parikh, R.M.; Price, T.R.; Robinson, R.G. Relationship between Anxiety Disorders and Depressive Disorders in Patients with Cerebrovascular Injury. Arch. Gen. Psychiatry 1990, 47, 246–251. [Google Scholar] [CrossRef] [PubMed]
  17. Chemerinski, E.; Robinson, R.G. The Neuropsychiatry of Stroke. Psychosomatics 2000, 41, 5–14. [Google Scholar] [CrossRef] [PubMed]
  18. Woranush, W.; Moskopp, M.L.; Sedghi, A.; Stuckart, I.; Noll, T.; Barlinn, K.; Siepmann, T. Preventive Approaches for Post-Stroke Depression: Where Do We Stand? A Systematic Review. Neuropsychiatr. Dis. Treat. 2021, 17, 3359–3377. [Google Scholar] [CrossRef]
  19. Mikami, K.; Jorge, R.E.; Moser, D.J.; Arndt, S.; Jang, M.; Solodkin, A.; Small, S.L.; Fonzetti, P.; Hegel, M.T.; Robinson, R.G. Prevention of Post-Stroke Generalized Anxiety Disorder, Using Escitalopram or Problem-Solving Therapy. J. Neuropsychiatry Clin. Neurosci. 2014, 26, 323–328. [Google Scholar] [CrossRef]
  20. Eack, S.M.; Greeno, C.G.; Lee, B.-J. Limitations of the Patient Health Questionnaire in Identifying Anxiety and Depression: Many Cases Are Undetected. Res. Soc. Work Pract. 2006, 16, 625–631. [Google Scholar] [CrossRef]
  21. Maters, G.A.; Sanderman, R.; Kim, A.Y.; Coyne, J.C. Problems in Cross-Cultural Use of the Hospital Anxiety and Depression Scale: “No Butterflies in the Desert”. PLoS ONE 2013, 8, e70975. [Google Scholar] [CrossRef]
  22. Premsagar, P.; Aldous, C.; Esterhuizen, T.M.; Gomes, B.J.; Gaskell, J.W.; Tabb, D.L. Comparing Conventional Statistical Models and Machine Learning in a Small Cohort of South African Cardiac Patients. Inform. Med. Unlocked 2022, 34, 101103. [Google Scholar] [CrossRef]
  23. Desai, R.J.; Wang, S.V.; Vaduganathan, M.; Evers, T.; Schneeweiss, S. Comparison of Machine Learning Methods With Traditional Models for Use of Administrative Claims With Electronic Medical Records to Predict Heart Failure Outcomes. JAMA Netw. Open 2020, 3, e1918962. [Google Scholar] [CrossRef] [PubMed]
  24. Ryu, Y.H.; Kim, S.Y.; Kim, T.U.; Lee, S.J.; Park, S.J.; Jung, H.-Y.; Hyun, J.K. Prediction of Poststroke Depression Based on the Outcomes of Machine Learning Algorithms. J. Clin. Med. 2022, 11, 2264. [Google Scholar] [CrossRef]
  25. Fast, L.; Temuulen, U.; Villringer, K.; Kufner, A.; Ali, H.F.; Siebert, E.; Huo, S.; Piper, S.K.; Sperber, P.S.; Liman, T.; et al. Machine Learning-Based Prediction of Clinical Outcomes after First-Ever Ischemic Stroke. Front. Neurol. 2023, 14, 1114360. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, J.; Zhao, D.; Lin, M.; Huang, X.; Shang, X. Post-Stroke Anxiety Analysis via Machine Learning Methods. Front. Aging Neurosci. 2021, 13, 657937. [Google Scholar] [CrossRef]
  27. Chen, Y.-M.; Chen, P.-C.; Lin, W.-C.; Hung, K.-C.; Chen, Y.-C.B.; Hung, C.-F.; Wang, L.-J.; Wu, C.-N.; Hsu, C.-W.; Kao, H.-Y. Predicting New-Onset Post-Stroke Depression from Real-World Data Using Machine Learning Algorithm. Front. Psychiatry 2023, 14, 1195586. [Google Scholar] [CrossRef] [PubMed]
  28. Zhanina, M.Y.; Druzhkova, T.A.; Yakovlev, A.A.; Vladimirova, E.E.; Freiman, S.V.; Eremina, N.N.; Guekht, A.B.; Gulyaeva, N.V. Development of Post-Stroke Cognitive and Depressive Disturbances: Associations with Neurohumoral Indices. Curr. Issues Mol. Biol. 2022, 44, 6290–6305. [Google Scholar] [CrossRef] [PubMed]
  29. Li, J.; Oakley, L.D.; Brown, R.L.; Li, Y.; Luo, Y. Properties of the Early Symptom Measurement of Post-Stroke Depression: Concurrent Criterion Validity and Cutoff Scores. J. Nurs. Res. 2020, 28, e107. [Google Scholar] [CrossRef]
  30. Khazaal, W.; Taliani, M.; Boutros, C.; Abou-Abbas, L.; Hosseini, H.; Salameh, P.; Sadier, N.S. Psychological Complications at 3 Months Following Stroke: Prevalence and Correlates among Stroke Survivors in Lebanon. Front. Psychol. 2021, 12, 663267. [Google Scholar] [CrossRef]
  31. Stern, A.F. The Hospital Anxiety and Depression Scale. Occup. Med. 2014, 64, 393–394. [Google Scholar] [CrossRef]
  32. Kroenke, K.; Spitzer, R.L.; Williams, J.B.W. The PHQ-9. J. Gen. Intern. Med. 2001, 16, 606–613. [Google Scholar] [CrossRef] [PubMed]
  33. Wu, Y.; Levis, B.; Sun, Y.; He, C.; Krishnan, A.; Neupane, D.; Bhandari, P.M.; Negeri, Z.; Benedetti, A.; Thombs, B.D. Accuracy of the Hospital Anxiety and Depression Scale Depression Subscale (HADS-D) to Screen for Major Depression: Systematic Review and Individual Participant Data Meta-Analysis. BMJ 2021, 373, n972. [Google Scholar] [CrossRef]
  34. Burton, L.-J.; Tyson, S. Screening for Mood Disorders after Stroke: A Systematic Review of Psychometric Properties and Clinical Utility. Psychol. Med. 2015, 45, 29–49. [Google Scholar] [CrossRef]
  35. Urtasun, M.; Daray, F.M.; Teti, G.L.; Coppolillo, F.; Herlax, G.; Saba, G.; Rubinstein, A.; Araya, R.; Irazola, V. Validation and Calibration of the Patient Health Questionnaire (PHQ-9) in Argentina. BMC Psychiatry 2019, 19, 291. [Google Scholar] [CrossRef]
  36. Kroenke, K.; Spitzer, R.L.; Williams, J.B.W. The Patient Health Questionnaire-2: Validity of a Two-Item Depression Screener. Med. Care 2003, 41, 1284–1292. [Google Scholar] [CrossRef] [PubMed]
  37. Jakobsen, J.C.; Gluud, C.; Wetterslev, J.; Winkel, P. When and How Should Multiple Imputation Be Used for Handling Missing Data in Randomised Clinical Trials—A Practical Guide with Flowcharts. BMC Med. Res. Methodol. 2017, 17, 162. [Google Scholar] [CrossRef]
  38. Blagus, R.; Lusa, L. Class Prediction for High-Dimensional Class-Imbalanced Data. BMC Bioinform. 2010, 11, 523. [Google Scholar] [CrossRef]
  39. Blagus, R.; Lusa, L. SMOTE for High-Dimensional Class-Imbalanced Data. BMC Bioinform. 2013, 14, 106. [Google Scholar] [CrossRef] [PubMed]
  40. Tibshirani, R. Regression Shrinkage and Selection via the Lasso. J. R. Stat. Soc. Ser. B Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  41. Fürnkranz, J. Decision Tree. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2010; pp. 263–267. [Google Scholar] [CrossRef]
  42. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  43. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  44. Bühlmann, P.; Hothorn, T. Boosting Algorithms: Regularization, Prediction and Model Fitting. Stat. Sci. 2007, 22, 477–505. [Google Scholar] [CrossRef]
  45. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  46. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  47. Freund, Y.; Schapire, R.E. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  48. Dorogush, A.V.; Ershov, V.; Gulin, A. CatBoost: Gradient Boosting with Categorical Features Support. arXiv 2018, arXiv:1810.11363. [Google Scholar] [CrossRef]
  49. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
  50. Martinez-Cantin, R. BayesOpt: A Bayesian Optimization Library for Nonlinear Optimization, Experimental Design and Bandits. arXiv 2014, arXiv:1405.7430. [Google Scholar] [CrossRef]
  51. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian Optimization of Machine Learning Algorithms. arXiv 2012, arXiv:1206.2944. [Google Scholar] [CrossRef]
  52. Youden, W.J. Index for Rating Diagnostic Tests. Cancer 1950, 3, 32–35. [Google Scholar] [CrossRef]
  53. Lundberg, S.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar] [CrossRef]
  54. McKinney, W. Pandas: A Foundational Python Library for Data Analysis and Statistics; Academic Publishers: Singapore, 2011. [Google Scholar]
  55. Array Programming with NumPy|Nature. Available online: https://www.nature.com/articles/s41586-020-2649-2 (accessed on 20 June 2023).
  56. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python|Nature Methods. Available online: https://www.nature.com/articles/s41592-019-0686-2 (accessed on 20 June 2023).
  57. Seabold, S.; Perktold, J. Statsmodels: Econometric and Statistical Modeling with Python; SCIPY: Austin, TX, USA, 2010; pp. 92–96. [Google Scholar] [CrossRef]
  58. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  59. Droettboom, M.; Hunter, J.; Firing, E.; Caswell, T.A.; Dale, D.; Lee, J.-J.; Elson, P.; McDougall, D.; Straw, A.; Root, B.; et al. Matplotlib, version 1.4.0; CERN Data Center: Meyrin, Switzerland, 2014. [Google Scholar] [CrossRef]
  60. Pinto-Meza, A.; Serrano-Blanco, A.; Peñarrubia, M.T.; Blanco, E.; Haro, J.M. Assessing Depression in Primary Care with the PHQ-9: Can It Be Carried Out over the Telephone? J. Gen. Intern. Med. 2005, 20, 738–742. [Google Scholar] [CrossRef] [PubMed]
  61. Lyu, Y.; Li, W.; Tang, T. Prevalence Trends and Influencing Factors of Post-Stroke Depression: A Study Based on the National Health and Nutrition Examination Survey. Med. Sci. Monit. Int. Med. J. Exp. Clin. Res. 2022, 28, e933367-e1–e933367-e8. [Google Scholar] [CrossRef]
  62. Wang, Z.; Zhu, M.; Su, Z.; Guan, B.; Wang, A.; Wang, Y.; Zhang, N.; Wang, C. Post-Stroke Depression: Different Characteristics Based on Follow-up Stage and Gender–A Cohort Perspective Study from Mainland China. Neurol. Res. 2017, 39, 996–1005. [Google Scholar] [CrossRef] [PubMed]
  63. Zhang, H.; Ma, J.; Sun, Y.; Xiao, L.D.; Yan, F.; Tang, S. Anxiety Subtypes in Rural Ischaemic Stroke Survivors: A Latent Profile Analysis. Nurs. Open 2023, 10, 4083–4092. [Google Scholar] [CrossRef] [PubMed]
  64. Park, E.-Y.; Kim, J.-H. An Analysis of Depressive Symptoms in Stroke Survivors: Verification of a Moderating Effect of Demographic Characteristics. BMC Psychiatry 2017, 17, 132. [Google Scholar] [CrossRef] [PubMed]
  65. Lin, F.-H.; Yih, D.N.; Shih, F.-M.; Chu, C.-M. Effect of Social Support and Health Education on Depression Scale Scores of Chronic Stroke Patients. Medicine 2019, 98, e17667. [Google Scholar] [CrossRef]
  66. Egan, M.; Kubina, L.-A.; Dubouloz, C.-J.; Kessler, D.; Kristjansson, E.; Sawada, M. Very Low Neighbourhood Income Limits Participation Post Stroke: Preliminary Evidence from a Cohort Study. BMC Public Health 2015, 15, 528. [Google Scholar] [CrossRef]
  67. Bi, H.; Wang, M. Role of Social Support in Poststroke Depression: A Meta-Analysis. Front. Psychiatry 2022, 13, 924277. [Google Scholar] [CrossRef]
  68. Kruithof, W.J.; van Mierlo, M.L.; Visser-Meily, J.M.A.; van Heugten, C.M.; Post, M.W.M. Associations between Social Support and Stroke Survivors’ Health-Related Quality of Life—A Systematic Review. Patient Educ. Couns. 2013, 93, 169–176. [Google Scholar] [CrossRef]
  69. Knapp, P.; Hewison, J. The Protective Effects of Social Support against Mood Disorder after Stroke. Psychol. Health Med. 1998, 3, 275–283. [Google Scholar] [CrossRef]
Figure 1. Flow diagram of the proposed model.
Figure 1. Flow diagram of the proposed model.
Sensors 23 07946 g001
Figure 2. Flow diagram of model development.
Figure 2. Flow diagram of model development.
Sensors 23 07946 g002
Figure 3. Performance of RFE using different ML algorithms.
Figure 3. Performance of RFE using different ML algorithms.
Sensors 23 07946 g003
Figure 4. Feature importance (using SHAP values) of the developed PSAMO model. A higher SHAP value indicates that the feature has a larger contribution to the model prediction. The figure only shows the top 10 important features.
Figure 4. Feature importance (using SHAP values) of the developed PSAMO model. A higher SHAP value indicates that the feature has a larger contribution to the model prediction. The figure only shows the top 10 important features.
Sensors 23 07946 g004
Table 2. Prediction performance obtained using our proposed models.
Table 2. Prediction performance obtained using our proposed models.
Train SetTest Set (after Youden’s J Adjustment)
AccuracyAUROCF1-ScoreAccuracyAUROCF1-Score
Logistic
Regression
0.696
(0.647–0.738)
0.870
(0.864–0.877)
0.658
(0.646–0.673)
0.6400.5730.304
Gradient-boosted tree0.973
(0.958–0.982)
0.946
(0.932–0.957)
0.950
(0.924–0.964)
0.7470.6200.341
Light GBM0.905
(0.878–0.922)
0.942
(0.924–0.964)
0.882
(0.872–0.894)
0.6590.6460.355
Decision tree0.729
(0.660–0.760)
0.831
(0.810–0.844)
0.705
(0.656–0.729)
0.7130.6220.338
Random forest0.800
(0.779–0.820)
0.911
(0.906–0.915)
0.779
(0.770–0.787)
0.5060.6710.359
XGBoost0.899
(0.874–0.920)
0.956
(0.934–0.968)
0.877
(0.868–0.889)
0.5390.6410.346
AdaBoost0.843
(0.825–0.860)
0.946
(0.939–0.948)
0.828
(0.811–0.841)
0.7270.6340.354
CatBoost0.966
(0.950–0.980)
0.966
(0.952–0.970)
0.957
(0.945–0.971)
0.631
0.5920.318
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Oei, C.W.; Ng, E.Y.K.; Ng, M.H.S.; Tan, R.-S.; Chan, Y.M.; Chan, L.G.; Acharya, U.R. Explainable Risk Prediction of Post-Stroke Adverse Mental Outcomes Using Machine Learning Techniques in a Population of 1780 Patients. Sensors 2023, 23, 7946. https://doi.org/10.3390/s23187946

AMA Style

Oei CW, Ng EYK, Ng MHS, Tan R-S, Chan YM, Chan LG, Acharya UR. Explainable Risk Prediction of Post-Stroke Adverse Mental Outcomes Using Machine Learning Techniques in a Population of 1780 Patients. Sensors. 2023; 23(18):7946. https://doi.org/10.3390/s23187946

Chicago/Turabian Style

Oei, Chien Wei, Eddie Yin Kwee Ng, Matthew Hok Shan Ng, Ru-San Tan, Yam Meng Chan, Lai Gwen Chan, and Udyavara Rajendra Acharya. 2023. "Explainable Risk Prediction of Post-Stroke Adverse Mental Outcomes Using Machine Learning Techniques in a Population of 1780 Patients" Sensors 23, no. 18: 7946. https://doi.org/10.3390/s23187946

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop