Next Article in Journal
Comparative Analysis of Small Nerve Fiber Density in Fibromyalgia Syndrome and Small Fiber Neuropathy
Previous Article in Journal
Distinct Plasma LPC Signatures Differentiate COVID-19 Sepsis from Other Sepsis Aetiologies
Previous Article in Special Issue
Histological Evaluation for Collagen Expression Prior to LVAD Implantation Is Useful to Estimate Weaning Success
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Predictive Performance of Machine Learning Models for Heart Failure Readmission: A Systematic Review

1
Medical Surgical Department, College of Nursing, University of Hail, Ha’il 81451, Saudi Arabia
2
Medical Surgical Department, College of Nursing, Taif University, Taif 21944, Saudi Arabia
3
Mental Health Nursing Department, College of Nursing, University of Hail, Ha’il 81451, Saudi Arabia
4
Medical Surgical Department, College of Nursing, King Khalid University, Abha 61421, Saudi Arabia
5
College of Nursing, University of Florida, Gainesville, FL 32610, USA
6
College of Nursing-Jeddah, King Saud bin Abdulaziz University for Health Sciences, Jeddah 21423, Saudi Arabia
7
Nursing Administration and Education Department, College of Nursing, University of Tabuk, Tabuk 47311, Saudi Arabia
8
Medical Surgical Nursing Department, Prince Sultan Military College of Health Sciences, Dhahran 34313, Saudi Arabia
*
Author to whom correspondence should be addressed.
Biomedicines 2025, 13(9), 2111; https://doi.org/10.3390/biomedicines13092111
Submission received: 25 June 2025 / Revised: 20 July 2025 / Accepted: 4 August 2025 / Published: 29 August 2025

Abstract

Background: Patients with heart failure (HF) are at high risk of readmission, contributing to substantial healthcare costs. This study investigated machine learning (ML) approaches to predict HF readmissions. Methods: A systematic review was conducted using several medical databases, adhering to the PRISMA guidelines, to identify studies employing ML to predict HF readmissions. Three reviewers independently screened the articles and extracted data. Results: Twenty-two studies from six countries were included in this study. Some studies examined 30-day readmissions, whereas others assessed 90-day, 180-day, or 1- to 3-year readmissions. Fourteen studies used supervised learning algorithms, with area under the curve (AUC) values ranging from 0.70 to 0.99, and unsupervised algorithms had AUCs of 0.69 to 0.72. The average age of the patients was 73 years, with approximately equal numbers of males and females. Conclusions: ML can predict HF-related hospitalization across various time frames. Supervised ML approaches and the incorporation of clinical knowledge may enhance model performance. Collaboration between providers and data scientists is needed to improve patient outcomes and reduce costs by using more accurate predictive models.

1. Introduction

Heart failure (HF) is a significant global health problem that affects millions of people and is associated with high readmission rates, particularly within the first month of discharge [1,2]. In the United States, nearly one in five patients with HF are readmitted within a month, contributing approximately USD 30.7 billion in annual costs [3]. Both clinical and non-clinical factors influence post-discharge outcomes. Studies have shown that in addition to clinical predictors, socioeconomic status, frailty, and behavioral factors also affect readmission risk. Incorporating patient-reported psychosocial and socioeconomic factors improves predictive modeling for 30-day readmission [4], and frailty is especially relevant among older adults [5]. Symptom trajectory patterns within the first month after discharge have also been linked to a higher risk of unplanned readmissions [6].
Despite these advances, most existing studies on machine learning (ML) models for HF readmission prediction have several limitations. First, most models have been developed and validated using US- or European-centered cohorts, which raises concerns about their generalizability to diverse populations and healthcare settings. Second, external or multicenter validation, a key step for assessing model robustness and real-world applicability, is infrequently undertaken. Third, many models depend on structured data, such as billing codes, with less attention paid to unstructured clinical notes, social determinants, or psychosocial factors, which may limit accuracy and equity in predictions. Finally, studies vary widely in algorithm selection, predictor sets, and outcome definitions, impeding direct comparisons and the synthesis of findings across the literature.
ML has emerged as a promising approach for predicting HF readmissions, with a higher predictive accuracy (AUC 0.70–0.99) than older statistical models [7,8]. ML can analyze complex nonlinear relationships and enhance risk stratification beyond conventional methods [9]. However, most ML models lack external validation and are primarily developed in the U.S. and European populations, limiting generalizability [10,11,12,13].
Healthcare systems and patient demographics vary substantially by region, influencing model transportability and the risk of perpetuating inequities in precision health [14,15,16,17]. Accurate HF readmission risk forecasting enables targeted interventions and improved outcomes; however, addressing gaps in external validation, population diversity, and the integration of broader social and clinical features is critical for enhancing model reliability and equity [18,19,20].
This systematic review addresses these limitations by systematically evaluating studies across multiple regions, incorporating both clinical and non-clinical predictors, critically synthesizing ML methodological quality, and appraising external validation strategies. By focusing on population diversity, healthcare system differences, and ethical and practical barriers to implementation, this review distinctly contributes to the ongoing effort to develop robust, reliable, and equitable ML models for HF readmission risk prediction.

2. Materials and Methods

2.1. Protocol and Registration

This systematic review was conducted and reported in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) [21] 2020 statement, as can be seen in Supplementary File S1. The review protocol was prospectively registered in the International Prospective Register of Systematic Reviews (PROSPERO) registration number CRD42021247198.

2.2. Eligibility Criteria

Studies published in English from 2015 to 2024 were included if they used machine learning (ML) methods to predict heart failure (HF) readmission risk in acute care settings. Eligible studies utilized applied ML algorithms, including supervised learning models (logistic regression, decision trees, random forest, gradient boosting, support vector machines, and neural networks), unsupervised learning methods (clustering, principal component analysis, and autoencoders), ensemble approaches (bagging, boosting, and stacking), and deep learning architectures (convolutional or recurrent neural networks and attention-based models). Studies were excluded if they used only traditional statistical techniques without ML enhancements, relied solely on manual rule-based or expert systems, or did not clearly describe the ML-based approach. Studies combining ML with standard statistical models were only included if the ML component was central to prediction. Studies have reported HF readmission as a primary outcome, used any clinically relevant prediction window (e.g., 30, 90, 180 days, one year or longer), and provided at least one ML performance metric (AUC, accuracy, precision, recall, or F1-score). Studies comparing ML with non-ML approaches for HF readmission prediction were also eligible.

2.3. Information Sources

A systematic literature search was conducted across CINAHL, EMBASE, MEDLINE/PubMed, Cochrane Library, Web of Science, CNKI, SciELO, clinical trial registries, preprint repositories, conference proceedings, and other gray literature sources. Initial searches were performed between 7 January and 28 February 2024, with a final update on 10 March 2024, to capture the most recent publications before screening.

2.4. Search Strategy

The search strategy combined MeSH terms (where applicable) and free-text keywords related to heart failure, readmission, and machine learning. An example search string for PubMed is as follows: (“heart failure” OR “cardiac failure”) AND (“readmission” OR “rehospitalization”) AND (“machine learning” OR “deep learning” OR “transformer models” OR “ensemble learning”).
For specific databases, free-text terms were searched in the title and abstract. This involved the use of a title and abstract tag in PubMed. ti, ab. in EMBASE, and TI (Title) and AB (Abstract) in CINAHL.

2.5. Selection Process

Two reviewers (NRA and HAM) independently screened the titles and abstracts of the retrieved records using Covidence systematic review software (2024). The same reviewers independently assessed the full texts of potentially eligible articles based on the predefined inclusion criteria. Disagreements during screening were resolved through discussion or, if necessary, by consultation with a third expert reviewer (CW). Studies combining traditional and ML methods were carefully evaluated for relevance. Automation tools were not used in the screening or selection process.
In total, 320 records were recorded. After removing duplicates, 285 unique records were screened. Of these, 251 were excluded based on the titles and abstracts because of an irrelevant study population (n = 105), irrelevant outcomes (n = 80), or irrelevant methods (n = 66). Studies were excluded if they enrolled participants outside the targeted demographic or clinical group, reported outcomes that were not aligned with the review objectives, or employed inconsistent study designs or methodologies.
The full texts of the remaining 34 articles were also assessed. Twelve articles were excluded at this stage: six were excluded due to an incorrect population (e.g., participants did not match age, diagnosis, or setting), three due to incorrect outcomes (primary or secondary outcomes of interest were not reported), and three due to incorrect study design (e.g., non-comparative or qualitative studies). The exclusion numbers and reasons for each stage are summarized in Figure 1 (PRISMA flow diagram).

2.6. Data Collection Process

Following the selection of studies, three reviewers (NRA, HAM, and CW) independently extracted data from the included articles using a standardized data extraction form. Any discrepancies in the extracted data were resolved through discussion and consensus between reviewers. If consensus could not be reached, a third reviewer (CW) was consulted to make a final decision. The extracted data included the study characteristics (first author, publication year, journal, and country), participant characteristics (number of participants, age range/mean, sex distribution, and method of HF diagnosis confirmation), ML model details (algorithm type/s and prediction window/s), and performance metrics (AUC-ROC, accuracy, precision, recall, and F1-score).

2.7. Data Items

The primary outcome was readmission for HF. Data were extracted from the reported model performance for predicting this outcome, focusing on the Area Under the Receiver Operating Characteristic Curve (AUC-ROC), chosen for its ability to assess discrimination across thresholds and robustness to class imbalance; precision, chosen for its ability to evaluate correct positive predictions and minimize false alarms; recall (sensitivity), chosen for its ability to measure the model’s ability to identify all positive cases, crucial for patient safety; and the F1-score, chosen as a balanced measure of precision and recall, particularly useful for imbalanced datasets. All reported results compatible with these outcome domains were obtained for each study. Other extracted variables included study and participant characteristics, ML algorithms, and prediction time windows.

2.8. Study Risk of Bias Assessment

Two reviewers independently evaluated the methodological quality and risk of bias of included studies using the modified CHARMS checklist [22], which is tailored for prediction model studies including those employing machine learning. This assessment focused on potential sources of bias relevant for ML research, such as participant selection, predictor and outcome measurement, the handling of data, feature selection, model development, and validation methods. Discrepancies were resolved through consensus or adjudication by a third reviewer.

2.9. Reporting Bias Assessment

Methods for formally assessing bias from missing results, such as funnel plot analysis, were not used in this review. This is because the review focused on a qualitative synthesis of ML model performance rather than meta-analysis, which is a prerequisite for funnel plots. In addition, the systematic assessment of outcome reporting bias across studies was not feasible because of the heterogeneous reported outcome metrics and a focus on machine learning model performance rather than specific clinical outcomes.

3. Results

3.1. Study Characteristics

The PRISMA flow diagram (Figure 1) shows that 320 records were obtained using the systematic search approach. Of these, 35 duplicates were excluded, which resulted in 285 unique records. Titles and abstracts were checked against the inclusion criteria, leading to the exclusion of 251 records that did not meet the criteria for the study population, outcomes, or methods. The full text of the remaining 34 articles was reviewed for eligibility. Twelve articles were excluded at this stage: six had an incorrect population, three had incorrect outcomes, and three had incorrect study designs. Eventually, 22 studies were included in the qualitative synthesis because they satisfied all inclusion criteria.

3.2. Synthesis of Findings

The qualitative synthesis included 22 studies published between 2015 and 2022, involving 463,270 patients across six countries. Most studies (15) originated in the U.S., with Italy and China contributing two each, and Australia, Korea, and Canada contributing one each. This geographical distribution confirms the bias identified in the introduction, potentially limiting generalizability to diverse global healthcare contexts. Across all included studies, the mean patient age was 73 years, with a nearly equal representation of women (49%).
Regarding the prediction windows, 14 studies (63.6%) focused on 30-day readmissions, involving patients aged 65–81.5 years with varied sex distributions (49–97.6% male). These studies reported moderate to high predictive accuracy, with AUC values ranging from 0.70 to 0.95 (median AUC: 0.82). One study examined both 90-day and 180-day readmissions (demographics not provided). Five studies (22.7%) tracked one-year readmission rates among patients aged 72–78 years, with a higher proportion of females (52.2–63%). One study reported on three-year outcomes, involving patients with a mean age of 72 years and equal sex proportions.
Supervised learning algorithms (SLA) consistently showed higher performance than unsupervised approaches across all prediction windows. Algorithms such as Random Forest and Gradient Boosting are frequently employed, often achieving high AUC values (often > 0.85), suggesting their suitability for capturing complex data relationships. Simple models, such as Logistic Regression, also perform adequately in certain scenarios, particularly with careful feature engineering. Model performance varied by prediction window; models predicting 30-day readmissions generally had higher AUCs than those predicting longer-term outcomes (e.g., 1-year), possibly because of the stronger influence of short-term post-discharge factors. SLA demonstrated AUCs between 0.70 and 0.99, while unsupervised methods showed AUCs ranging from 0.69 to 0.72 (see Table 1). The prevalence of 30-day prediction windows indicates the researchers’ focus in the reviewed studies.
To enhance clarity, a dedicated column indicating the type of machine learning algorithm (e.g., deep learning, traditional ML, unsupervised) used in each study has been added to Tables 3–5. This column specifies whether the approach is a traditional machine learning model (such as logistic regression, random forest, or gradient boosting), deep learning architecture (such as neural networks or attention-based models), or unsupervised approach (such as clustering or autoencoder-based representation learning).

3.3. Subgroup Analysis of Model Performance

Subgroup analyses were conducted to clarify AUC variability (Table 1) using the reported study characteristics. Models with a 30-day prediction window (n = 14) exhibited higher mean/median AUCs (median: 0.82; range: 0.70–0.95) than those using 1-year or longer windows (median: 0.77). Larger studies (>1000 participants) tended to achieve higher average AUCs (mean, 0.85) than smaller cohorts. Automated or systematic feature selection has also been associated with improved performance. The patterns are shown in Figure 2.
Table 2 shows that among the 17 studies included, 13 (76%) had a low overall risk of bias and three (18%) had a medium risk. One study [28] had a high risk for participant selection. The six studies that did not report (NR) risk of bias ratings in the adapted CHARMS checklist were not included in this table [11,19,23,24,36,39].
The evidentiary weights of the four studies with a moderate risk of bias should be interpreted with caution, as this level may affect the reliability and generalizability of their findings. In conclusion, the results of these studies should be considered within the limitations of their methodology.
Table 3 compares five pivotal ML studies on heart failure readmission prediction, detailing the diversity in algorithm types, prediction horizons, key findings, strengths, and limitations. SLA was most frequently applied, as reflected in Allam et al. [23] and Frizzell et al. [7], both of which evaluated 30-day readmissions and reported moderate discrimination (AUC 0.62 to 0.64) in U.S.-based cohorts. Angraal et al. employed a random forest ensemble to predict 3-year readmission and mortality [24], achieving an AUC of 0.76 but displaying a limited sample size and less demographic diversity. Jiang et al. utilized a novel unsupervised clustering approach to characterize dynamic risk trajectories [26], identifying new patient phenotypes with distinct readmission risks, though this approach posed clinical translation challenges. The comparison reveals that while sample size, geographic and demographic representation, and methodology enhance model robustness, generalizability, and external validation remain common limitations. These findings highlight the importance of carefully considering algorithm selection, validation strategy, and population diversity in designing and deploying ML models for HF readmission prediction.
Table 4 shows that HF readmission prediction studies attempted to improve predictions through various means, including comprehensive reviews, long-term prediction models, Electronic Health Record (EHR) data usage, algorithm comparisons, multicenter designs, and novel approaches. However, these studies had certain limitations that hindered their progress. These limitations include a U.S.-centric focus, homogeneous populations, single-center designs, reliance on billing codes, short prediction windows, and implementation complexity, which can limit the generalizability, accuracy in diverse populations, the consideration of social factors, and clinical application.
Table 5 highlights the range of ML methods used to predict HF readmission, including logistic regression, gradient boosting, random forest, deep learning, and unsupervised clustering. The prediction windows span from 30 days to three years, with short-term (30-day) models being the most common. Larger datasets typically enable more advanced ML models, whereas smaller or single-center cohorts favor simpler methods.
ML models use diverse feature sets, from EHR billing and administrative codes to more detailed clinical data, laboratory values, demographics, and imaging such as echocardiography and ECG signals. However, most studies are from the US and Europe, often relying on billing codes and less frequently incorporating psychosocial or socioeconomic factors, which limit generalizability.
Model performance varied: supervised models achieved AUCs between 0.70 and 0.99 (median AUC ≈ 0.82 for 30-day models), while unsupervised methods had AUCs of 0.69 to 0.72. Models with broader, heterogeneous, or externally validated cohorts showed lower but more generalizable accuracies. The observed variability in AUCs reflects differences in data sources, prediction windows, study populations, and evaluation methods, indicating the need for rigorous external validation, the incorporation of diverse features, and the adoption of standardized performance metrics for reliable clinical applications of ML-based readmission models.

4. Discussion

4.1. Strengths of ML in Predicting HF Readmissions

As shown (Table 1) by the high AUC values ranging from 0.70 to 0.99, this recent study highlights ML’s considerable promise in predicting HF readmissions within various healthcare contexts. The ability of ML to achieve accurate predictions in identifying patients likely to be readmitted validates descriptive ML-enabled resource allocation and targeted intervention frameworks. In addition to previous studies, ML models have outperformed standard statistical frameworks in predicting HF patient readmissions, with AUCs for 30-day readmission prediction ranging from 0.546 to 0.784 [44]. This finding further contributes to the consensus that ML-based strategies improve risk stratification and enhance clinical decision support across multiple healthcare environments [44,45]. Therefore, it is advised that these clinically validated ML models be integrated within workflows to streamline the identification of high-risk HF patients and that subsequent investigations concentrate on validating these ML models in varying patient populations and healthcare environments to establish model generalizability.

4.2. Effectiveness of Supervised Learning Methods

This review consistently demonstrated the effectiveness of SLA in predicting HF readmissions, with AUCs ranging from 0.70 to 0.99, highlighting their potential for clinical decision-making and resource allocation. For instance, Shameer et al. achieved an AUC of 0.72 using elastic net logistic regression [31]. However, reliance on billing codes may limit the inclusion of psychosocial predictors, thereby emphasizing the importance of feature selection and data sources. Supporting this, Sabouri et al. [44] and Mortazavi et al. [10] found that ML methods outperformed traditional models (AUCs 0.546–0.784 and up to 0.678, respectively). However, Jahangiri et al. [13] reported lower AUCs (0.576–0.607) using a nationwide database, suggesting that larger heterogeneous datasets may decrease model performance. Therefore, based on the observed variability in ML model performance for HF readmission prediction across studies (AUCs 0.576–0.99), future research should prioritize diverse data source integration, rigorous feature selection, and validation on heterogeneous datasets to enhance model accuracy and generalizability for clinical applications.

4.3. Expanding Unsupervised Learning Potential

Although supervised methods dominated the reviewed studies (82%), unsupervised approaches demonstrated unique capabilities for HF risk stratification. For example, Jiang et al. identified four novel patient phenotypes through clustering, including a “rapid decompensator” group with 22% 30-day readmission rates and “social determinant-driven” subgroups exhibiting a 3× higher readmission risk [26]. Furthermore, Chen et al. demonstrated that autoencoder-derived features boosted supervised model performance with 0.08 AUC, suggesting that hybrid approaches could maximize clinical utility [38].
However, the current unsupervised models face interpretability challenges. Lv et al. achieved 89% timing prediction accuracy through survival clustering but struggled to translate identified patterns into actionable clinical protocols [39]. This aligns with the findings of Flores et al., who reported that unsupervised clustering identifies prognostically distinct subgroups in coronary artery disease and improves risk stratification compared to traditional methods [46]. Bednarski et al. found that unsupervised learning outperformed quantitative ischemia assessment, revealing that conventional approaches for high-risk cardiac events were deficient [47]. In heart failure, self-supervised learning on echocardiography images has shown promise in effectively predicting event timing, even with limited data, surpassing established deep-learning architectures [48].
To bridge the gap between performance and clinical applicability, future research should prioritize developing interpretable unsupervised and hybrid models, focusing on methods that translate complex patterns into actionable insights for HF management.

5. Analysis of Machine Learning Approaches

5.1. Algorithm Types and Prediction Windows

Although SLA was the most common approach in the reviewed studies (consistent with Allam et al. [23] and Frizzell et al. [7], other ML strategies also contributed to HF readmission prediction. Allam et al. [23] and Frizzell et al. [7] demonstrated the effectiveness of supervised learning for predicting 30-day readmissions, achieving modest AUC values of 0.64 and 0.62, respectively. However, this contrasts with other studies, such as the work of Huang et al. [43], which highlighted the higher performance (e.g., AUC = 0.76) in US-centric supervised models, suggesting potential geographic and demographic biases.
Angraal et al. highlighted the value of ensemble methods by using a random forest model to predict long-term (3-year) outcomes with an AUC of 0.76, suggesting their ability to capture complex relationships over extended periods [24]. Furthermore, several studies have explored the utility of unsupervised learning. Jiang et al. employed this approach to identify dynamic readmission risk trajectories, offering a different perspective focused on patterns and changes in risk over time [26]. In contrast to the findings of Jiang et al. [26], Lv et al. [39] encountered challenges in translating unsupervised patterns into actionable clinical protocols, reflecting broader concerns about the interpretability of these methods. Friz et al. reported a lower AUC (0.69) in Italy using supervised models with LACE index variables, further underscoring how regional healthcare practices can influence model performance [37].

5.2. Strengths and Limitations

Table 3 summarizes the strengths and limitations of each study, including sample size, methodology, and generalizability. Allam et al. used a large sample size, which increased generalizability [23]. Angraal et al. used a sophisticated algorithm but a smaller sample, potentially limiting generalizability [24]. These points are crucial for interpreting findings and identifying future research areas.

5.3. Additional Performance Metrics

The AUC measures overall accuracy, but precision (minimizing false positives) and recall (identifying high-risk patients) are critical for clinical utility. Researchers have found that precision–recall curves provide additional insight into imbalanced cohorts, with optimal F1 scores occurring at probability thresholds 18–32% higher than the standard 0.5 cutoffs. This underscores the importance of considering multiple performance metrics, particularly with imbalanced datasets common to HF readmission prediction.

6. Methodological Considerations

6.1. Integration Recommendation for Methodological Considerations

ML shows promise in predicting HF readmissions; however, methodological limitations exist. Geographic and population biases are evident: Huang et al.’s review showed a U.S.-centric focus [43], and Angraal et al.’s study used a homogeneous cohort (88% White), potentially limiting the generalizability to diverse populations and non-Western healthcare systems (Table 4) [24]. Data source and feature selection issues were highlighted by Shameer et al.’s use of single-center data and prioritization of billing codes over clinical narratives, potentially overlooking crucial social determinants [31]. The validation scope varies, with many studies relying on single-center validation; multicenter studies, such as Frizzell et al. [7], offer more robust generalizability but remain US-limited. Prediction windows ranged from 30 days to 3 years, reflecting HF readmission complexity, but complicating direct model comparisons. Algorithm selection varied widely, from logistic regression to complex neural networks, showing ML versatility but highlighting the need for standardized performance comparisons. These limitations underscore the need for multicenter validation across diverse healthcare ecosystems, the incorporation of socioeconomic variables and clinical narratives, the standardized reporting of cohort demographics and model performance, and the exploration of both short- and long-term prediction windows. Addressing these constraints can enhance future ML clinical utility and the generalizability of future ML models for predicting HF readmissions across diverse patient populations and healthcare settings.

6.2. Factors Contributing to Discrepancies in ML Model Performance

A significant discrepancy existed in the reported ML model performance in predicting heart failure remissions, with AUC values ranging from 0.69 0.99. The variability stems from several factors. First, data heterogeneity is crucial; models trained on the U.S. EHR systems may differ significantly from those that use variables such as family support, as seen in some Chinese studies. Second, temporal factors, such as predicting 30-day or 3-year readmissions, necessitate distinct algorithm architectures and influence the performance. Finally, metric selection impacts performance evaluation, and while AUC is commonly reported, precision–recall curves offer better insight into clinical risk thresholds, especially in imbalanced cohorts, where F1-optimized thresholds can be significantly higher than the default 0.5 cutoffs. To mitigate this variability and enhance comparability, standardizing evaluation protocols according to guidelines such as TRIPOD-AI is essential for preserving clinical relevance.

7. Clinical Context

7.1. Prediction Windows and the Complexity of HF Readmissions

The variation in prediction windows across studies reflects the complex nature of HF readmissions and the evolving healthcare needs. Most studies focused on 30-day readmissions, but some extended predictions to 90 days, 180 days, one year, or even three years. This range highlights the importance of both short- and long-term prediction models in the management of patients with HF. Short-term forecasts (30–90 days) are crucial for immediate post-discharge care as patients at this stage are the most vulnerable. For instance, Wideqvist et al. reported that up to 22% of patients with HF are readmitted within one month, emphasizing timely interventions [49]. Long-term forecasts (six months to three years) provide insights into HF’s chronic nature of HF and guide ongoing care planning. Angraal et al. used random forests to predict three-year outcomes with an AUC of 0.76 [24]. The choice of prediction timeframe should be guided by specific healthcare system goals, data availability, and algorithm performance across various time horizons.

7.2. Patient Demographics and Risk Factors

The reviewed studies offer insights into HF patient demographics at risk of readmission. Patients aged 65–81.5 years (average 73 years) were most likely to be readmitted. Sex distributions varied, aligning with Savarese et al. [2] and Lam et al. [50] regarding higher HF prevalence in older adults and potential gender-based risk factors.
Specifically, 14 studies (63.6%) focused on 30-day readmissions, involving patients aged 65–81.5 years with varied sex distributions (49–97.6% male). Five studies (22.7%) tracked one-year readmission rates among patients aged 72–78 years, with a higher proportion of females (52.2–63%).
These findings underscore the importance of long-term patient monitoring and targeted interventions considering age, sex, and other demographic variables. Healthcare providers should tailor care plans to address the specific needs of different patient subgroups including older adults and those with sex-specific risk factors.

7.3. Addressing Geographic Considerations

Geographic bias is a significant challenge in ML-based HF readmission models, with a predominance of US-centric research. In this review, 68% of the included studies originated in the United States, constraining their direct applicability to countries with differing healthcare infrastructures, patient demographics, and data availability. Comparative analysis revealed notable variability in model performance across regions: U.S. models leveraging comprehensive EHR data, such as Golas et al. [25], achieved higher AUCs (0.76) for 30-day prediction than Italian models, such as Friz et al. [37], which relied on LACE index variables and reported lower AUCs (0.69). This suggests that regional differences in discharge practices and data structure significantly influence predictive accuracy. Furthermore, Chinese studies, exemplified by Lv et al. [39], have incorporated family support as a predictor—an important contextual factor typically omitted from Western models—highlighting the value of integrating culturally and regionally relevant variables. To enhance the transferability and equity of ML implementation worldwide, future models should undergo local calibration, address regional differences in data infrastructure, and include locally significant predictors. Multicenter international collaborations are encouraged, with the adoption of standardized outcome definitions and flexibility for local adaptation, as outlined in frameworks such as WHO STEPS, to ensure robust and globally applicable predictive tools.

8. Translational Implications and Implementation Considerations

8.1. Implications for Healthcare Organizations

The findings of this systematic review have important implications for healthcare organizations that use ML to predict and manage HF readmissions. Supervised ML models can estimate the likelihood of hospitalization in patients with HF, enabling risk stratification and targeted interventions in high-risk individuals. However, ethical implementation is crucial and requires strong data governance and continuous monitoring to mitigate potential biases and ensure fairness in patient care. Collaboration among data scientists, clinicians, and IT teams is essential to overcome the challenges of infrastructure investment, workflow integration, and ethical considerations. The effective implementation of ML models can potentially reduce readmission rates, improve patient outcomes, and optimize resource allocation.

8.2. Ethical Considerations in ML Implementation

This review did not systematically extract or synthesize evidence regarding specific ethical frameworks, explainability tools, or bias auditing mechanisms (e.g., AI Fairness 360, Google What-If Tool, LIME, or SHAP). Any mention of these methods is provided solely as a forward-looking recommendation based on broader machine learning best practices rather than as findings derived from the included studies. The deployment of ML models for heart failure readmission prediction presents notable ethical challenges, including the risk of algorithmic bias, particularly when minority populations are underrepresented in the training data, and privacy risks associated with unstructured EHR data. For instance, some studies have shown reduced predictive performance for minority groups and that a portion of the predictive power may depend on potentially identifiable free-text fields. Moreover, high predictive accuracy can sometimes lead to overreliance on models, potentially overriding clinician judgment in borderline cases. To address these concerns, future research and implementation efforts should prioritize pre-deployment fairness audits, emphasize model transparency and explainability, and develop robust patient consent protocols for ML-driven clinical care. Although specific fairness and interpretability tools have not been systematically reviewed here, their adoption and ongoing assessment remain essential for trustworthy and equitable ML-enabled decision-making support in heart failure care.

8.3. Infrastructure and Clinical Integration Challenges

Beyond the model performance, successful ML implementation for heart failure readmission prediction faces substantial technical and operational barriers. Three core challenges were identified: (1) EHR interoperability issues (affecting 68% of studies), hindering data integration; (2) operational and computational cost burdens, presenting financial barriers for hospitals; and (3) workflow disruptions, as ML tools may increase staff workloads and require the adjustment of established clinical processes. To address these challenges, implementation blueprints, such as the modular API architecture proposed by Golas et al. [25], can streamline integration compared with traditional monolithic systems. Effective deployments also require the alignment of model outputs with clinician workflows, such as triggering nurse-led interventions at empirically validated risk thresholds, and the incorporation of regular model updates through continuous feedback loops. Resource-limited settings may necessitate the development of lightweight models that function with minimal infrastructure while maintaining an acceptable predictive performance.

8.4. Implementation Roadmap for Clinical Integration

Successfully integrating ML models for HR readmission prediction into clinical practice requires that the key implementation barriers be addressed. These include technical challenges such as adopting modular API architectures to improve EHR interoperability and ensuring the use of feasible computational requirements for real-world deployment. Operational challenges involve integrating models into clinical workflows, such as facilitating nurse-led interventions at empirically validated risk thresholds and establishing regular model updates through clinician feedback. Ethical considerations are also crucial, necessitating pre-deployment fairness audits using tools such as AI Fairness 360, and developing patient consent protocols for ML-driven care adjustments.

8.5. Barriers to Clinical Adoption: Technical vs. Sociocultural Perspectives

Despite advances in predictive accuracy, the clinical adoption of machine learning (ML) models for heart failure (HF) is limited by technical and sociocultural barriers. Technical challenges include EHR interoperability, significant computational and financial demands, and difficulties in integrating ML tools into existing clinical workflows. Sociocultural barriers include clinician skepticism toward “black box” algorithms, differing levels of digital literacy among clinical staff, data privacy concerns, and the need for culturally and regionally tailored solutions. These challenges are not unique to HF; as highlighted by Cersosimo et al. [51], similar implementation barriers are observed in arrhythmia detection and automated echocardiography analysis. Overcoming these obstacles requires not only robust technical validation, but also strong clinician engagement, interpretable AI outputs, and effective organizational change management. In the context of coronary artery disease phenotyping, Ajiboye et al. demonstrated that collaborative interpretation and the demonstration of clinical value improve clinician trust. Collectively, these multidisciplinary experiences emphasize that efforts to address EHR integration, resource constraints, clinician education, explainability, and local adaptation are essential for scalable and sustainable ML integration in HF and across the broader spectrum of cardiovascular care.

8.6. Methodological Recommendations for Future Research

For future research, standardizing evaluation metrics, such as reporting AUC-ROC together with precision, recall, F1-score, and calibration plots, would enable a more complete assessment of model performance, especially with imbalanced datasets common in HF readmission studies. It is also recommended that future studies incorporate clinical text variables from discharge summaries or physician notes as well as socioeconomic and psychosocial factors to capture important predictors often missed by structured data alone. Robust external validation using data from multiple institutions or regions is essential to confirm that the models generalize beyond their developmental settings. The adoption of established reporting frameworks such as TRIPOD-AI can further improve consistency, transparency, and reproducibility. Implementing these methodological enhancements will help guide the development of more practical, reliable, and clinically meaningful ML models for predicting readmissions for HF.

8.7. Limitations and Future Directions

This review highlights several recurring limitations that constrain the generalizability and comparability of current ML models for HF readmission prediction (see Table 4). Most included studies relied predominantly on U.S.-based or single-center cohorts, which restricts the applicability of the findings to other healthcare systems and diverse patient populations. Dependence on administrative billing codes and structured EHR data with the infrequent inclusion of unstructured clinical notes or social determinants may result in important predictors being missed. External and multicenter validations have rarely been performed, raising concerns regarding overfitting and the robustness of model transportability. Additionally, substantial methodological heterogeneity exists, with variations in algorithm choice, predictor sets, prediction time windows, and outcome definitions, all of which complicate direct comparisons across studies. As detailed in Table 5, this review included studies published up to March 2024 and therefore did not account for more recent methodological advances, such as transformer architecture and federated learning. Addressing these limitations will require future research to incorporate multicenter and regionally diverse cohorts, expand predictor variables to include clinical texts and socioeconomic features, standardize outcome measures and reporting frameworks, and rigorously evaluate new methodological approaches. These steps are critical for the development of more robust, reliable, and equitable machine learning models for predicting HF readmissions.

9. Conclusions

This systematic review demonstrates ML’s potential to predict hospital readmissions among patients with HF. SLA showed a promising performance, with AUC values ranging from 0.70–0.99. However, several key areas require attention for future research and implementation. These include the need for demographic audits to address potential biases, temporal validation to account for evolving treatments, the implementation of science to bridge the gap between research and practice, standardized evaluation methods, and diverse geographic representations to enhance generalizability. Addressing these priorities will facilitate the development of more robust and equitable ML models and ultimately improve patient outcomes and reduce healthcare costs.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/biomedicines13092111/s1.

Author Contributions

N.A. and E.P.-a. were assigned the roles of conceptualization, investigation, writing—original draft, and writing—review and editing, suggesting a leading role in the research project, from its conception to the writing of the manuscript. R.M.J. and P.P. are assigned roles related to data management, analysis, and visualization, implying a key role in handling and interpreting the research data. H.A. and H.C. are assigned roles related to methodology, software, and validation, indicating a significant contribution to the technical aspects of the study. A.G. and S.A. are assigned to supervision and project administration, indicating a role in overseeing the project and ensuring its smooth execution. A.A.M.A. and N.A.H.A. are assigned to writing—review and editing, implying a contribution to refining the manuscript. H.A. is assigned to investigation and validation, suggesting involvement in conducting the research and verifying the findings. S.T. and K.A. are assigned to software and formal analysis, indicating a contribution to the technical aspects of data analysis. F.A.A. and R.M.J. are assigned to resources and data curation, suggesting a role in providing resources and managing the data. All authors have read and agreed to the published version of the manuscript.

Funding

No funding agency in the public, commercial, or not-for-profit sectors have provided a specific grant to support this research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This review was registered in the International Prospective Register of Systematic Reviews (ID #CRD42021247198) and contains a clear and detailed summary of the review protocol. This data is available upon request.

Conflicts of Interest

The authors declare no competing financial interests or personal relationships that may have influenced the work reported in this study.

References

  1. Shahim, B.; Kapelios, C.J.; Savarese, G.; Lund, L.H. Global Public Health Burden of Heart Failure: An Updated Review. Card. Fail. Rev. 2023, 9, e11. [Google Scholar] [CrossRef] [PubMed]
  2. Savarese, G.; Lund, L.H. Global Public Health Burden of Heart Failure. Card. Fail. Rev. 2017, 3, 7. [Google Scholar] [CrossRef] [PubMed]
  3. McLaren, D.P.; Jones, R.; Plotnik, R.; Zareba, W.; McIntosh, S.; Alexis, J.; Chen, L.; Block, R.; Lowenstein, C.J.; Kutyifa, V. Prior hospital admission predicts thirty-day hospital readmission for heart failure patients. Cardiol. J. 2016, 23, 155–162. [Google Scholar] [CrossRef]
  4. Krumholz, H.M.; Chaudhry, S.I.; Spertus, J.A.; Mattera, J.A.; Hodshon, B.; Herrin, J. Do Non-Clinical Factors Improve Prediction of Readmission Risk? JACC Heart Fail. 2016, 4, 12–20. [Google Scholar] [CrossRef] [PubMed]
  5. Keeney, T.; Jette, D.U.; Cabral, H.; Jette, A.M. Frailty and Function in Heart Failure: Predictors of 30-Day Hospital Readmission? J. Geriatr. Phys. Ther. 2021, 44, 101–107. [Google Scholar] [CrossRef]
  6. Lv, Q.; Zhang, X.; Wang, Y.; Xu, X.; He, Y.; Liu, J.; Chang, H.; Zhao, Y.; Zang, X. Multi-trajectories of symptoms and their associations with unplanned 30-day hospital readmission among patients with heart failure: A longitudinal study. Eur. J. Cardiovasc. Nurs. 2024, 23, 737–745. [Google Scholar] [CrossRef]
  7. Frizzell, J.D.; Liang, L.; Schulte, P.J.; Yancy, C.W.; Heidenreich, P.A.; Hernandez, A.F.; Bhatt, D.L.; Fonarow, G.C.; Laskey, W.K. Prediction of 30-Day All-Cause Readmissions in Patients Hospitalized for Heart Failure: Comparison of Machine Learning and Other Statistical Approaches. JAMA Cardiol. 2017, 2, 204–209. [Google Scholar] [CrossRef]
  8. Ahmad, T.; Lund, L.H.; Rao, P.; Ghosh, R.; Warier, P.; Vaccaro, B.; Dahlström, U.; O’Connor, C.M.; Felker, G.M.; Desai, N.R. Machine Learning Methods Improve Prognostication, Identify Clinically Distinct Phenotypes, and Detect Heterogeneity in Response to Therapy in a Large Cohort of Heart Failure Patients. J. Am. Heart Assoc. 2018, 7, e008081. [Google Scholar] [CrossRef]
  9. Sarker, I.H. Machine Learning: Algorithms, Real-World Applications and Research Directions. SN Comput. Sci. 2021, 2, 160. [Google Scholar] [CrossRef]
  10. Mortazavi, B.J.; Downing, N.S.; Bucholz, E.M.; Dharmarajan, K.; Manhapra, A.; Li, S.-X.; Negahban, S.N.; Krumholz, H.M. Analysis of Machine Learning Techniques for Heart Failure Readmissions. Circ. Cardiovasc. Qual. Outcomes 2016, 9, 629–640. [Google Scholar] [CrossRef]
  11. Jing, L.; Ulloa Cerna, A.E.; Good, C.W.; Sauers, N.M.; Schneider, G.; Hartzel, D.N.; Leader, J.B.; Kirchner, H.L.; Hu, Y.; Riviello, D.M.; et al. A Machine Learning Approach to Management of Heart Failure Populations. JACC Heart Fail. 2020, 8, 578–587. [Google Scholar] [CrossRef]
  12. Yu, M.-Y.; Son, Y.-J. Machine learning–based 30-day readmission prediction models for patients with heart failure: A systematic review. Eur. J. Cardiovasc. Nurs. 2024, 23, 711–719. [Google Scholar] [CrossRef] [PubMed]
  13. Jahangiri, S.; Abdollahi, M.; Rashedi, E.; Azadeh-Fard, N. A machine learning model to predict heart failure readmission: Toward optimal feature set. Front. Artif. Intell. 2024, 7, 1363226. [Google Scholar] [CrossRef]
  14. Hernandez, L.M.; Blazer, D.G. The Impact of Social and Cultural Environment on Health; National Academies Press: Washington, DC, USA, 2020. Available online: https://www.ncbi.nlm.nih.gov/books/NBK19924/ (accessed on 24 January 2024).
  15. Dawkins, B.; Renwick, C.; Ensor, T.; Shinkins, B.; Jayne, D.; Meads, D. What Factors Affect Patients’ Ability to Access healthcare? An Overview of Systematic Reviews. Trop. Med. Int. Health 2021, 26, 1177–1188. [Google Scholar] [CrossRef] [PubMed]
  16. Pavlou, M.; Ambler, G.; Seaman, S.R.; Guttmann, O.; Elliott, P.; King, M.; Omar, R.Z. How to develop a more accurate risk prediction model when there are few events. BMJ 2015, 351, h3868. [Google Scholar] [CrossRef] [PubMed]
  17. Stødle, K.; Flage, R.; Guikema, S.D.; Aven, T. Data-driven predictive modeling in risk assessment: Challenges and directions for proper uncertainty representation. Risk Anal. 2023, 43, 2644–2658. [Google Scholar] [CrossRef]
  18. Kleiner Shochat, M.; Fudim, M.; Shotan, A.; Blondheim, D.S.; Kazatsker, M.; Dahan, I.; Asif, A.; Rozenman, Y.; Kleiner, I.; Weinstein, J.M.; et al. Prediction of readmissions and mortality in patients with heart failure: Lessons from the IMPEDANCE-HF extended trial. ESC Heart Fail. 2018, 5, 788–799. [Google Scholar] [CrossRef]
  19. Hu, Y.; Ma, F.; Hu, M.; Shi, B.; Pan, D.; Ren, J. Development and validation of a machine learning model to predict the risk of readmission within one year in HFpEF patients: Short title: Prediction of HFpEF readmission. Int. J. Med. Inform. 2024, 194, 105703. [Google Scholar] [CrossRef]
  20. Yordanov, T.R.; Lopes, R.R.; Ravelli, A.C.; Vis, M.; Houterman, S.; Marquering, H.; Abu-Hanna, A. An integrated approach to geographic validation helped scrutinize prediction model performance and its variability. J. Clin. Epidemiol. 2023, 157, 13–21. [Google Scholar] [CrossRef]
  21. PRISMA. Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA). 2020. Available online: https://www.prisma-statement.org/ (accessed on 24 January 2024).
  22. Moons, K.G.M.; de Groot, J.A.H.; Bouwmeester, W.; Vergouwe, Y.; Mallett, S.; Altman, D.G.; Reitsma, J.B.; Collins, G.S. Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies: The CHARMS Checklist. PLoS Med. 2014, 11, e1001744. [Google Scholar] [CrossRef]
  23. Allam, A.; Nagy, M.; Thoma, G.; Krauthammer, M. Neural networks versus Logistic regression for 30 days all-cause readmission prediction. Sci. Rep. 2019, 9, 9277. [Google Scholar] [CrossRef]
  24. Angraal, S.; Mortazavi, B.J.; Gupta, A.; Khera, R.; Ahmad, T.; Desai, N.R.; Jacoby, D.L.; Masoudi, F.A.; Spertus, J.A.; Krumholz, H.M. Machine Learning Prediction of Mortality and Hospitalization in Heart Failure with Preserved Ejection Fraction. JACC Heart Fail. 2020, 8, 12–21. [Google Scholar] [CrossRef]
  25. Golas, S.B.; Shibahara, T.; Agboola, S.; Otaki, H.; Sato, J.; Nakae, T.; Hisamitsu, T.; Kojima, G.; Felsted, J.; Kakarmath, S.; et al. A machine learning model to predict the risk of 30-day readmissions in patients with heart failure: A retrospective analysis of electronic medical records data. BMC Med. Inform. Decis. Mak. 2018, 18, 44. [Google Scholar] [CrossRef]
  26. Jiang, W.; Siddiqui, S.; Barnes, S.; Barouch, L.A.; Korley, F.; Martinez, D.A.; Toerper, M.; Cabral, S.; Hamrock, E.; Levin, S. Readmission risk trajectories for patients with heart failure using a dynamic prediction approach: Retrospective study. JMIR Med. Inform. 2019, 7, e1475. [Google Scholar] [CrossRef] [PubMed]
  27. Mahajan, S.M.; Mahajan, A.S.; King, R.; Negahban, S. Predicting Risk of 30-Day Readmissions Using Two Emerging Machine Learning Methods. Stud. Health Technol. Inform. 2018, 250, 250–255. Available online: https://pubmed.ncbi.nlm.nih.gov/29857454/ (accessed on 24 January 2024). [PubMed]
  28. Mahajan, S.M.; Ghani, R. Using ensemble machine learning methods for predicting risk of readmission for heart failure. In MEDINFO 2019: Health and Wellbeing e-Networks for All; IOS Press: Amsterdam, The Netherlands, 2019; pp. 243–247. [Google Scholar] [CrossRef]
  29. Pishgar, M.; Harford, S.; Theis, J.; Galanter, W.; Rodríguez-Fernández, J.M.; Chaisson, L.H.; Zhang, Y.; Trotter, A.; Kochendorfer, K.M.; Boppana, A.; et al. A process mining-deep learning approach to predict survival in a cohort of hospitalized COVID-19 patients. BMC Med. Inform. Decis. Mak. 2022, 22, 194. [Google Scholar] [CrossRef] [PubMed]
  30. Desai, R.J.; Wang, S.V.; Vaduganathan, M.; Evers, T.; Schneeweiss, S. Comparison of machine learning methods with traditional models for use of administrative claims with electronic medical records to predict heart failure outcomes. JAMA Netw. Open 2020, 3, e1918962. [Google Scholar] [CrossRef]
  31. Shameer, K.; Johnson, K.W.; Yahi, A.; Miotto, R.; Li, L.I.; Ricks, D.; Jebakaran, J.; Kovatch, P.; Sengupta, P.P.; Gelijns, S.; et al. Predictive modeling of hospital readmission rates using electronic medical record-wide machine learning: A case-study using Mount Sinai heart failure cohort. Pac. Symp. Biocomput. 2017, 22, 276–287. [Google Scholar] [CrossRef]
  32. Tukpah, A.M.C.; Cawi, E.; Wolf, L.; Nehorai, A.; Cummings-Vaughn, L. Development of an Institution-Specific Readmission Risk Prediction Model for Real-time Prediction and Patient-Centered Interventions. J. Gen. Intern. Med. 2021, 36, 3910–3912. [Google Scholar] [CrossRef]
  33. Turgeman, L.; May, J.H. A mixed-ensemble model for hospital readmission. Artif. Intell. Med. 2016, 72, 72–82. [Google Scholar] [CrossRef]
  34. Yu, S.; Farooq, F.; Van Esbroeck, A.; Fung, G.; Anand, V.; Krishnapuram, B. Predicting readmission risk with institution-specific prediction models. Artif. Intell. Med. 2015, 65, 89–96. [Google Scholar] [CrossRef]
  35. Sarijaloo, F.; Park, J.; Zhong, X.; Wokhlu, A. Predicting 90-day acute heart failure readmission and death using machine learning-supported decision analysis. Clin. Cardiol. 2021, 44, 230–237. [Google Scholar] [CrossRef]
  36. Lorenzoni, G.; Sabato, S.S.; Lanera, C.; Bottigliengo, D.; Minto, C.; Ocagli, H.; De Paolis, P.; Gregori, D.; Iliceto, S.; Pisanò, F. Comparison of Machine Learning Techniques for Prediction of Hospitalization in Heart Failure Patients. J. Clin. Med. 2019, 8, 1298. [Google Scholar] [CrossRef] [PubMed]
  37. Friz, P.H.; Esposito, V.; Marano, G.; Primitz, L.; Bovio, A.; Delgrossi, G.; Bombelli, M.; Grignaffini, G.; Monza, G.; Boracchi, P. Machine learning and LACE index for predicting 30-day readmissions after heart failure hospitalization in elderly patients. Intern. Emerg. Med. 2022, 17, 1727–1737. [Google Scholar] [CrossRef] [PubMed]
  38. Chen, P.; Dong, W.; Wang, J.; Lu, X.; Kaymak, U.; Huang, Z. Interpretable clinical prediction via attention-based neural network. BMC Med. Inform. Decis. Mak. 2020, 20, 131. [Google Scholar] [CrossRef] [PubMed]
  39. Lv, H.; Yang, X.; Wang, B.; Wang, S.; Du, X.; Tan, Q.; Hao, Z.; Xia, Y. Machine learning–driven models to predict prognostic outcomes in patients hospitalized with heart failure using electronic health records: Retrospective study. J. Med. Internet Res. 2021, 23, e24996. [Google Scholar] [CrossRef]
  40. Bat-Erdene, B.I.; Zheng, H.; Son, S.H.; Lee, J.Y. Deep learning-based prediction of heart failure rehospitalization during 6, 12, 24-month follow-ups in patients with acute myocardial infarction. Health Inform. J. 2022, 28, 14604582221101529. [Google Scholar] [CrossRef]
  41. Awan, S.E.; Bennamoun, M.; Sohel, F.; Sanfilippo, F.M.; Chow, B.J.; Dwivedi, G. Feature selection and transformation by machine learning reduce variable numbers and improve prediction for heart failure readmission or death. PLoS ONE 2019, 14, e0218760. [Google Scholar] [CrossRef]
  42. Sharma, V.; Kulkarni, V.; Mcalister, F.; Eurich, D.; Keshwani, S.; Simpson, S.H.; Voaklander, D.; Samanani, S. Predicting 30-Day Readmissions in Patients with Heart Failure Using Administrative Data: A Machine Learning Approach. J. Card. Fail. 2021, 28, 710–722. [Google Scholar] [CrossRef]
  43. Huang, Y.; Talwar, A.; Chatterjee, S.; Aparasu, R.R. Application of machine learning in predicting hospital readmissions: A scoping review of the literature. BMC Med. Res. Methodol. 2021, 21, 96. [Google Scholar] [CrossRef]
  44. Sabouri, M.; Rajabi, A.; Hajianfar, G.; Gharibi, O.; Mohebi, M.; Avval, A.; Naderi, N.; Shiri, I. Machine learning based readmission and mortality prediction in heart failure patients. Sci. Rep. 2023, 13, 18671. [Google Scholar] [CrossRef]
  45. Hidayaturrohman, Q.A.; Hanada, E. Predictive Analytics in Heart Failure Risk, Readmission, and Mortality Prediction: A Review. Cureus 2024, 16, 11. [Google Scholar] [CrossRef]
  46. Flores, A.M.; Schuler, A.; Eberhard, A.V.; Olin, J.W.; Cooke, J.P.; Leeper, N.J.; Shah, N.H.; Ross, E.G. Unsupervised Learning for Automated Detection of Coronary Artery Disease Subgroups. J. Am. Heart Assoc. 2021, 10, e021976. [Google Scholar] [CrossRef] [PubMed]
  47. Bednarski, B.; Williams, M.C.; Pieszko, K.; Miller, R.J.H.; Huang, C.; Kwiecinski, J.; Sharir, T.; Di Carli, M.; Fish, M.B.; Ruddy, T.D.; et al. Unsupervised machine learning improves risk stratification of patients with visual normal SPECT myocardial perfusion imaging assessments. Eur. Heart J. 2022, 43 (Suppl. 2), ehac544.300. [Google Scholar] [CrossRef]
  48. Bell-Navas, A.; Villalba-Orero, M.; Lara-Pezzi, E.; Garicano-Mena, J.; Clainche, S.L. Heart Failure Prediction using Modal Decomposition and Masked Autoencoders for Scarce Echocardiography Databases. arXiv 2025, arXiv:2504.07606. [Google Scholar] [CrossRef]
  49. Wideqvist, M.; Cui, X.; Magnusson, C.; Schaufelberger, M.; Fu, M. Hospital readmissions of patients with heart failure from real world: Timing and associated risk factors. ESC Heart Fail. 2021, 8, 1388–1397. [Google Scholar] [CrossRef] [PubMed]
  50. Lam, C.S.P.; Piña, I.L.; Zheng, Y.; Bonderman, D.; Pouleur, A.-C.; Saldarriaga, C.; Pieske, B.; Blaustein, R.O.; Nkulikiyinka, R.; Westerhout, C.M.; et al. Age, Sex, and Outcomes in Heart Failure with Reduced EF: Insights from the VICTORIA Trial. JACC Heart Fail. 2023, 11, 1246–1257. [Google Scholar] [CrossRef]
  51. Cersosimo, A.; Zito, E.; Pierucci, N.; Matteucci, A.; La Fazia, V.M. A Talk with ChatGPT: The Role of Artificial Intelligence in Shaping the Future of Cardiology and Electrophysiology. J. Pers. Med. 2025, 15, 205. [Google Scholar] [CrossRef]
Figure 1. PRISMA flow diagram of study selection process.
Figure 1. PRISMA flow diagram of study selection process.
Biomedicines 13 02111 g001
Figure 2. Meta-visualization of mean AUC values by prediction window, sample size, and feature selection methods for heart failure readmission prediction. Legend: blue for prediction window, orange for sample size, and tan for feature selection method.
Figure 2. Meta-visualization of mean AUC values by prediction window, sample size, and feature selection methods for heart failure readmission prediction. Legend: blue for prediction window, orange for sample size, and tan for feature selection method.
Biomedicines 13 02111 g002
Table 1. Summary of ML models for HF readmission prediction (2015–2022).
Table 1. Summary of ML models for HF readmission prediction (2015–2022).
Study/CountryNumber of Patients% of GenderAverage Age (yrs)AlgorithmAUCAccuracyPrecisionReadmission Days
Allam et al. [23], USA272,77849 females73SLA0.64NRNR30 days
Angraal et al. [24], USA176750 females72SLA0.76NRNR3-year
Frizzell et al. [7], USA238,58154.5 females80SLA0.62NRNR30 days
Golas et al. [25], USA28,03153 males65SLA0.70NRNR30 days
Jiang et al. [26], USA53464 females75ULA0.73 NRNR30 days
Mahajan et al. [27], USA177897.6 males72ULA0.72NRNR30 days
Mahajana et al. [28], USA36,245NANAELT0.70NRNR30 days
Pishgar et al. [29], USA38,59746.3 females70SLA0.930.840.8930 days
Desai et al. [30], USA950245 female78SLA0.76NRNR1-year
Shameer et al. [31], USA1068NANASLA0.78NRNR30 days
Tukpah et al. [32], USA965NANASLA0.690.780.5830 days
Turgeman & May [33], USA965NA79SLANR0.85NR30 days
Yu et al. [34],
USA
20,588NA65SLA0.65NRNR30 days
Sarijaloo et al. [35], USA2441NA65SLA0.750.75NR90 days
Mortazavi et al. [10], USA1653NANASLA0.67NRNR180 days
Lorenzoni et al. [36], Italy38060 females73SLA0.810.81NR1-year
Friz et al. [37], Italy 307955.3 females81SLA0.740.600.7030 days
Chen et al. [38], China736NA72SLA0.670.670.711-year
Lv et al. [39], China13,60252 females72SLA0.810.770.761-year
Bat-Erdene et al. [40], Korea11,011NANASLA0.990.990.981-year
Awan et al. [41], Australia 10,75749 males81SLA0.6248.420.7030 days
Sharma et al. [42],
Canada
984556 males71SLA0.65NRNR30 days
SLA: supervised learning algorithm; ULA: unsupervised learning algorithm; ELT: ensemble learning techniques; NR: not reported; NA: not available.
Table 2. Results of adapted CHARMS checklist for assessing risk of bias in included studies.
Table 2. Results of adapted CHARMS checklist for assessing risk of bias in included studies.
Study Name Participant Selection Predictor Assessment Outcome Assessment Model DevelopmentAnalysis
Allam et al. [23] LLLLL
Awan et al. [41] MLLLL
Frizzell et al. [7] LLLLL
Golas et al. [25] MLLLL
Jiang et al. [26] LLLLL
Mahajan et al. [27]LLLLL
Mahajana et al. [28] HLLLL
Pishgar et al. [29]LLLLL
Polo Friz et al. [37]MLLLL
Shameer et al. [31] LLLLL
Sharma et al. [42] LLLLL
Turgeman & May [33] LLLLL
Yu et al. [34] LLLLL
Sarijaloo et al. [35] LLLLL
Mortazavi et al. [10]LLLLL
Bat-Erdene et al. [40]LLLLL
Chen et al. [38]LLLLL
Legend: L = Low; M = Medium; H = High.
Table 3. Comparison of key machine learning studies.
Table 3. Comparison of key machine learning studies.
StudyPrediction WindowSample SizeKey FindingsStrengthsLimitationsPopulation Diversity (% Non-White)
Allam et al. [23]30 daysLargeDeep Learning (Neural Network), Traditional ML (Logistic Regression)Neural networks showed slightly better performance than logistic regression for 30-day readmission risk.Large dataset; comparison of deep learning and traditional approaches.Homogeneous cohort; dependent on billing codes.
Frizzell et al. [7]30 daysLarge multicenterTraditional ML (Logistic Regression, Random Forest, Gradient Boosting)Ensemble methods did not substantially outperform logistic regression, with AUC ~0.62–0.72.Multicenter design; rigorous validation.U.S.-centric data; moderate discrimination.
Angraal et al. [24]3 yearsMediumEnsemble (Random Forest)Random forest achieved reasonable predictive power (AUC 0.76) for long-term (3-year) readmission risk.Emphasis on long-term outcomes; advanced ML pipeline.Relatively small/less diverse sample; limited generalizability.
Jiang et al. [26]Dynamic (varied)Not Reported (NR)Unsupervised ML (Clustering: k-means)Identified risk trajectories and clusters (e.g., “rapid decompensators”); segmentation associated with markedly different readmission risks.Novel dynamic prediction; insight into patient heterogeneity.Unsupervised results are harder to translate into protocols; lack of clinical actionability.
Golas et al. [25]30 days11,510 patients, 27,334 admissionsTraditional ML (Random Forest, Logistic Regression, SVM, Gradient Boosting)Random forest and logistic regression had similar AUCs (0.76), supporting use of EHR data for prediction.Large EHR dataset; real-time application design.Limited external validation; single institution setting.
Shameer et al. [31]30 daysLargeTraditional ML (Elastic Net Logistic Regression)AUC 0.72; demonstrated EHR-wide ML is feasible and valuable for readmission prediction.Comprehensive variable set; relevant to clinical workflows.Single-center design; focus on billing code predictors.
Bat-Erdene et al. [40]6, 12, 24 monthsModerateDeep LearningDeep learning outperformed traditional approaches for 6–24-month readmission prediction.Extended follow-up window; leveraged advanced neural networks.Lacked clinical interpretability; smaller dataset.
Chen et al. [38]1 yearNot Reported (NR)Deep Learning (Attention-based Neural Network)Attention mechanisms improved interpretability and prediction with AUC 0.82.Introduced model interpretability; highlighted features via attention weights.Lacked comparison to other ML approaches; cohort size NR.
Lv et al. [39]DynamicNot Reported (NR)Unsupervised ML (Clustering for trajectory patterns)High timing prediction (89% accuracy) through symptom trajectory clustering.Focus on dynamic, interpretable trajectories; novel approach.Hard to translate unsupervised findings into actionable clinical tools; sample size NR.
Sarijaloo et al. [35]90 daysModerateEnsemble (Random Forest, Gradient Boosting)ML models improved prediction of 90-day readmission and death versus clinical risk models.Included robust clinical and administrative data.Model complexity limits bedside application.
Table 4. Studies addressing specific limitations in HF readmission prediction.
Table 4. Studies addressing specific limitations in HF readmission prediction.
Study Key StrengthCritical Limitation Clinical Limitations
Huang et al. [43]Comprehensive scoping review of 42 studiesU.S.-centered sample (82% of included studies; no quality assessment of primary studiesLimited generalizability to non-Western healthcare systems
Angraal et al. [24]Long-term (3-year prediction capability)Homogeneous cohort (72% White participants); no SGLT2 inhibitor dataUnderestimates risk in Asian/younger populations
Shameer et al. [31]Health Electronic records (HER)-wide feature engineeringSingle-center design; reliance on billing codes over clinical narrativesMay miss social determinants affecting readmission
Allam et.al. [23]Comparison of neural networks vs. logistic regressionLimited to 30-day readmission predictionProvide insight on algorithm selection for short-term risk assessment
Frizzell et al. [7]Multicenter study designFocus on traditional statistical approachesEstablishes baseline for comparing ML to conventional methods
Jiang et al. [26]Novel unsupervised approach for dynamic risk trajectoriesComplex implementation in clinical settingsOffers new perspective on evolving readmission risk over time
Table 5. Comparative performance of ML models in HF readmission prediction.
Table 5. Comparative performance of ML models in HF readmission prediction.
StudyML Algorithm(s) UsedPrediction WindowAUCKey Features Used
Allam et al. [23]Neural Network, Logistic Regression30 days0.64Billing codes, labs
Frizzell et al. [7]Random Forest, Gradient Boosting, Logistic Regression30 days0.62–0.72EHR, demographics
Golas et al. [25]Random Forest, Logistic Regression, SVM, Gradient Boosting30 days0.76EHR, demographic, clinical, admission data
Chen et al. [38]Deep Learning (Attention-based Neural Network)1 year0.82EHR, text
Jiang et al. [26]Unsupervised k-Means ClusteringDynamic0.73 Trajectory patterns
Shameer et al. [31]Elastic Net Logistic Regression30 days0.72EHR-wide features, billing codes
Bat-Erdene et al. [40]Deep Learning6, 12, 24 months0.80–0.85Epidemiologic, labs, admission/discharge data
Sarijaloo et al. [35]Random Forest, Gradient Boosting90 days0.76Clinical, administrative, labs
Lv et al. [39]Unsupervised Clustering for Trajectory PatternsDynamicNot ReportedSymptom trajectories
Angraal et al. [24]Random Forest (Ensemble)3 years0.76Demographic, clinical variables
Polo Friz et al. [37]Supervised ML (Random Forest, SVM, Logistic Regression)30 days~0.69LACE index, administrative, clinical
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alnomasy, N.; Pangket, P.; Mostoles, R., Jr.; Alrashedi, H.; Pasay-an, E.; Cho, H.; Alsayed, S.; Gonzales, A.; Alharbi, A.A.M.; Alatawi, N.A.H.; et al. Predictive Performance of Machine Learning Models for Heart Failure Readmission: A Systematic Review. Biomedicines 2025, 13, 2111. https://doi.org/10.3390/biomedicines13092111

AMA Style

Alnomasy N, Pangket P, Mostoles R Jr., Alrashedi H, Pasay-an E, Cho H, Alsayed S, Gonzales A, Alharbi AAM, Alatawi NAH, et al. Predictive Performance of Machine Learning Models for Heart Failure Readmission: A Systematic Review. Biomedicines. 2025; 13(9):2111. https://doi.org/10.3390/biomedicines13092111

Chicago/Turabian Style

Alnomasy, Nader, Petelyne Pangket, Romeo Mostoles, Jr., Habib Alrashedi, Eddieson Pasay-an, Hwayoung Cho, Sharifah Alsayed, Analita Gonzales, Amal A. Mohammad Alharbi, Nuha Ayad H. Alatawi, and et al. 2025. "Predictive Performance of Machine Learning Models for Heart Failure Readmission: A Systematic Review" Biomedicines 13, no. 9: 2111. https://doi.org/10.3390/biomedicines13092111

APA Style

Alnomasy, N., Pangket, P., Mostoles, R., Jr., Alrashedi, H., Pasay-an, E., Cho, H., Alsayed, S., Gonzales, A., Alharbi, A. A. M., Alatawi, N. A. H., Torres, S., Abudawood, K., & Alamoudi, F. A. (2025). Predictive Performance of Machine Learning Models for Heart Failure Readmission: A Systematic Review. Biomedicines, 13(9), 2111. https://doi.org/10.3390/biomedicines13092111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop