Next Article in Journal
Hepatic Involvement across the Metabolic Syndrome Spectrum: Non-Invasive Assessment and Risk Prediction Using Machine Learning
Next Article in Special Issue
Neutrophil Extracellular Traps Upregulate p21 and Suppress Cell Cycle Progression to Impair Endothelial Regeneration after Inflammatory Lung Injury
Previous Article in Journal
Cleavage Patterns of 9600 Embryos: The Importance of Irregular Cleavage
Previous Article in Special Issue
Does Prehospital Suspicion of Sepsis Shorten Time to Administration of Antibiotics in the Emergency Department? A Retrospective Study in One University Hospital
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Machine Learning-Based Early Prediction of Sepsis Using Electronic Health Records: A Systematic Review

by
Khandaker Reajul Islam
1,
Johayra Prithula
2,
Jaya Kumar
1,*,
Toh Leong Tan
3,
Mamun Bin Ibne Reaz
4,
Md. Shaheenur Islam Sumon
5 and
Muhammad E. H. Chowdhury
6,*
1
Department of Physiology, Faculty of Medicine, University Kebangsaan Malaysia, Kuala Lumpur 56000, Malaysia
2
Department of Electrical and Electronics Engineering, University of Dhaka, Dhaka 1000, Bangladesh
3
Department of Emergency Medicine, Faculty of Medicine, Universiti Kebangsaan Malaysia, Kuala Lumpur 56000, Malaysia
4
Department of Electrical and Electronic Engineering, Independent University, Bangladesh Bashundhara, Dhaka 1229, Bangladesh
5
Department of Biomedical Engineering, Military Institute of Science and Technology (MIST), Dhaka 1216, Bangladesh
6
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
*
Authors to whom correspondence should be addressed.
J. Clin. Med. 2023, 12(17), 5658; https://doi.org/10.3390/jcm12175658
Submission received: 13 July 2023 / Revised: 13 August 2023 / Accepted: 23 August 2023 / Published: 30 August 2023
(This article belongs to the Special Issue Trends and Prospects in Sepsis and Septic Shock)

Abstract

:
Background: Sepsis, a life-threatening infection-induced inflammatory condition, has significant global health impacts. Timely detection is crucial for improving patient outcomes as sepsis can rapidly progress to severe forms. The application of machine learning (ML) and deep learning (DL) to predict sepsis using electronic health records (EHRs) has gained considerable attention for timely intervention. Methods: PubMed, IEEE Xplore, Google Scholar, and Scopus were searched for relevant studies. All studies that used ML/DL to detect or early-predict the onset of sepsis in the adult population using EHRs were considered. Data were extracted and analyzed from all studies that met the criteria and were also evaluated for their quality. Results: This systematic review examined 1942 articles, selecting 42 studies while adhering to strict criteria. The chosen studies were predominantly retrospective (n = 38) and spanned diverse geographic settings, with a focus on the United States. Different datasets, sepsis definitions, and prevalence rates were employed, necessitating data augmentation. Heterogeneous parameter utilization, diverse model distribution, and varying quality assessments were observed. Longitudinal data enabled early sepsis prediction, and quality criteria fulfillment varied, with inconsistent funding–article quality correlation. Conclusions: This systematic review underscores the significance of ML/DL methods for sepsis detection and early prediction through EHR data.

1. Introduction

Sepsis is a potentially fatal illness that occurs when the body’s response to an infection causes tissue and organ damage [1]. It is a complex syndrome characterized by a dysregulated host immune response, organ dysfunction, and a high risk of mortality [2]. Sepsis can affect people of all ages, from infants to the elderly, and can arise from various types of infections, including bacterial, viral, and fungal infections [3]. When an infection occurs, the body’s immune system releases chemicals to combat the invading pathogens. In sepsis, the immune response becomes dysregulated, leading to widespread inflammation throughout the body. This inflammatory response can damage tissues and impair organ function [4]. Sepsis can develop into severe sepsis or septic shock, which are life-threatening illnesses with high fatality rates if ignored [5]. Fever, a faster heartbeat, rapid breathing, changed mental status, decreased urine output, low blood pressure, and general malaise are some of the symptoms of sepsis that might vary [6]. An intensive care unit (ICU) must provide urgent medical care and treatment for sepsis [7]. The standard treatment involves administering both intravenous antibiotics to combat the underlying infection as well as intravenous fluids to maintain adequate blood pressure and organ perfusion [8]. In severe cases, additional interventions such as vasopressor medications to raise blood pressure, mechanical ventilation to support breathing, and renal replacement therapy may be necessary [9].
Sepsis imposes a substantial burden on healthcare systems worldwide. Globally, it stands as a leading cause of illness and mortality, incurring substantial treatment expenses and leaving survivors with lasting repercussions [10,11,12]. Sepsis significantly elevates mortality rates. Fleischmann et al. (2016) conducted a systematic review and meta-analysis and found that there are about 48.9 million cases of sepsis every year, with a 19.4% death rate [13]. The recorded death rate was about 15.7% in high-income countries, but it was much higher in low- and middle-income countries, where it stood at 34.7%. Angus et al. found in 2001 that the death rate for serious sepsis and septic shock is between 40 and 60% [10]. Long-term physical, cognitive, and psychological impairments are common among survivors of sepsis, resulting in a significant burden of morbidity. The Surviving Sepsis Campaign (SSC) conducted a study involving over 1000 sepsis survivors and discovered that one year after discharge, 33% of patients had cognitive dysfunction, 43% had new functional limitations, and 27% had symptoms of post-traumatic stress disorder (PTSD) [14]. These long-term complications can have significant impacts on survivors’ quality of life and result in substantial healthcare requirements. Sepsis places a significant burden on healthcare resources, increasing hospitalizations, ICU stays, and healthcare costs. According to a study by Rudd et al. (2013) [10], the total annual cost of sepsis hospitalizations in the United States was approximately $24 billion, or 13.3% of all hospital expenditures. Patients with sepsis require intensive monitoring, invasive procedures, and broad-spectrum antibiotics, all of which increase the utilization of healthcare resources. In addition, sepsis has been linked to an increased risk of readmissions and hospital-acquired infections, which further strains healthcare systems. Rhee et al. (2017) [15] found in a retrospective cohort study that sepsis survivors had a 38% higher risk of hospital readmission within 90 days compared to non-sepsis patients. Not only do these readmissions increase healthcare costs, but they also add to the overall burden on hospitals and healthcare facilities. For healthcare providers, sepsis presents significant challenges, including the need for prompt diagnosis, intervention, and the management of complications. The complexity and unpredictability of sepsis necessitate a multidisciplinary approach and place a significant burden on healthcare teams. The emotional toll of caring for critically ill septic patients, high mortality rates, and the risk of healthcare provider burnout are emerging concerns [12]. Providing healthcare providers with the necessary resources, support systems, and education is essential to addressing these challenges and improving patient outcomes.
Early identification and rapid treatment are, therefore, essential for enhancing patient outcomes. Prevention of sepsis involves measures such as proper hygiene practices, vaccination against infectious diseases, the prompt treatment of infections, and the appropriate use of antibiotics [16]. Additionally, the early identification of individuals at risk, such as those with compromised immune systems or chronic medical conditions, can help in implementing preventive strategies and prompt treatment [17].
In recent decades, data-driven biomarker discovery has garnered traction as an alternative to conventional methods with the potential to overcome existing obstacles. This strategy seeks to harvest and exploit health data using quantitative computer methods like machine learning. High-resolution digital data are becoming more available to persons at risk and patients with sepsis [18]. These include laboratory, vital, genetic, genomic, clinical, and health history data. Improved patient outcomes can be achieved with the early detection and prediction of sepsis [19]. In recent years, machine learning (ML) and deep learning (DL) algorithms have shown promise in sepsis diagnosis and early prediction [20] by analyzing large-scale patient data. These algorithms can be used with electronic health records (EHRs) and other clinical data to identify patterns and indicators of sepsis that may not be immediately apparent to human practitioners [21]. This enables the early detection of sepsis and the beginning of prompt therapies.
Machine learning techniques have greatly aided in sepsis detection and prediction. Logistic regression, support vector machines (SVMs), random forests, and gradient boosting are examples of common machine learning approaches. These algorithms make use of vital signs (such as heart rate and blood pressure), laboratory results (such as white blood cell count and lactate level), and clinical factors (such as age and comorbidities) to predict the likelihood of sepsis. High sensitivity, specificity, and area under the receiver operating characteristic curve (AUC-ROC) values have been reported for these models by researchers, demonstrating their propensity for sepsis prediction [18,22,23,24,25,26,27,28,29]. In response to the persisting challenge of sepsis-related fatalities, Wang et al. [30] sought to develop an AI algorithm for early sepsis prediction, successfully creating a random forest model utilizing 55 clinical features from ICU patient data, yielding an AUC of 0.91, 87% sensitivity, and 89% specificity, with potential wider applicability pending external validation. Similarly, Kijpaisalratana et al. (2022) [31] have developed novel machine learning-based sepsis screening tools and compared their performance with traditional methods for early risk prediction of sepsis. Using retrospective electronic health record data from emergency department visits, the machine learning models, including logistic regression, gradient boosting, random forest, and neural network models, exhibited significantly better predictive performance (e.g., AUROC 0.931) compared to reference models (AUROC 0.635 for qSOFA (quick Sepsis Related Organ Failure Assessment), 0.688 for MEWS (Modified Early Warning Score), and 0.814 for SIRS (Systemic Inflammatory Response Syndrome)), highlighting their potential to enhance sepsis diagnosis in emergency patients. The ability of DL, a subset of ML, to automatically develop hierarchical representations from complicated data has generated a great deal of interest in sepsis prediction. DL models, particularly convolutional neural networks (CNNs) and recurrent neural networks (RNNs) have shown encouraging results in several medical applications, including in sepsis detection. Nemati et al. (2018) combined CNN and long short-term memory (LSTM) networks to construct a deep learning model to predict sepsis using EHR data. The model beat traditional ML models with an AUC-ROC of 0.83 [32]. Addressing the pressing issue of sepsis’s exponential impact and increased mortality in ICU patients, Singh et al. [33] presents a highly accurate machine learning model for early detection, outperforming existing approaches with a proposed ensemble model achieving a balanced accuracy of 0.96.
This article’s goal is to give readers an overview of recent studies on ML- and DL-based sepsis diagnosis and early prediction. Due to the rapid pace of progress in this field of study, it is essential to review and evaluate the present state of the art in the field of sepsis detection and onset prediction. By analyzing the pertinent research, we assessed the effectiveness of various ML and DL models, the features employed for prediction, and the difficulties and future directions in this area. The purpose of this research was to provide an in-depth investigation of the most recent research by utilizing data taken from the electronic health records of adults and employing ML and DL models. To achieve this, we thoroughly searched for relevant articles in four popular electronic databases and developed a new assessment criterion for the quality assessment of different articles. The characteristics of different selected articles are thoroughly studied and summarized in this article. To the best of our knowledge, this article covers a systematic review of the largest number of research articles published between Jun 2016 and March 2023. In addition, we have discussed the current challenges and future direction of research in this domain, along with the limitations of the current study.

2. Background and Fundamental Concepts

The subsequent portion of this section provides concise explanations of key concepts that are essential for a deeper comprehension of the subjects addressed in the following sections. It is important to note that the selection of methods discussed in this section mirrors the methodologies utilized in the chosen articles as a direct outcome of the systematic review process.

2.1. Sepsis Definition

Sepsis criteria and definitions have changed over time to improve early care and identification. Sepsis-2 and Sepsis-3 are the two main definitions of sepsis that have gained widespread support.
The Sepsis-2 definition [34] is based on the existence of a systemic inflammatory response syndrome (SIRS) brought on by infection and was first published in 2001. According to Sepsis-2, sepsis is characterized by the presence of at least two SIRS criteria, which include leukopenia or leukocytosis in the case of abnormal white blood cell counts, as well as abnormal body temperature (fever or hypothermia), increased heart rate (tachycardia), increased respiratory rate (tachypnea), and abnormal body temperature (fever or hypothermia). Septic shock is defined as severe sepsis with prolonged hypotension despite fluid resuscitation, and severe sepsis combined with organ dysfunction is known as severe sepsis.
The Sepsis-3 definition [35] was put up in 2016 to increase the precision and clinical utility of sepsis detection. Instead of depending primarily on the SIRS criteria, it emphasizes organ failure as sepsis’ defining feature. According to Sepsis-3, sepsis is characterized as a potentially fatal organ malfunction brought on by an improperly managed host defense against infection. The Sequential Organ Failure Assessment (SOFA) score, which assesses the function of the respiratory, cardiovascular, hepatic, renal, coagulation, and central nervous systems, is used to measure organ dysfunction. If the SOFA score rises by two or more points as a result of infection, sepsis is thought to be present. A blood lactate level of more than two mmol/L and persistent hypotension requiring vasopressors to maintain a mean arterial pressure of 65 mmHg or higher are both considered signs of septic shock. These definitions play a crucial role in standardizing the identification and management of sepsis, aiding clinicians in making timely and accurate diagnoses and facilitating effective interventions to improve patient outcomes. The most widely adopted definition of sepsis is Sepsis-3 [36]. The key advantages of the Sepsis-3 definition include the following:
  • By incorporating organ dysfunction criteria, the Sepsis-3 definition improves the specificity of sepsis diagnosis. This helps differentiate sepsis from other conditions that may present with signs of infection but do not involve organ dysfunction.
  • The Sepsis-3 definition simplifies the criteria for sepsis by focusing on organ dysfunction rather than the systemic inflammatory response syndrome (SIRS) criteria used in previous definitions. This simplification reduces the potential for misdiagnosis and facilitates a more targeted approach to sepsis identification.
  • The early prediction of sepsis is crucial for timely intervention. The SOFA score, which is part of the Sepsis-3 definition, provides a tool for assessing organ dysfunction and predicting patient outcomes. Higher SOFA scores are associated with increased mortality rates and can help identify patients at higher risk who require immediate attention.
  • The Sepsis-3 definition has facilitated the standardization of sepsis diagnosis and research. The use of a consistent definition enables better comparison of studies, data sharing, and the development of evidence-based management strategies.

2.2. Common Machine Learning Models and Performance Metrics

A summary of common classical machine learning (ML) and deep learning models, as well as of performance metrics commonly used for the early prediction of sepsis, is listed below.
  • Classical ML Models:
    • Decision Trees (DT)
    • Random Forest (RF)
    • Support Vector Machine (SVM)
    • Logistic Regression (LR)
    • Gradient Boosting (GB)
    • Naïve Bayes (NB)
    • k-Nearest Neighbor (kNN)
  • Deep Learning Models:
    • Long Short-Term Memory (LSTM) Networks
    • Convolutional Neural Network (CNN)
    • Gated Recurrent Unit (GRU)
    • Neural Network (NN)
    • Multitask Gaussian Process and Attention-based Deep Learning Model (MGP-AttTCN)
    • Temporal Convolutional Network (TCN)
    • Recurrent Neural Network (RNN)
    • CNN-LSTM
    • CNN-GRU
  • Performance Metrics:
    • Area Under the Curve (AUC) or AUROC (Receiver Operating Characteristics Curve)
    • Sensitivity (Recall)
    • Specificity
    • Accuracy
    • Precision
    • F1 score
    • Matthews Correlation Coefficient (MCC)
    • Mean Average Precision (mAP)
    • Positive Predictive Value (PPV)
    • Negative Predictive Value (NPV)
    • Positive Likelihood Ratio (PLR)
    • Negative Likelihood Ratio (NLR)
These models and metrics have been widely used to assess the predictive performance of algorithms for early sepsis prediction using EHRs. The selection of models and metrics may vary depending on the specific study and dataset under consideration. Supplementary Tables S1 and S2 provide short definitions of these ML and DL models and performance metrics.

3. Materials and Methods

3.1. Search Strategy

A comprehensive search strategy was employed to identify relevant studies from electronic databases including PubMed, IEEE Xplore, Google Scholar, and Scopus (Figure 1). The bibliographic research for this systematic review was conducted by a team of two experienced researchers with expertise in the fields of medical informatics and machine learning. Additionally, the team consulted an information specialist with extensive knowledge in database searching and retrieval. The search terms used included variations of “sepsis”, “prediction”, “machine learning”, “deep learning”, and “electronic health records”. The search was limited to articles published in English between Jun 2016 and Mar 2023. This systematic review is not registered in PROSPERO or any other database.
We combined our search terms using Boolean operators (e.g., AND, OR) to create an effective search query. We used parentheses to group related terms: for example, “(sepsis OR prediction) AND (machine learning OR prediction) AND (electronic health records OR prediction)”. We executed our search query in each selected database. We applied the necessary filters, such as publication date or language, to refine our results, and saved the search strategy for reporting purposes, as shown in Supplementary Table S3.

3.2. Inclusion and Exclusion Criteria

The following inclusion and exclusion criteria were used for this study:

3.2.1. Inclusion Criteria

  • Studies that focused on the application of machine learning and deep learning algorithms for the early prediction of sepsis.
  • Studies utilizing electronic health records (EHR) data as the primary sources of information.
  • Studies that involved adult human subjects (i.e., age ≥ 18).
  • Studies that reported on the performance metrics (e.g., sensitivity, specificity, area under the curve) of the machine learning models for sepsis prediction.
  • Studies published in peer-reviewed journals.
  • Studies available in the English language.
  • Studies published within a specific time frame (e.g., Jun 2016 and March 2023).

3.2.2. Exclusion Criteria

  • Studies that did not focus on early prediction of sepsis.
  • Studies that focused only on using clinical notes.
  • Studies that did not involve the use of machine learning/deep learning algorithms.
  • Studies that did not utilize electronic health records as data sources.
  • Studies that primarily focused on non-human subjects or experimental setups not related to human healthcare.
  • Studies that did not report on performance metrics for machine learning models.
  • Studies that were not published in peer-reviewed journals.
  • Studies published in languages other than English.
It should be noted that Sepsis definition was not considered in the inclusion or exclusion criteria. The Population, Intervention, Comparator, Outcome, and Study Design (PICOS) criteria for the systematic review, including both the inclusion and exclusion criteria, are detailed in Supplementary Table S4.

3.3. Study Selection

After retrieving search results using the above-mentioned search strategy in the selected databases, we used manual screening to identify and remove duplicate articles from the search results. We reviewed the titles and abstracts of the remaining articles to determine their relevance to our research question and inclusion/exclusion criteria. We excluded clearly irrelevant articles at this stage. We excluded randomized controlled trials, study overviews and protocols, and meta-epidemiological studies. We also checked the systematic reviews and meta-analysis papers published in the time span and cross-checked them with what we had collected, and added any missing article as “More article selected from other sources”.
The study selection process was carried out by two independent authors, and discrepancies were resolved through discussion and consensus. In cases where disagreements arose, a third author was consulted to make a final decision. The authors assessed the relevance of each study based on predetermined inclusion and exclusion criteria. The initial screening involved reviewing the titles and abstracts of the identified articles, followed by a full-text assessment of potentially eligible studies. This process was conducted in duplicate to ensure a thorough and unbiased selection of studies for inclusion in the systematic review. The inter-rater agreement between the two authors was assessed using Cohen’s Kappa coefficient, and a score of above 0.85 indicated substantial agreement. We aimed to minimize bias and ensure the robustness of study selection through this rigorous and collaborative process.

3.4. Data Extraction

Data extraction was carried out independently by two authors, and any disparities were resolved through consensus. In cases where differences persisted, a third author was consulted to reach a final decision. The authors utilized a standardized data extraction form to systematically collect relevant information from the included studies. This process involved extracting key details such as study characteristics, participant demographics, intervention specifics, outcome measures, and relevant results. The data extraction was performed in duplicate to ensure accuracy and reliability. The level of agreement between the two authors was assessed using Cohen’s Kappa coefficient, with a value exceeding 0.85 indicating substantial concordance. We prioritized consistency and quality in data extraction through this collaborative and meticulous approach. To assess the quality of the articles and perform our systematic review, we extracted the following data from each article:
Publication characteristics: Collected the last name of the first author and the year of publication.
Study Design: Identified the study design used in the article (e.g., prospective cohort study, retrospective analysis).
Objectives: Determined the specific research objectives or aims stated by the authors, which were supposed to match our study criteria.
Cohort Selection: Extracted information about the characteristics of the study participants such as sample size, demographics, prevalence of sepsis, and any relevant inclusion/exclusion criteria.
Data Source: Identified the source of the data used in the study (e.g., electronic health records, administrative databases, clinical trials) and whether it is a publicly accessible dataset or not.
Model Selection: Extracted information about machine learning or deep learning methods employed for data analysis, including feature engineering for classical ML, hyperparameters, ways to handle overfitting, etc.
Reproducibility: Assessed whether the article provides sufficient information to reproduce the study, including details on data availability, code availability, and software/hardware specifications.
Performance Measures and Explainability: Determined whether the performance matrices are reported in the study or not and whether the reported model is explainable or not.
Limitations: Noted the limitations or potential biases acknowledged by the authors in the article, such as sample size limitations, selection bias, or confounding factors.
Funding Source: Identified any sources of funding or financial support disclosed by the authors.

3.5. Quality Assessment

The risk of bias in the included studies was assessed using a predefined comprehensive approach to assessing the quality of eligible articles. The criteria include unmet needs, reproducibility, robustness, generalizability, and clinical significance. Based on Moor et al. [18] and Qiao et al. [37], modified quality assessment criteria are presented, which include sample size (>50), data availability, code availability, mobile/web deployment, handling of missing data, sepsis prevalence, feature engineering, machine learning model, hyperparameters, overfitting prevention technique, reporting of performance metrics, validation using external data, explainability, limitation of the study in question, and discussion of clinical application. The inclusion of 16 relevant criteria and the use of a quality assessment table with “yes” or “no” ratings for each category contribute to a systematic and transparent evaluation of study quality.
The quality assessment of the included studies was conducted by two authors independently, and any discrepancies were resolved through discussion and consensus. If a consensus could not be reached, a third author was involved to make a final decision. To evaluate inter-rater agreement, Cohen’s Kappa coefficient was computed, yielding a substantial agreement level of 0.85 between the two authors. This demonstrated a high level of concordance in their assessments. We placed significant emphasis on evaluating the methodological quality of the included studies, aiming to ensure the reliability and validity of the overall findings.

3.6. Impact of Funding Source

The funding source should not influence the design, conduct, analysis, interpretation, or reporting of a systematic review. It is crucial to maintain the independence and objectivity of the review process to ensure the integrity of the findings. Depending on the funding source, there may be implications for the generalizability of the findings. If the funding is specific to a particular setting, population, or intervention, it is important to consider the applicability of the results to other contexts.

4. Results

4.1. Selection Process

The initial search yielded a total of 1942 articles (Figure 1). Out of these initially selected articles, only 42 articles met the complete inclusion criteria. Most studies (n = 1900) were disregarded because they did not meet the inclusion criteria. These criteria included: research not involving machine learning or deep learning; conducting research on the wrong population (e.g., pediatric, or neonatal); conducting research on a topic outside the scope of the current review (e.g., mortality prediction); and conducting research using a study design different from that used in the current article, i.e., not peer reviewed.

4.2. Study Characteristics

Among the 42 studies selected for this systematic review, 38 were retrospective studies, while 4 were prospective studies. Most of the studies (35) use populations from the United States, while two studies are from China, two are from South Korea, and one each is from Singapore, Israel, and Denmark. The majority of the studies (34) used ICU data, while 11 studies used ED data, and the remaining six studies used general ward data (Figure 2). Some studies used data from multiple sources.
Figure 3 shows that the most commonly used data source is MIMIC-III (n = 14, 29.8%), while the second most used dataset is the 2019 PhysioNet/CinC Challenge (n = 6, 12.8%), and the Emory Healthcare System (n = 4, 8.5%) is the third most used dataset. Each of the University of Pennsylvania Health System, University of California, San Francisco (UCSF), Methodist Le Bonheur Healthcare (MLH) System, and MIMIC-II datasets were used for two studies. It is worth noting here that some studies used multiple datasets. The Sepsis-2 (33.3%, n = 14) or Sepsis-3 (52.4%, n = 22) definitions of sepsis were utilized in the majority of the investigations. Depending on the nature of these studies, the time windows that were used have changed, and numerous researchers revised the Sepsis-2 or -3 definition. ICD-9 was utilized in certain studies (n = 4, 9.5%), while ICD-10 was used in others (n = 2, 4.8%). Intensive care unit specialists diagnosed sepsis in some of these investigations.
The proportion of patients with sepsis varied from 0.41% to 63.6%; however, most of the studies used imbalanced datasets, with very small numbers of patients with sepsis compared to the study populations. Figure 4 shows that in only 12 studies out of 42 studies, the prevalence of sepsis was more than 20%, while the median of sepsis prevalence was 9.5%. For three studies (18th, 22nd, and 31st), the sepsis prevalence data were not provided. Therefore, most of the ML or DL studies need to adopt some form of data augmentation or balancing techniques to develop reliable machine learning models. This systematic review only focused on the studies carried out on adult patients. Supplementary Figure S1 shows the sample size and sepsis-positive population in numbers and percentages, which made it possible to have a clear visual depiction of the imbalance distribution of the dataset. Two studies were discarded from this plot as those studies would make the plot biased, and a clear picture of the sample size and sepsis prevalence becomes unclear if those studies are incorporated.
Most of the studies were carried out using vital signs (73.1%), laboratory data (65.4%), and demographics (55.8%), while some studies also used clinical notes, treatments or medications, comorbidities, imaging, clinical context, diagnosis, etc., as shown in Figure 5.
A wide range of parameters is used in different studies, from as low as two to as high as one hundred and sixty eight. However, the median number of parameters used in the studies is 22. Figure 6 clearly shows that eight studies used more than fifty parameters, while most of the studies used a smaller number of parameters. Sixteen studies reported the feature importance among the variables available in the respective datasets.
After the careful evaluation of important features identified by the studies, the Top 20 features were shown in Figure 7 based on their appearance in the important feature lists of 16 articles. It is noticed that heart rate, temperature, white blood cell (WBC) count, SBP, age, DBP, and RR are the six most ranked features reported in these studies.

4.3. Machine Learning Models

The selected studies exhibited a diverse array of machine learning and deep learning models, encompassing a total of 29 different approaches. Notably, within this spectrum, 58% of the articles opted for classical machine learning models, while the remaining 42% delved into the application of various deep learning models. This distribution underscores the comprehensive exploration of both traditional and cutting-edge techniques to address the complexities of early sepsis prediction.

4.4. Cross-Validation (CV) Techniques

Two cross-validation strategies were used in the articles: train-val-test split and N-fold cross-validation. Among the 42 selected articles, 14 articles used train-val-test split techniques, while 23 articles reported N-fold cross-validation techniques. Four different types of N-fold CV techniques are used: 4-fold CV (n = 4), 5-fold CV (n = 6), 6-fold CV (n = 1), and 10-fold CV (n = 12). It is clear that a 10-fold CV is popular among researchers for CV, while a large number of articles used a train-val-test split, which is equivalent to a single-fold CV.

4.5. Performance Metrics

Although different articles used different performance metrics, some performance metrics were common in the majority of articles, such as area under the curve (AUC) or area under the receiver operating characteristics curve (AUROC) (29.8%), sensitivity or recall (R) (19.8%), specificity (S) (21.5%), and accuracy (A) (9.9%). One metric was specific to the dataset, such as in the example of the Utility score being suitable for the 2019 PhysioNet/CinC Challenge. Other used metrics were precision, F1 score, F2 score, MCC, mAP, PPV, NPV, PLR, NLR, and response rate. AUROC values range between 0.80 and 0.97 in different studies. However, comparing model performance based solely on specific metrics can be limiting due to dataset variations. Given diverse datasets and experimental setups, assessing article performance solely based on such metrics is inappropriate. The models demonstrated improved accuracy and performance compared to traditional methods in identifying patients at risk of sepsis. Several studies reported high sensitivity and specificity, highlighting the potential for early detection and timely intervention. We have reported them in Table 1; however, we should not judge these articles based on these metrics. Moreover, it is beyond the scope of this systematic review to evaluate the articles using these performance metrics; rather, we have reported a quality evaluation in Section 3.5 to evaluate the quality of the articles based on 16 performance measures.

4.6. Early Prediction of Sepsis Onset

Among the 42 articles, 30 (71.4%) articles used longitudinal data for the early prediction of sepsis onset, while the remaining 12 (28.6%) articles did not report early prediction (Figure 8). Only eight articles reported sepsis onset prediction 12 or more hours earlier. Five articles reported 24 h earlier, while one article reported 40 h earlier. The majority of the articles reported sepsis onset 2–6 h earlier.

4.7. Quality Assessment of Included Studies

Table 2 displays the outcomes of the quality analysis. The quality evaluation was meant to evaluate how well certain prediction models were implemented and reported. Some studies reported a challenge or did not properly report their ML/DL model. Reyna et al. [54] presented the findings of the 2019 PhysioNet/Computing in Cardiology Challenge, which made such an article difficult to assess. The remaining studies had quality ratings ranging from extremely low (meeting 30% of assessment criteria) to very good (more than 80% of assessment criteria). There was no study that met all 16 criteria. Three requirements were met by every single study; all of them had used a sample size of more than 50, used a sepsis definition that adhered to ours in the study design, and reported performance metric(s). One requirement was fulfilled by 95% of the studies, which was reporting sepsis prevalence, while the reporting of the machine learning model in use had been provided in 93% of studies. Out of all the studies, 69% had reported the techniques adopted for handling missing data, 83% of the studies had reported feature engineering techniques used, and 74% of the studies had reported the limitations of their contents. Only a small percentage of research studies (n = 2 (5%)) had deployed their model for prospective study or real-world validation. Only six studies had made their data-cleaning, analysis, and ML/DL code available. Interestingly, 52% of studies had been carried out on publicly available datasets, while only 17% of studies had validated the machine learning models on external datasets, and 21% of studies had reported explainable AI models. Out of all the studies, 46% had reported hyperparameters, 43% had reported techniques to avoid overfitting, and 50% had discussed the clinical applicability of their proposed approaches. We have categorized the articles according to our newly adopted tool (Table 2) into four categories: low quality (LQ) (0–40%), average quality (AQ) (40–60%), above-average quality (AAQ) (60–80%), and high quality (HQ) (80–100%). According to our evaluation criteria, there are 5 HQ, 19 AAQ, 13 AQ, and 5 LQ articles among the 42 articles investigated in this study.

4.8. Impact of Funding Source

In our evaluation process, we meticulously assessed the quality of the articles by considering various factors including the availability of funding information. However, our comprehensive analysis revealed that there was no discernible correlation between the presence of funding information and the overall quality of the articles. This observation underscores the independence of funding disclosure from the overall rigor and validity of the research findings, emphasizing the importance of a comprehensive evaluation encompassing various dimensions of article quality, which is reported in Table 2.

5. Discussion

This section sheds light on the current state of research, revealing several important insights. However, upon comparison with the existing literature, we acknowledge the existence of other comprehensive systematic reviews and meta-analyses that delve into intricate technical and clinical aspects of sepsis prediction algorithms, thereby offering clinicians an understanding of the challenges and opportunities within this domain. This section also discusses the limitations of our study and challenges and future directions of research in this domain.

5.1. Key Findings

The systematic review conducted a comprehensive search resulting in 1942 articles, from which 42 studies were ultimately selected based on stringent inclusion criteria. These criteria ensured the relevance of the chosen studies, with non-machine learning or -deep learning studies and those with pediatric populations, divergent topics, and non-peer-reviewed designs excluded. The predominance of retrospective designs (n = 38) among the selected studies highlighted the utilization of historical data for analysis, with geographical diversity observed, including a concentration of investigations in the United States (n = 35). The variety of healthcare settings utilized, including ICU (n = 34), ED (n = 11), and general ward data, underscored the complexity of sepsis detection across different contexts. Notably, various datasets were incorporated, with MIMIC-III, PhysioNet/CinC Challenge, and the Emory Healthcare System being primary sources. The adoption of Sepsis-2 and Sepsis-3 definitions for diagnosis indicated the evolving nature of sepsis classification. The range of sepsis prevalence (0.41% to 63.6%) exposed the inherent dataset imbalance, leading many studies to employ data augmentation techniques. The range of parameter utilization and feature importance analysis highlighted the heterogeneity of approaches within the field, emphasizing the ongoing quest for the optimal model.
The analyzed studies reveal a balanced distribution between classical machine learning models (58%) and various deep learning models (42%), indicating a diverse spectrum of approaches. Validation strategies were categorized into train-validation-test split and N-fold cross-validation, with the latter being prominent, especially 10-fold CV. Common metrics like AUROC, Sensitivity, Specificity, and Accuracy were recurrently used. Large disparities are found among sepsis definitions, making it impossible to compare the AUROC value of every study to find the best machine learning model. AUROC values range between 0.80 and 0.97 in different studies. However, comparing model performance based solely on specific metrics can be limiting due to dataset variations. The analysis demonstrates that 71.4% of studies utilized longitudinal data for early sepsis onset prediction, often forecasting sepsis 2–6 h ahead. Quality assessments ranged from extremely low to very good, showcasing the multifaceted nature of article quality. Notably, no study met all 16 quality criteria, highlighting the complexity of evaluation. Funding information showed no consistent correlation with article quality.
A substantial portion of EHR data resides within unstructured clinical notes, encompassing clinician insights not captured by physiological variables. Several studies have explored leveraging natural language processing (NLP) techniques to extract predictive features from clinical notes, targeting sepsis detection [49,52,64,72]. However, most of these investigations have treated NLP features in isolation without integrating them with physiological data. The study conducted by Goh et al. [49] showcases the synergy of NLP and physiological features when early predicting sepsis. This study highlights that combining NLP and physiological attributes yields superior classification performance compared to utilizing NLP or physiological features in isolation.
To ensure a comprehensive evaluation of the quality of the articles, several key aspects should be considered. Firstly, the sample size in a study should be a reasonable size to enhance the statistical robustness of the findings. Additionally, data accessibility and the availability of code should be prioritized to ensure the transparency and replicability of the study. Secondly, for practicality and broader usability, exploring the potential of mobile or web deployment for the developed models can enhance their real-world applicability. Thirdly, the handling of missing data should be addressed using effective strategies to maintain data integrity. Moreover, evaluating the developed models against the prevalence of sepsis is crucial, and meticulous feature engineering should be emphasized to extract relevant insights. Ensuring the adoption of suitable machine learning models and careful tuning of hyperparameters is essential to prevent the risk of overfitting. Furthermore, the transparent reporting of performance metrics enables a clear assessment of model efficacy. External validation is imperative to establish the generalizability of the models, and their explainability should be a priority. Lastly, the discussion should encompass both the clinical implications of the findings and a candid exploration of the study’s limitations.
The quality assessment conducted in this study, as reported in Table 2, reveals that among the studies included, five were designated as high quality, five as low quality, nineteen as above-average quality, and thirteen as average quality.

5.2. Summaries of Recent Systematic Reviews in Relevant Fields

Eight systematic reviews and meta-analyses have been identified in the literature (Supplementary Table S5), each focusing on distinct facets of sepsis prediction, precluding direct comparison with our study. Our investigation encompassed a substantial volume of peer-reviewed journal articles, contrasting with some literature sources that also included conference papers. Some articles lacked quality assessment reporting. Nonetheless, our findings align with the majority of these articles except for meta-analysis papers that explored a narrower selection of articles but conducted in-depth investigations.
Jahandideh et al. (2023) [25] explored the use of ML techniques for predicting patient clinical deterioration in hospitals. A total of 29 primary studies were identified, utilizing various ML models, including supervised, unsupervised, and classical techniques. The models exhibited diverse performance, with area-under-the-curve values ranging from 0.55 to 0.99, highlighting the potential for automated patient deterioration identification, although further real-world investigations are needed. Deng et al. (2022) [26] introduced new evaluation criteria and reporting standards for assessing 21 machine learning models based on PRISMA, revealing inconsistent sepsis definitions, varied data sources, preprocessing methods, and models, with AUROC improvement being linked to machine learning’s role in feature engineering. Deep neural networks coupled with Sepsis-3 criteria show promise for time-series data from sepsis patients, aiding clinical model enhancements. Yan et al. (2022) [28] assessed the impact of using unstructured clinical text in machine learning for sepsis prediction. Various databases were searched for articles using clinical text for ML or natural language processing (NLP) to predict sepsis. Findings indicated that combining text with structured data improved sepsis prediction accuracy compared to structured data alone, with varying methods and definitions influencing outcomes. However, the lack of comparable measurements prevented meta-analysis. Giacobbe et al. (2021) [27] focused on the impact of sepsis definition, input features, model performance, and AI’s role in healthcare, with potential benefits in medical decision-making. Sepsis prediction studies in the ICU often rely on MIMIC-II or MIMIC-III data, but insufficient code-sharing hampers reproducibility [18]. Bias in datasets is observed with predominantly Western cohorts, impacting sepsis label creation due to demographic and policy differences. Inconsistent study parameters and metrics prevent meta-analyses, highlighting the need for improved methods and shared code to enhance predictive accuracy and research comparability. Fleuren et al. (2020) [29] demonstrated ML model accuracy in predicting sepsis onset in retrospective cohorts with clinically relevant variables. While individual models outperform traditional tools, study heterogeneity limits pooled performance assessment. Clinical implementation across diverse patient populations is crucial to assess real-world impact. Islam et al. (2019) [23] carried out a meta-analysis investigating ML model performance in predicting sepsis 3–4 h before onset. The ML model in the study outperformed traditional sepsis scoring tools like SIRS, MEWS, SOFA, and qSOFA in recognizing sepsis and non-sepsis cases, with higher ability for sepsis detection. Different datasets showed consistent performance, suggesting machine learning’s potential to reduce sepsis-related mortality and hospital stay by accurately identifying at-risk patients, despite the challenge of sepsis diagnosis due to organ dysfunction and preexisting conditions. Schinkel et al. (2019) [22] identified fifteen articles on sepsis diagnosis with AI models (best AUROC 0.97), seven on mortality prognosis (AUROC up to 0.895), and three on treatment assistance. However, 22 articles exhibited high risk of bias due to the overestimation of performance caused by predictor variables coinciding with sepsis definitions. The authors also reported that AI holds potential for early antibiotic administration, but bias, overfitting, and lack of standardization hinder clinical implementation.

5.3. Limitations

This systematic review is subject to publication bias as we have selected studies with significant or positive results. There is a lack of studies reporting negative or null results; therefore, the systematic review’s findings may overstate the effectiveness of machine learning-based early prediction models for sepsis. The studies included in this systematic review vary in terms of patient populations, healthcare settings, study designs, machine learning algorithms used, and outcome measures. This heterogeneity limits the comparability and generalizability of the results, making it challenging to draw definitive conclusions on the performance of different machine learning models used on diverse datasets. Additionally, issues related to data quality, feature availability, and the interpretability of models were identified as challenges in the field. To address these limitations, future studies should focus on the prospective validation of machine learning models, external validation across diverse healthcare systems, and comparative analyses of different algorithms. Efforts to improve data quality, feature engineering, and the interpretability of models are crucial. Furthermore, studies should evaluate the clinical impact of implementing machine learning-based sepsis prediction models, considering patient outcomes, healthcare resource utilization, and cost-effectiveness.
Handling missing data in EHRs, along with noting whether a study is retrospective or prospective, is pivotal. Imputing substantial missing values, even with robust statistical methods, as seen in MIMIC-based studies with up to 50% imputed data, can pose limitations and uncertainties for sepsis prediction—a critical concern for clinicians within machine learning approaches. Assessing the quality and risk of bias in included studies is an essential aspect of a systematic review. However, in this study, we have introduced a 16-parameter quality assessment tool based on previous studies [18,37] to assess the quality of the included articles. Although there are subjective metrics in the scoring system, due to the addition of a large number of metrics, this assessment tool made this evaluation more reproducible compared to earlier systematic reviews.
This systematic review relies on published studies only, which has led to the exclusion of relevant studies available only as preprints. This exclusion could impact the comprehensiveness of the review and potentially overlook recent advancements or findings in the field. Machine learning-based early prediction of sepsis is a rapidly evolving field, and this systematic review has used a cut-off date (March 2023) for the inclusion of studies. Consequently, the review may not capture the most recent advancements or developments in the field, limiting the review’s currency. This systematic review has language restrictions, such as including only studies published in English, which can introduce language bias. Additionally, systematic reviews may only include studies published in indexed journals, leading to potential publication bias by excluding relevant studies from non-indexed sources and conferences.

5.4. Challenges and Future Directions

Considering our findings in the context of the broader literature, we recognize several challenges that demand further exploration and investigation. The application of machine learning algorithms in the early prediction of sepsis using EHRs holds great potential for improving patient outcomes [32,70]. By leveraging large amounts of patient data, these models can identify subtle patterns and early indicators of sepsis that may go unnoticed by human clinicians. The early prediction of sepsis enables timely interventions, such as appropriate antibiotic therapy and fluid resuscitation, which can significantly reduce morbidity and mortality rates. However, several challenges need to be addressed for the successful implementation of machine learning-based sepsis prediction models in clinical practice. Here are some key challenges and potential future directions in the field [32,78,79,80]:
  • EHR data can be heterogeneous, incomplete, and prone to errors, which poses challenges for accurate prediction models. Future research should focus on improving data quality and standardization, integrating data from multiple sources, and developing techniques to handle missing data effectively.
  • EHR data contain a vast number of variables, and not all of them may be relevant for sepsis prediction. Feature selection techniques and advanced representation learning methods, such as deep learning, can help identify the most informative features and extract meaningful representations from the EHR data.
  • Sepsis is a relatively rare event compared to non-sepsis cases, leading to imbalanced datasets. Class imbalance can affect model performance, and handling this issue requires techniques such as oversampling, under-sampling, or employing advanced algorithms designed for imbalanced data.
  • Machine learning models trained on one healthcare system may not be generalized well to other institutions or patient populations. Future research should focus on the external validation and generalizability of sepsis prediction models across diverse healthcare settings to ensure their real-world effectiveness.
  • Black-box machine learning models may lack interpretability, which can limit their adoption in clinical practice. Developing interpretable models and providing explanations for model predictions can enhance trust and facilitate clinicians’ understanding of the underlying reasons for sepsis predictions.
  • Early detection and timely intervention are crucial for sepsis management. Future research should focus on developing real-time prediction models that integrate seamlessly into clinical workflows, triggering alerts to clinicians and facilitating prompt action.
  • Demonstrating the clinical impact of machine learning-based sepsis prediction models is essential. Prospective validation studies in clinical settings are needed to assess these models’ effectiveness, impact on patient outcomes, and cost-effectiveness compared to existing clinical practices.
  • EHR datasets defining sepsis onset time becomes crucial for predictive models’ clinical relevance. The challenge lies in aligning model predictions with actual clinical timelines, considering symptoms’ varying occurrence times. Symptoms manifesting hours before hospital arrival or in different healthcare settings pose complexities in early prediction models’ optimization, which necessitates detailed exploration and discussion regarding patient record alignment and optimization.

6. Conclusions

Sepsis exerts significant mortality, morbidity, and healthcare strains, necessitating strategies such as heightened sepsis awareness, early diagnosis, standardized care protocols, and post-sepsis monitoring to alleviate its impact. This systematic review aims to comprehensively synthesize the existing evidence, encompassing diverse classical machine learning (ML) and deep learning (DL) prediction models, performance metrics, key features, and limitations. Notably, around half of the evaluated articles demonstrate above-average quality. The systematic evaluation underscores the potential of ML models (AUROC: 0.80 to 0.97) in predicting sepsis onset using electronic health records (EHRs), often forecasting sepsis emergence 2–6 h beforehand. This research emphasizes the ability of ML-based early sepsis prediction to enhance patient care despite existing challenges. The progressive exploration of this domain promises the development of robust models for clinical integration, ultimately facilitating timely interventions and improved patient outcomes in sepsis management.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm12175658/s1, Table S1. Different classical ML and DL models commonly used in the early prediction of sepsis. Table S2. Performance metrics used to interpret results. Table S3. A literature retrieval strategy for sepsis prediction. Table S4. PICOS criteria for the systematic review. Table S5. Summary of systematic review and meta-analysis in the literature. Figure S1. Visual representation of sample size and sepsis-positive patient numbers and percentage. Note: Studies # 1 and 28 were removed while plotting this figure as these datasets are outliers, and including these two studies would make the plot non-representative to most of the studies.

Author Contributions

Conceptualization K.R.I., J.K. and M.E.H.C.; Data curation K.R.I., J.P. and M.S.I.S.; Formal analysis K.R.I., J.P. and M.S.I.S.; Project Management J.K., T.L.T., M.B.I.R. and M.E.H.C.; Investigation K.R.I., J.P., J.K. and M.E.H.C.; Methodology M.E.H.C., J.K., T.L.T. and M.B.I.R.; Software K.R.I., J.P. and M.S.I.S.; Project administration J.K. and M.E.H.C.; Resources M.E.H.C. and J.K.; Supervision J.K., M.E.H.C., T.L.T. and M.B.I.R.; Validation J.K., T.L.T., M.B.I.R. and M.E.H.C. All authors equally contributed to the reviewing and editing of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors can provide the dataset upon reasonable request.

Acknowledgments

This study was supported by Faculty of Medicine, Universiti Kebangsaan Malaysia and Qatar National Library (QNL).

Conflicts of Interest

There is no conflict of interest to declare.

References

  1. Vincent, J.L.; Opal, S.M.; Marshall, J.C.; Tracey, K.J. Sepsis definitions: Time for change. Lancet 2013, 381, 774–775. [Google Scholar] [CrossRef] [PubMed]
  2. Caraballo, C.; Jaimes, F. Focus: Death: Organ dysfunction in sepsis: An ominous trajectory from infection to death. Yale J. Biol. Med. 2019, 92, 629. [Google Scholar] [PubMed]
  3. Jain, A.; Jain, S.; Rawat, S. Emerging fungal infections among children: A review on its clinical manifestations, diagnosis, and prevention. J. Pharm. Bioallied Sci. 2010, 2, 314. [Google Scholar] [CrossRef]
  4. Arina, P.; Singer, M. Pathophysiology of sepsis. Curr. Opin. Anesthesiol. 2021, 34, 77–84. [Google Scholar] [CrossRef] [PubMed]
  5. Liang, S.Y.; Kumar, A. Empiric antimicrobial therapy in severe sepsis and septic shock: Optimizing pathogen clearance. Curr. Infect. Dis. Rep. 2015, 17, 493. [Google Scholar] [CrossRef] [PubMed]
  6. Dorsett, M.; Kroll, M.; Smith, C.S.; Asaro, P.; Liang, S.Y.; Moy, H.P. qSOFA has poor sensitivity for prehospital identification of severe sepsis and septic shock. Prehospital Emerg. Care 2017, 21, 489–497. [Google Scholar] [CrossRef]
  7. Levy, M.M.; Artigas, A.; Phillips, G.S.; Rhodes, A.; Beale, R.; Osborn, T.; Vincent, J.L.; Townsend, S.; Lemeshow, S.; Dellinger, R.P. Outcomes of the Surviving Sepsis Campaign in intensive care units in the USA and Europe: A prospective cohort study. Lancet Infect. Dis. 2012, 12, 919–924. [Google Scholar] [CrossRef]
  8. Hunt, A. Sepsis: An overview of the signs, symptoms, diagnosis, treatment and pathophysiology. Emerg. Nurse 2019, 27, 32–41. [Google Scholar] [CrossRef]
  9. Dellinger, R.P.; Levy, M.M.; Rhodes, A.; Annane, D.; Gerlach, H.; Opal, S.M.; Sevransky, J.E.; Sprung, C.L.; Douglas, I.S.; Jaeschke, R.; et al. Surviving Sepsis Campaign: International guidelines for management of severe sepsis and septic shock, 2012. Intensive Care Med. 2013, 39, 165–228. [Google Scholar] [CrossRef]
  10. Rudd, K.E.; Johnson, S.C.; Agesa, K.M.; Shackelford, K.A.; Tsoi, D.; Kievlan, D.R.; Colombara, D.V.; Ikuta, K.S.; Kissoon, N.; Finfer, S.; et al. Global, regional, and national sepsis incidence and mortality, 1990–2017: Analysis for the Global Burden of Disease Study. Lancet 2020, 395, 200–211. [Google Scholar] [CrossRef]
  11. Angus, D.C.; Linde-Zwirble, W.T.; Lidicker, J.; Clermont, G.; Carcillo, J.; Pinsky, M.R. Epidemiology of severe sepsis in the United States: Analysis of incidence, outcome, and associated costs of care. Crit. Care Med. 2001, 29, 1303–1310. [Google Scholar] [CrossRef] [PubMed]
  12. Evans, L.; Rhodes, A.; Alhazzani, W.; Antonelli, M.; Coopersmith, C.M.; French, C.; Machado, F.R.; Mcintyre, L.; Ostermann, M.; Prescott, H.C.; et al. Surviving sepsis campaign: International guidelines for management of sepsis and septic shock 2021. Intensive Care Med. 2021, 47, 1181–1247. [Google Scholar] [CrossRef] [PubMed]
  13. Fleischmann, C.; Scherag, A.; Adhikari, N.K.; Hartog, C.S.; Tsaganos, T.; Schlattmann, P.; Angus, D.C.; Reinhart, K. Assessment of global incidence and mortality of hospital-treated sepsis. Current Estimates and Limitations. Am. J. Respir. Crit. Care Med. 2016, 193, 259–272. [Google Scholar] [CrossRef] [PubMed]
  14. Iwashyna, T.J.; Ely, E.W.; Smith, D.M.; Langa, K.M. Long-term cognitive impairment and functional disability among survivors of severe sepsis. JAMA 2010, 304, 1787–1794. [Google Scholar] [CrossRef]
  15. Rhee, C.; Murphy, M.V.; Li, L.; Platt, R.; Klompas, M. Comparison of trends in sepsis incidence and coding using administrative claims versus objective clinical data. Clin. Infect. Dis. 2015, 60, 88–95. [Google Scholar] [CrossRef]
  16. Shakoor, S.; Warraich, H.J.; Zaidi, A.K. Infection prevention and control in the tropics. In Hunter’s Tropical Medicine and Emerging Infectious Diseases; Elsevier: Amsterdam, The Netherlands, 2020; pp. 159–165. [Google Scholar]
  17. Luu, S.; Spelman, D.; Woolley, I.J. Post-splenectomy sepsis: Preventative strategies, challenges, and solutions. Infect. Drug Resist. 2019, 12, 2839–2851. [Google Scholar] [CrossRef]
  18. Moor, M.; Rieck, B.; Horn, M.; Jutzeler, C.R.; Borgwardt, K. Early prediction of sepsis in the ICU using machine learning: A systematic review. Front. Med. 2021, 8, 607952. [Google Scholar] [CrossRef]
  19. Lauritsen, S.M.; Kalør, M.E.; Kongsgaard, E.L.; Lauritsen, K.M.; Jørgensen, M.J.; Lange, J.; Thiesson, B. Early detection of sepsis utilizing deep learning on electronic health record event sequences. Artif. Intell. Med. 2020, 104, 101820. [Google Scholar] [CrossRef]
  20. Ramlakhan, S.; Saatchi, R.; Sabir, L.; Singh, Y.; Hughes, R.; Shobayo, O.; Ventour, D. Understanding and interpreting artificial intelligence, machine learning and deep learning in Emergency Medicine. Emerg. Med. J. 2022, 39, 380–385. [Google Scholar] [CrossRef]
  21. Coggins, S.A.; Glaser, K. Updates in Late-Onset Sepsis: Risk Assessment, Therapy, and Outcomes. Neoreviews 2022, 23, 738–755. [Google Scholar] [CrossRef]
  22. Schinkel, M.; Paranjape, K.; Panday, R.N.; Skyttberg, N.; Nanayakkara, P.W. Clinical applications of artificial intelligence in sepsis: A narrative review. Comput. Biol. Med. 2019, 115, 103488. [Google Scholar] [CrossRef] [PubMed]
  23. Islam, M.M.; Nasrin, T.; Walther, B.A.; Wu, C.C.; Yang, H.C.; Li, Y.C. Prediction of sepsis patients using machine learning approach: A meta-analysis. Comput. Methods Programs Biomed. 2019, 170, 1–9. [Google Scholar] [CrossRef] [PubMed]
  24. Komorowski, M.; Green, A.; Tatham, K.C.; Seymour, C.; Antcliffe, D. Sepsis biomarkers and diagnostic tools with a focus on machine learning. EBioMedicine 2022, 86, 104394. [Google Scholar] [CrossRef] [PubMed]
  25. Jahandideh, S.; Ozavci, G.; Sahle, B.; Kouzani, A.; Magrabi, F.; Bucknall, T. Evaluation of machine learning-based models for prediction of clinical deterioration: A systematic literature review. Int. J. Med. Inform. 2023, 175, 105084. [Google Scholar] [CrossRef] [PubMed]
  26. Deng, H.F.; Sun, M.W.; Wang, Y.; Zeng, J.; Yuan, T.; Li, T.; Li, D.H.; Chen, W.; Zhou, P.; Wang, Q.; et al. Evaluating machine learning models for sepsis prediction: A systematic review of methodologies. Iscience 2022, 25, 103651. [Google Scholar] [CrossRef] [PubMed]
  27. Giacobbe, D.R.; Signori, A.; Del Puente, F.; Mora, S.; Carmisciano, L.; Briano, F.; Vena, A.; Ball, L.; Robba, C.; Pelosi, P.; et al. Early detection of sepsis with machine learning techniques: A brief clinical perspective. Front. Med. 2021, 8, 617486. [Google Scholar] [CrossRef]
  28. Yan, M.Y.; Gustad, L.T.; Nytrø, Ø. Sepsis prediction, early detection, and identification using clinical text for machine learning: A systematic review. J. Am. Med. Inform. Assoc. 2022, 29, 559–575. [Google Scholar] [CrossRef]
  29. Fleuren, L.M.; Klausch, T.L.; Zwager, C.L.; Schoonmade, L.J.; Guo, T.; Roggeveen, L.F.; Swart, E.L.; Girbes, A.R.; Thoral, P.; Ercole, A.; et al. Machine learning for the prediction of sepsis: A systematic review and meta-analysis of diagnostic test accuracy. Intensive Care Med. 2020, 46, 383–400. [Google Scholar] [CrossRef]
  30. Wang, D.; Li, J.; Sun, Y.; Ding, X.; Zhang, X.; Liu, S.; Han, B.; Wang, H.; Duan, X.; Sun, T. A machine learning model for accurate prediction of sepsis in ICU patients. Front. Public Health 2021, 9, 754348. [Google Scholar] [CrossRef]
  31. Kijpaisalratana, N.; Sanglertsinlapachai, D.; Techaratsami, S.; Musikatavorn, K.; Saoraya, J. Machine learning algorithms for early sepsis detection in the emergency department: A retrospective study. Int. J. Med. Inform. 2022, 160, 104689. [Google Scholar] [CrossRef]
  32. Nemati, S.; Holder, A.; Razmi, F.; Stanley, M.D.; Clifford, G.D.; Buchman, T.G. An interpretable machine learning model for accurate prediction of sepsis in the ICU. Crit. Care Med. 2018, 46, 547. [Google Scholar] [CrossRef] [PubMed]
  33. Singh, Y.V.; Singh, P.; Khan, S.; Singh, R.S. A machine learning model for early prediction and detection of sepsis in intensive care unit patients. J. Healthc. Eng. 2022, 2022, 9263391. [Google Scholar] [CrossRef] [PubMed]
  34. Levy, M.M.; Fink, M.P.; Marshall, J.C.; Abraham, E.; Angus, D.; Cook, D.; Cohen, J.; Opal, S.M.; Vincent, J.L.; Ramsay, G.; et al. 2001 sccm/esicm/accp/ats/sis international sepsis definitions conference. Inten-Sive Care Med. 2003, 29, 530–538. [Google Scholar] [CrossRef] [PubMed]
  35. Singer, M.; Deutschman, C.S.; Seymour, C.W.; Shankar-Hari, M.; Annane, D.; Bauer, M.; Bellomo, R.; Bernard, G.R.; Chiche, J.D.; Coopersmith, C.M.; et al. The third international consensus definitions for sepsis and septic shock (Sepsis-3). JAMA 2016, 315, 801–810. [Google Scholar] [CrossRef]
  36. Seymour, C.W.; Liu, V.X.; Iwashyna, T.J.; Brunkhorst, F.M.; Rea, T.D.; Scherag, A.; Rubenfeld, G.; Kahn, J.M.; Shankar-Hari, M.; Singer, M.; et al. Assessment of clinical criteria for sepsis: For the Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA 2016, 315, 762–774. [Google Scholar] [CrossRef]
  37. Qiao, N. A systematic review on machine learning in sellar region diseases: Quality and reporting items. Endocr. Connect. 2019, 8, 952–960. [Google Scholar] [CrossRef]
  38. Gholamzadeh, M.; Abtahi, H.; Safdari, R. Comparison of different machine learning algorithms to classify patients suspected of having sepsis infection in the intensive care unit. Inform. Med. Unlocked 2023, 38, 101236. [Google Scholar] [CrossRef]
  39. Duan, Y.; Huo, J.; Chen, M.; Hou, F.; Yan, G.; Li, S.; Wang, H. Early prediction of sepsis using double fusion of deep features and handcrafted features. Appl. Intell. 2023, 53, 17903–17919. [Google Scholar] [CrossRef]
  40. Strickler, E.A.; Thomas, J.; Thomas, J.P.; Benjamin, B.; Shamsuddin, R. Exploring a global interpretation mechanism for deep learning networks when predicting sepsis. Sci. Rep. 2023, 13, 3067. [Google Scholar] [CrossRef]
  41. Zhou, A.; Beyah, R.; Kamaleswaran, R. OnAI-comp: An online ai experts competing framework for early sepsis detection. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 19, 3595–3603. [Google Scholar] [CrossRef]
  42. Al-Mualemi, B.Y.; Lu, L. A deep learning-based sepsis estimation scheme. IEEE Access 2020, 9, 5442–5452. [Google Scholar] [CrossRef]
  43. Rosnati, M.; Fortuin, V. MGP-AttTCN: An interpretable machine learning model for the prediction of sepsis. PLoS ONE 2021, 16, e0251248. [Google Scholar] [CrossRef] [PubMed]
  44. Zhang, D.; Yin, C.; Hunold, K.M.; Jiang, X.; Caterino, J.M.; Zhang, P. An interpretable deep-learning model for early prediction of sepsis in the emergency department. Patterns 2021, 2, 100196. [Google Scholar] [CrossRef] [PubMed]
  45. Shashikumar, S.P.; Josef, C.S.; Sharma, A.; Nemati, S. DeepAISE–an interpretable and recurrent neural survival model for early prediction of sepsis. Artif. Intell. Med. 2021, 113, 102036. [Google Scholar] [CrossRef] [PubMed]
  46. Aşuroğlu, T.; Oğul, H. A deep learning approach for sepsis monitoring via severity score estimation. Comput. Methods Programs Biomed. 2021, 198, 105816. [Google Scholar] [CrossRef]
  47. Oei, S.P.; van Sloun, R.J.; van der Ven, M.; Korsten, H.H.; Mischi, M. Towards early sepsis detection from measurements at the general ward through deep learning. Intell. Based Med. 2021, 5, 100042. [Google Scholar] [CrossRef]
  48. Rafiei, A.; Rezaee, A.; Hajati, F.; Gheisari, S.; Golzan, M. SSP: Early prediction of sepsis using fully connected LSTM-CNN model. Comput. Biol. Med. 2021, 128, 104110. [Google Scholar] [CrossRef]
  49. Goh, K.H.; Wang, L.; Yeow, A.Y.; Poh, H.; Li, K.; Yeow, J.J.; Tan, G.Y. Artificial intelligence in sepsis early prediction and diagnosis using unstructured data in healthcare. Nat. Commun. 2021, 12, 711. [Google Scholar] [CrossRef]
  50. Bedoya, A.D.; Futoma, J.; Clement, M.E.; Corey, K.; Brajer, N.; Lin, A.; Simons, M.G.; Gao, M.; Nichols, M.; Balu, S.; et al. Machine learning for early detection of sepsis: An internal and temporal validation study. JAMIA Open 2020, 3, 252–260. [Google Scholar] [CrossRef]
  51. Yang, M.; Liu, C.; Wang, X.; Li, Y.; Gao, H.; Liu, X.; Li, J. An explainable artificial intelligence predictor for early detection of sepsis. Crit. Care Med. 2020, 48, e1091-6. [Google Scholar] [CrossRef]
  52. Yuan, K.C.; Tsai, L.W.; Lee, K.H.; Cheng, Y.W.; Hsu, S.C.; Lo, Y.S.; Chen, R.J. The development an artificial intelligence algorithm for early sepsis diagnosis in the intensive care unit. Int. J. Med. Inform. 2020, 141, 104176. [Google Scholar] [CrossRef] [PubMed]
  53. Kok, C.; Jahmunah, V.; Oh, S.L.; Zhou, X.; Gururajan, R.; Tao, X.; Cheong, K.H.; Gururajan, R.; Molinari, F.; Acharya, U.R. Automated prediction of sepsis using temporal convolutional network. Comput. Biol. Med. 2020, 127, 103957. [Google Scholar] [CrossRef] [PubMed]
  54. Reyna, M.A.; Josef, C.S.; Jeter, R.; Shashikumar, S.P.; Westover, M.B.; Nemati, S.; Clifford, G.D.; Sharma, A. Early Prediction of Sepsis from Clinical Data: The PhysioNet/Computing in Cardiology Challenge 2019. Crit. Care Med. 2020, 48, 210–217. [Google Scholar] [CrossRef]
  55. Choi, J.S.; Trinh, T.X.; Ha, J.; Yang, M.S.; Lee, Y.; Kim, Y.E.; Choi, J.; Byun, H.G.; Song, J.; Yoon, T.H. Implementation of complementary model using optimal combination of hematological parameters for sepsis screening in patients with fever. Sci. Rep. 2020, 10, 273. [Google Scholar] [CrossRef] [PubMed]
  56. Kim, J.; Chang, H.; Kim, D.; Jang, D.H.; Park, I.; Kim, K. Machine learning for prediction of septic shock at initial triage in emergency department. J. Crit. Care 2020, 55, 163–170. [Google Scholar] [CrossRef]
  57. Ibrahim, Z.M.; Wu, H.; Hamoud, A.; Stappen, L.; Dobson, R.J.; Agarossi, A. On classifying sepsis heterogeneity in the ICU: Insight using machine learning. J. Am. Med. Inform. Assoc. 2020, 27, 437–443. [Google Scholar] [CrossRef]
  58. Fagerström, J.; Bång, M.; Wilhelms, D.; Chew, M.S. LiSep LSTM: A machine learning algorithm for early detection of septic shock. Sci. Rep. 2019, 9, 15132. [Google Scholar] [CrossRef]
  59. Kaji, D.A.; Zech, J.R.; Kim, J.S.; Cho, S.K.; Dangayach, N.S.; Costa, A.B.; Oermann, E.K. An attention based deep learning model of clinical events in the intensive care unit. PLoS ONE 2019, 14, e0211057. [Google Scholar] [CrossRef]
  60. Giannini, H.M.; Ginestra, J.C.; Chivers, C.; Draugelis, M.; Hanish, A.; Schweickert, W.D.; Fuchs, B.D.; Meadows, L.; Lynch, M.; Donnelly, P.J.; et al. A machine learning algorithm to predict severe sepsis and septic shock: Development, implementation and impact on clinical practice. Crit. Care Med. 2019, 47, 1485. [Google Scholar] [CrossRef]
  61. Ginestra, J.C.; Giannini, H.M.; Schweickert, W.D.; Meadows, L.; Lynch, M.J.; Pavan, K.; Chivers, C.J.; Draugelis, M.; Donnelly, P.J.; Fuchs, B.D.; et al. Clinician perception of a machine learning-based early warning system designed to predict severe sepsis and septic shock. Crit. Care Med. 2019, 47, 1477. [Google Scholar] [CrossRef]
  62. Schamoni, S.; Lindner, H.A.; Schneider-Lindner, V.; Thiel, M.; Riezler, S. Leveraging implicit expert knowledge for non-circular machine learning in sepsis prediction. Artif. Intell. Med. 2019, 100, 101725. [Google Scholar] [CrossRef] [PubMed]
  63. Barton, C.; Chettipally, U.; Zhou, Y.; Jiang, Z.; Lynn-Palevsky, A.; Le, S.; Calvert, J.; Das, R. Evaluation of a machine learning algorithm for up to 48-hour advance prediction of sepsis using six vital signs. Comput. Biol. Med. 2019, 109, 79–84. [Google Scholar] [CrossRef]
  64. Delahanty, R.J.; Alvarez, J.; Flynn, L.M.; Sherwin, R.L.; Jones, S.S. Development and evaluation of a machine learning model for the early identification of patients at risk for sepsis. Ann. Emerg. Med. 2019, 73, 334–344. [Google Scholar] [CrossRef]
  65. Scherpf, M.; Gräßer, F.; Malberg, H.; Zaunseder, S. Predicting sepsis with a recurrent neural network using the MIMIC III database. Comput. Biol. Med. 2019, 113, 103395. [Google Scholar] [CrossRef] [PubMed]
  66. Bloch, E.; Rotem, T.; Cohen, J.; Singer, P.; Aperstein, Y. Machine learning models for analysis of vital signs dynamics: A case for sepsis onset prediction. J. Healthc. Eng. 2019, 2019, 5930379. [Google Scholar] [CrossRef] [PubMed]
  67. van Wyk, F.; Khojandi, A.; Kamaleswaran, R. Improving prediction performance using hierarchical analysis of real-time data: A sepsis case study. IEEE J. Biomed. Health Inform. 2019, 23, 978–986. [Google Scholar] [CrossRef]
  68. van Wyk, F.; Khojandi, A.; Mohammed, A.; Begoli, E.; Davis, R.L.; Kamaleswaran, R. A minimal set of physiomarkers in continuous high frequency data streams predict adult sepsis onset earlier. Int. J. Med. Inform. 2019, 122, 55–62. [Google Scholar] [CrossRef]
  69. Yee, C.R.; Narain, N.R.; Akmaev, V.R.; Vemulapalli, V. A data-driven approach to predicting septic shock in the intensive care unit. Biomed. Inform. Insights 2019, 11, 1178222619885147. [Google Scholar] [CrossRef]
  70. Mao, Q.; Jay, M.; Hoffman, J.L.; Calvert, J.; Barton, C.; Shimabukuro, D.; Shieh, L.; Chettipally, U.; Fletcher, G.; Kerem, Y.; et al. Multicentre validation of a sepsis prediction algorithm using only vital sign data in the emergency department, general ward and ICU. BMJ Open 2018, 8, e017833. [Google Scholar] [CrossRef]
  71. Taneja, I.; Reddy, B.; Damhorst, G.; Dave Zhao, S.; Hassan, U.; Price, Z.; Jensen, T.; Ghonge, T.; Patel, M.; Wachspress, S.; et al. Combining biomarkers with EMR data to identify patients in different phases of sepsis. Sci. Rep. 2017, 7, 10800. [Google Scholar] [CrossRef]
  72. Horng, S.; Sontag, D.A.; Halpern, Y.; Jernite, Y.; Shapiro, N.I.; Nathanson, L.A. Creating an automated trigger for sepsis clinical decision support at emergency department triage using machine learning. PLoS ONE 2017, 12, e0174708. [Google Scholar] [CrossRef]
  73. Kam, H.J.; Kim, H.Y. Learning representations for the early detection of sepsis with deep neural networks. Comput. Biol. Med. 2017, 89, 248–255. [Google Scholar] [CrossRef] [PubMed]
  74. Shashikumar, S.P.; Li, Q.; Clifford, G.D.; Nemati, S. Multiscale network representation of physiological time series for early prediction of sepsis. Physiol. Meas. 2017, 38, 2235. [Google Scholar] [CrossRef] [PubMed]
  75. Calvert, J.S.; Price, D.A.; Chettipally, U.K.; Barton, C.W.; Feldman, M.D.; Hoffman, J.L.; Jay, M.; Das, R. A computational approach to early sepsis detection. Comput. Biol. Med. 2016, 74, 69–73. [Google Scholar] [CrossRef] [PubMed]
  76. Desautels, T.; Calvert, J.; Hoffman, J.; Jay, M.; Kerem, Y.; Shieh, L.; Shimabukuro, D.; Chettipally, U.; Feldman, M.D.; Barton, C.; et al. Prediction of sepsis in the intensive care unit with minimal electronic health record data: A machine learning approach. JMIR Med. Inf. 2016, 4, e28. [Google Scholar] [CrossRef] [PubMed]
  77. Brown, S.M.; Jones, J.; Kuttler, K.G.; Keddington, R.K.; Allen, T.L.; Haug, P. Prospective evaluation of an automated method to identify patients with severe sepsis or septic shock in the emergency department. BMC Emerg. Med. 2016, 16, 31. [Google Scholar] [CrossRef] [PubMed]
  78. Henry, K.E.; Hager, D.N.; Pronovost, P.J.; Saria, S. A targeted real-time early warning score (TREWScore) for septic shock. Sci. Transl. Med. 2015, 7, 299ra122. [Google Scholar] [CrossRef]
  79. Taylor, S.P.; Bray, B.C.; Chou, S.H.; Burns, R.; Kowalkowski, M.A. Clinical subtypes of sepsis survivors predict readmission and mortality after hospital discharge. Ann. Am. Thorac. Soc. 2022, 19, 1355–1363. [Google Scholar] [CrossRef]
  80. Tang, B.M.; Eslick, G.D.; Craig, J.C.; McLean, A.S. Accuracy of procalcitonin for sepsis diagnosis in critically ill patients: Systematic review and meta-analysis. Lancet Infect. Dis. 2007, 7, 210–217. [Google Scholar] [CrossRef]
Figure 1. Search and selection process using PRISMA flowchart.
Figure 1. Search and selection process using PRISMA flowchart.
Jcm 12 05658 g001
Figure 2. Summary of data sources (A), study types (B), and study populations (C).
Figure 2. Summary of data sources (A), study types (B), and study populations (C).
Jcm 12 05658 g002
Figure 3. Distribution of data sources used to develop machine learning models.
Figure 3. Distribution of data sources used to develop machine learning models.
Jcm 12 05658 g003
Figure 4. Sepsis prevalence (%) in different selected studies. The red line shows the median prevalence.
Figure 4. Sepsis prevalence (%) in different selected studies. The red line shows the median prevalence.
Jcm 12 05658 g004
Figure 5. Different data types used in different studies.
Figure 5. Different data types used in different studies.
Jcm 12 05658 g005
Figure 6. Number of variables used in different studies. The red line shows the median of the number of variables used in different studies.
Figure 6. Number of variables used in different studies. The red line shows the median of the number of variables used in different studies.
Jcm 12 05658 g006
Figure 7. Top-ranked 20 features among the studies.
Figure 7. Top-ranked 20 features among the studies.
Jcm 12 05658 g007
Figure 8. Early prediction of sepsis onset among different studies.
Figure 8. Early prediction of sepsis onset among different studies.
Jcm 12 05658 g008
Table 1. Summary of the selected articles.
Table 1. Summary of the selected articles.
NumberAuthorsYear of PublicationDataset(s)Sepsis DefinitionSample SizeSepsis Positive PatientsSepsis Prevalence (%)Data TypesNumber of VariablesModel(s)AUROCRetrospective (R) or Prospective (P) ?Output MetricsHours before Onset
1Gholamzadeh et al. [38]2023MIMIC-IIISepsis-31,552,21027,9161.79Demographics,
labs, vitals
38DT, RF, XGB0.918RAUROC, A, P, R, S, F1-
2Duan et al. [39]2023Shanghai HospitalSepsis-228211440.43Demographics,
labs, vitals
29Double fusion DL framework0.92RAUC, A, S, R24
3Strickler et al. [40]20232019 PhysioNet/CinC ChallengeSepsis-340,33629327.33Demographics,
labs, vitals
40CSE, MSC-LSTM,
MSC-CSE
-RA, P, S, R, F1, MCC, Utility-
4Zhou et al. [41]20212019 PhysioNet/CinC ChallengeSepsis-340,33629327.33Demographics,
labs, vitals
40RG, XGB, LR, RF, SVC-RAverage regret-
5Al-Mualemi and Lu [42]20212019 PhysioNet/CinC ChallengeSepsis-340,33629327.33Demographics,
labs, vitals
40SVM, RNN-LSTM, Adaptive CNN-RA, S, R-
6Rosnati and Fortuin [43]2021MIMIC-IIISepsis-322,007793636.06Labs, vitals24MGP-AttTCN0.660RAUC5
7Zhang et al. [44]20212019 DII National
Data Science Challenge
Sepsis-2178,84352,80229.5Demographics,
labs, vitals
145GRU, RETAIN, Dipole, LSTM0.892RAUC4
8Shashikumar et al. [45]2021Emory Healthcare System,
UCSD,
MIMIC-III
Sepsis-385,04647945.64Demographics,
labs, vitals
65DeepAISE0.90RAUC4
9Aşuroğlu and Oğul [46] 2021MIMIC-IIISepsis-35154140427.24Vitals7DSPA0.97RAUC6
10Oei et al. [47]2021MIMIC-IIISepsis-348,63213,93528.65Labs, vitals8Deep learning0.86RAUC3
11Rafiei et al. [48]20212019 PhysioNet/CinC ChallengeSepsis-340,33629327.33Demographics,
labs, vitals
40Smart Sepsis Predictor 0.86RAUC, S, A12
12Goh et al. [49]2021Singapore government-based hospitalSepsis-353173276.15Demographic, vitals, notes, labs, treatment 15SERA algorithm0.94PAUC, R, S12
13Bedoya et al. [50]2020Duke University Health SystemSepsis-242,979816019.0Demographics, vital, lab, comorbidities,
medications
86RNN, LR, RF0.88RAUROC5
14Yang et al. [51]20202019 PhysioNet/CinC ChallengeSepsis-340,33629327.33Demographics,
labs, vitals
168XGB, EASP0.85RUtility score, AUROC, R, S6
15Yuan et al. [52]2020Taipei
Medical University Hospital
Sepsis-2158844427.96Vitals, lab, exam reports,
text and images
18GB0.89PF1 score, A, R, S,
PPV
-
16Kok et al. [53]20202019 PhysioNet/CinC ChallengeSepsis-340,00029327.33Demographics,
labs, vitals
40TCN 0.99RA, R, S, AUROC, AUPRC-
17Reyna et al. [54]2020Emory Healthcare System,
MIMIC-III
Sepsis-360,00029327.3Demographics,
labs, vitals
40ML Models0.815RAUROC, Utility 6
18Lauritsen et al. [19]2020CROSSTRACKSSepsis-23126--Diagnoses, labs, vitals, imaging,
medications,
treatment
6CNN-LSTM
0.88RAUROC, mAP3
19Choi et al. [55]2020Yonsei University (YU) Severance HospitalSepsis
(ICD-10)
7743113614.67laboratory data36LR0.86RA, AUROC, R, S, PPV, NPV -
20Kim et al. [56]2020Tertiary academic
hospital
Sepsis-349,56048179.7Demographics,
labs, vitals
29SVM, GB, MARS, LASSO, Ridge, RF0.93RAUROC, R, S, PPV, NPV-
21Ibrahim et al. [57]2020MIMIC-IIISepsis-213,728425631.0Vitals, labs 63RF, GB, SVM0.96RR, S, AUC, PLR, NLR-
22Fagerstrom et al. [58]2019MIMIC-IIISepsis-259,000--Demographics,
labs, vitals, treatment, medications, diagnoses
-LSTM 0.83RAUROC 40
23Kaji et al. [59]2019MIMIC-IIISepsis-236,17623,00863.6Demographics,
labs, vitals, medications
119LSTM0.88RAUROC, R, PPV24
24Giannini et al. [60]2019University of Pennsylvania Health SystemSepsis (ICD-9)172,7009500.55Demographics,
labs, vitals
20RF0.88RAUROC, R, S6
25Ginestra et al. [61]2019University of Pennsylvania Health SystemSepsis (ICD-9)162,2129430.58Vitals, comorbidity, labs-Early Warning Systems 2.0 (EWS 2.0)-PResponse rate6
26Schamoni et al. [62]2019University Medical Center, MannheimSepsis-362020032.3Demographics,
Labs, clinical data
55Non-Linear ordinal regression0.84RAUROC 4
27Barton et al. [63]2019MIMIC-III, UCSFSepsis-3112,95236733.3Vitals6XGBoost0.88RAUROC, R, S24
28Delahanty et al. [64] 2019Tenet
Healthcare System
Sepsis-32,759,52954,6611.98Demographics,
labs, vitals, medications, nursing notes
13GB 0.97RAUROC, R, S, P 24
29Scherpf et al. [65]2019MIMIC-IIISepsis-246,52027247.7Lab, vitals 10RNN-GRU, Insight0.81RAUROC3
30Bloch et al. [66]2019RMCSepsis-260030050Vitals 4NN, SVM, LR0.88RAUROC, R, S, PPV, NPV 4
31VanWyk et al. [67]2019MLH SystemSepsis
(ICD10)
586--Lab, vitals7RF, RNN-RF2 score, A, R, S, PPV4
32van Wyk et al. [68]2019MLH SystemSepsis-2116137732.47Demographics, vitals8RF-RR, F15
33Yee et al. [69]2019MIMIC-IIISepsis-391658729.5Demographics,
labs, vitals, diagnosis
Bayesian network0.81RAUROC, R, S, PPV, NPV24
34Mao et al. [70]2018MIMIC-III, UCSFSepsis-290,35319659.1Vitals30Insight0.92RAUROC, R, S4
35Nemati et al. [32]2018Emory Healthcare SystemSepsis-327,52723758.6Demographics,
labs, vitals, clinical data
65AISE 0.85RA, R, S, AUC4
36Taneja et al. [71]2017Carle Foundation HospitalSepsis-34447617.11Demographics,
labs, vitals
15SVM0.81RR, S, AUC -
37Horng et al. [72]2017Beth Israel Deaconess Medical CenterSepsis (ICD-9)230,93632,33114Vitals, demographics, and notes 12SVM0.86RAUROC, R, S, PPV -
38Kam and Kim [73]2017MIMIC-IISepsis-263623606.2Demographics,
labs, vitals
9Sepsis LSTM (SepLSTM)0.92RA, R, S, AUC 3
39Shashikumar et al. [74]2017Emory Healthcare SystemSepsis-325010040.0Demographics, comorbidity, clinical context, vitals2SVM0.80PAUROC, S2
40Calvert et al. [75]2016MIMIC-IISepsis-2139415911.4Demographics,
labs, vitals
9InSight0.92RA, AUROC, R, S3
41Desautels et al. [76]2016MIMIC-IIISepsis-322,85318409.7Demographics,
labs, vitals
8InSight0.88RAUROC4
42Brown et al. [77]2016LDS Hospital and Intermountain
Medical Center
Sepsis (ICD-9)132,7485490.41Demographics,
labs, vitals
9NB0.953RAUROC, R, S, PPV, NPV-
Note: MIMIC—Medical Information Mart for Intensive Care, DT—Decision tree, RF—Random Forest, XGB—Extreme Gradient Boosting (XGboost), ROC—Receiver Operating Characteristic Curve, AUROC—Area under the ROC Curve, A—Accuracy, P—Precession, R—Recall (Sensitivity), S—Specificity, F1—Harmonic mean of precision and recall, AUC—Area under the ROC Curve, CSE—Computational sepsis expert, MSC—Multi-set classifier, MSC-CSE—Multi-set classifier–Computational sepsis expert, MCC—Matthews’s correlation coefficient, RG—Random Guess, LR—Logistic Regression, SVC—Support Vector Classifier, SVM—Support Vector Machine, LSTM—Long short-term memory networks, CNN—Convolutional neural network, MGP-AttTCN—Multitask Gaussian Process and attention-based deep learning model, GRU—Gated Recurrent Unit, RETAIN—REverse Time AttentIoN model (RETAIN), UCSF—University of California, San Francisco, UCSD—University of California, San Diego, MLH—Methodist Le Bonheur Healthcare, DeepAISE—Deep Artificial Intelligence Sepsis Expert, DSPA—Deep SOFA-Sepsis Prediction Algorithm, SERA—Sequential Element Rejection and Admission, EASP—explainable AI sepsis predictor, GB—Gradient Boosting, MARS—Multivariate Adaptive Regression Splines, LASSO—Least Absolute Shrinkage and Selection Operator, PPV—Positive predictive value, TCN—Temporal Convolutional Network, AUPRC—Area under Precision Recall Curve, ML—Machine Learning, mAP—Mean average precision, NPV—Negative predictive value, PLR—Positive likelihood ratio, NLR—Negative likelihood ratio, RNN—Recurrent Neural Network, NN—Neural Network, AISE—Artificial Intelligence Sepsis Expert, NB—Naïve Bayes Classifier, Latter-day Saints (LDS) Hospital and Intermountain Medical Center.
Table 2. A quality assessment tool for the included studies.
Table 2. A quality assessment tool for the included studies.
NumberAuthorsYear of PublicationSample Size > 50 Data Availability Code Availability Mobile or Web DeploymentHandling of Missing DataSepsis PrevalenceFeature EngineeringMachine Learning ModelsHyperparameters Sepsis Definition Adhered in Study DesignValid Methods to Prevent OverfittingReporting of Performance MatrixExternal Data ValidationExplainabilityLimitations of the StudyClinical Applicability DiscussionScore (%)Quality Category
1Gholamzadeh et al. [38]2023×××××68.75AAQ
2Duan et al. [39]2023×××××××56.25AQ
3Strickler et al. [40]2023××××75.00AAQ
4Zhou et al. [41]2021×××81.25HQ
5Al-Mualemi and Lu [42]2021××××××62.50AAQ
6Rosnati and Fortuin [43]2021×××××68.75AAQ
7Zhang et al. [44]2021××××75.00AAQ
8Shashikumar et al. [45]2021××87.50HQ
9Aşuroğlu and Oğul [46] 2021×××××68.75AAQ
10Oei et al. [47]2021××××××62.50AAQ
11Rafiei et al. [48]2021×××××68.75AAQ
12Goh et al. [49]2021×××××××56.25AQ
13Bedoya et al. [50]2020××××××62.50AAQ
14Yang et al. [51]2020×××××68.75AAQ
15Yuan et al. [52]2020××××××××50.00AQ
16Kok et al. [53]2020×××××68.75AAQ
17* Reyna et al. [54]2020×××××××××43.75AQ
18Lauritsen et al. [19]2020×××××68.75AAQ
19Choi et al. [55]2020××××××××50.00AQ
20Kim et al. [56]2020××××××62.50AAQ
21Ibrahim et al. [57]2020××××××62.50AAQ
22Fagerstrom et al. [58]2019×××81.25HQ
23Kaji et al. [59]2019×××81.25HQ
24Giannini et al. [60]2019×××××××××××31.25LQ
25Ginestra et al. [61]2019××××××××××37.50LQ
26Schamoni et al. [62]2019××××××62.50AAQ
27Barton et al. [63]2019××××75.00AAQ
28Delahanty et al. [64] 2019××××××××50.00AQ
29Scherpf et al. [65]2019××××××62.50AAQ
30Bloch et al. [66]2019×××××××××43.75AQ
31VanWyk et al. [67]2019××××××××50.00AQ
32van Wyk et al. [68]2019×××××××56.25AQ
33Yee et al. [69]2019×××××××××43.75AQ
34Mao et al. [70]2018×××××68.75AAQ
35Nemati et al. [32]2018×××81.25HQ
36Taneja et al. [71]2017××××××××××37.50LQ
37Horng et al. [72]2017××××××××50.00AQ
38Kam and Kim [73]2017××××××62.50AAQ
39Shashikumar et al. [74]2017××××××××××37.50LQ
40Calvert et al. [75]2016××××××××50.00AQ
41Desautels et al. [76]2016×××××××56.25AQ
42Brown et al. [77]2016××××××××××37.50LQ
% prevalence of each category 100%52%14%5%69%95%83%93%46%100%43%100%17%21%74%50%
Note: * this study has reported the outcome of a challenge.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Islam, K.R.; Prithula, J.; Kumar, J.; Tan, T.L.; Reaz, M.B.I.; Sumon, M.S.I.; Chowdhury, M.E.H. Machine Learning-Based Early Prediction of Sepsis Using Electronic Health Records: A Systematic Review. J. Clin. Med. 2023, 12, 5658. https://doi.org/10.3390/jcm12175658

AMA Style

Islam KR, Prithula J, Kumar J, Tan TL, Reaz MBI, Sumon MSI, Chowdhury MEH. Machine Learning-Based Early Prediction of Sepsis Using Electronic Health Records: A Systematic Review. Journal of Clinical Medicine. 2023; 12(17):5658. https://doi.org/10.3390/jcm12175658

Chicago/Turabian Style

Islam, Khandaker Reajul, Johayra Prithula, Jaya Kumar, Toh Leong Tan, Mamun Bin Ibne Reaz, Md. Shaheenur Islam Sumon, and Muhammad E. H. Chowdhury. 2023. "Machine Learning-Based Early Prediction of Sepsis Using Electronic Health Records: A Systematic Review" Journal of Clinical Medicine 12, no. 17: 5658. https://doi.org/10.3390/jcm12175658

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop