Next Article in Journal
Exploring the Social Networks of Women Bereaved by Stillbirth: A Descriptive Qualitative Study
Next Article in Special Issue
Seasonal Variation and Severity of Acute Abdomen in Japan: A Nine-Year Retrospective Analysis
Previous Article in Journal
High Skin Sympathetic Nerve Activity in Patients with Recurrent Syncope
Previous Article in Special Issue
Monocyte Distribution Width, Neutrophil-to-Lymphocyte Ratio, and Platelet-to-Lymphocyte Ratio Improves Early Prediction for Sepsis at the Emergency
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Model to Identify Sepsis Patients in the Emergency Department: Algorithm Development and Validation

1
Graduate Institute of Biomedical Informatics, College of Medicine Science and Technology, Taipei Medical University, Taipei 106, Taiwan
2
Emergency Department, Taoyuan General Hospital, Ministry of Health and Welfare, Taoyuan 330, Taiwan
3
Emergency Department, Chi-Mei Medical Center, Tainan 710, Taiwan
4
Department of Neurosurgery, Taipei Medical University-Wan Fang Hospital, Taipei 116, Taiwan
5
Taipei Neuroscience Institute, Taipei Medical University, Taipei 110, Taiwan
6
International Center for Health Information Technology (ICHIT), Taipei Medical University, Taipei 110, Taiwan
7
Research Center of Big Data and Meta-Analysis, Wan Fang Hospital, Taipei Medical University, Taipei 116, Taiwan
8
Department of Neurosurgery, Shuang Ho Hospital, Taipei Medical University, New Taipei City 235, Taiwan
*
Author to whom correspondence should be addressed.
Current Address: Graduate Institute of Biomedical Informatics, Taipei Medical University, Taipei 106, Taiwan.
J. Pers. Med. 2021, 11(11), 1055; https://doi.org/10.3390/jpm11111055
Submission received: 14 September 2021 / Revised: 11 October 2021 / Accepted: 18 October 2021 / Published: 21 October 2021
(This article belongs to the Special Issue New Advances in Treatment of Sepsis)

Abstract

:
Accurate stratification of sepsis can effectively guide the triage of patient care and shared decision making in the emergency department (ED). However, previous research on sepsis identification models focused mainly on ICU patients, and discrepancies in model performance between the development and external validation datasets are rarely evaluated. The aim of our study was to develop and externally validate a machine learning model to stratify sepsis patients in the ED. We retrospectively collected clinical data from two geographically separate institutes that provided a different level of care at different time periods. The Sepsis-3 criteria were used as the reference standard in both datasets for identifying true sepsis cases. An eXtreme Gradient Boosting (XGBoost) algorithm was developed to stratify sepsis patients and the performance of the model was compared with traditional clinical sepsis tools; quick Sequential Organ Failure Assessment (qSOFA) and Systemic Inflammatory Response Syndrome (SIRS). There were 8296 patients (1752 (21%) being septic) in the development and 1744 patients (506 (29%) being septic) in the external validation datasets. The mortality of septic patients in the development and validation datasets was 13.5% and 17%, respectively. In the internal validation, XGBoost achieved an area under the receiver operating characteristic curve (AUROC) of 0.86, exceeding SIRS (0.68) and qSOFA (0.56). The performance of XGBoost deteriorated in the external validation (the AUROC of XGBoost, SIRS and qSOFA was 0.75, 0.57 and 0.66, respectively). Heterogeneity in patient characteristics, such as sepsis prevalence, severity, age, comorbidity and infection focus, could reduce model performance. Our model showed good discriminative capabilities for the identification of sepsis patients and outperformed the existing sepsis identification tools. Implementation of the ML model in the ED can facilitate timely sepsis identification and treatment. However, dataset discrepancies should be carefully evaluated before implementing the ML approach in clinical practice. This finding reinforces the necessity for future studies to perform external validation to ensure the generalisability of any developed ML approaches.

Graphical Abstract

1. Introduction

Sepsis, defined as “life threatening organ dysfunction caused by a dysregulated host response to infection” [1], is a global health problem with high mortality and morbidity [2]. Epidemiologic estimates have reported that the crude mortality of sepsis patients is over 20% [2,3,4,5], and the global cost of sepsis is estimated to be $16.7 billion [6,7]. Numerous studies [8,9,10] have demonstrated that timely identification of sepsis and initiation of an evidenced-based treatment protocol could decrease in-hospital mortality, shorten length of stay and reduce healthcare costs. Nevertheless, because of the heterogeneity of infectious insults and the diversity of hosts, efficiently recognising and treating sepsis remains highly challenging for physicians [11].
Early sepsis identification relies upon clinical data that is readily available during hospitalisation [12]. The currently available clinical sepsis risk scores, namely the Systemic Inflammatory Response Syndrome (SIRS) and quick Sequential Organ Failure Assessment (qSOFA), have several shortcomings, which hamper their utilisation in identifying the early signs of organ failure [13,14,15]. Therefore, it is urgently needed to develop a more precise and personalised tool to recognise sepsis in a timely manner. The increasing availability of electronic health records (EHR) and advancing machine learning (ML) techniques has stimulated attempts to identify patient conditions through the automated analysis of medical records. Previous studies have shown that the ML approach can facilitate the detection of sepsis and septic shock [10,16,17,18]. However, the clinical utility of these models in the emergency department (ED) setting remains uncertain. The majority of previous studies developed and validated the ML models using clinical data only from the intensive care unit (ICU) [18,19,20,21,22,23].
To better evaluate the clinical utility of ML approaches for identifying sepsis patients in the ED, we developed an ML technique to correctly identify sepsis patients, using clinical predictors available in the electronic health record (EHR). Afterward, we validated our model externally using a distinct dataset from a geographically separate institute that provided a different level of care. Finally, we compared model performance with currently available risk scores.

2. Materials and Methods

Overview of the study: The ML model was developed and externally validated to identify sepsis patients in the ED. The overall process of our study is depicted in Figure 1.

2.1. Study Population

From the EHR, we retrospectively collected clinical information from all adults (≥20 years old) admitted to the ED as inpatients (July 2016 to October 2016) at Chi-Mei Medical Center, a tertiary teaching hospital located in Southern Taiwan. Sepsis cases were assessed based on a manual chart review according to the Rhee clinical surveillance criteria [24]. Two experienced clinicians independently reviewed the medical records of the study cohorts, throughout the clinical course from ED arrival to hospital charge or death, to determine whether a patient had sepsis. Patients were excluded if they were (a) less than 20 years old, (b) identified as septic patients before ED admission. This study was reviewed and approved by the Institutional Review Board of Human Research at both the Chi-Mei Medical Centre and the Taoyuan General Hospital (IRB No: TYGH107014).

2.2. Sepsis Definitions

Sepsis was confirmed when either one of the following two conditions were fulfilled (Table 1): (1) the Sepsis-3 definition [1], that is, having a suspected infection (prescription of antibiotics and sampling of bodily fluids for microbiological culture) combined with evidence of organ dysfunction, defined by an increase in the Sequential Organ Failure Assessment (SOFA) score greater than or equal to two, and (2) having a suspected infection combined with evidence of hypoperfusion and shock, including lactate >2 mmol/L and the presence of vasopressor medications.

2.3. Predictor Variables

We collected the following clinical information which was available in the EHR: patient’s vitals upon arrival acquired by a triage nurses including; systolic blood pressure (SBP), diastolic blood pressure (DBP), respiratory rate (RR), Glasgow coma scale (GCS), body temperature (BT), heart rate (HR), and the first acquired laboratory study results during the patient’s stay at the emergency department, including complete blood count, lactate level, C-reactive protein (CRP), random glucose, sodium level (Na), potassium level (K), blood urea nitrogen (BUN), creatinine (Cr), glutamic oxaloacetic transaminase (GOT), glutamate pyruvate transaminase (GPT), total bilirubin (T.bil), high sensitivity Troponin I (hs-TnI), and creatine kinase-MB (CK-MB).

2.4. Model Development and Validation

The development dataset was split into training and testing sets (internal validation) with an 80–20 ratio in a stratified fashion to preserve the same prevalence of sepsis cases as in the development dataset. We developed eXtreme Gradient Boosting (XGBoost), a highly scalable end-to-end tree boosting system proposed by Chen and Guestrin [25], on the training cohort using all clinical variables and validated this model internally to stratify sepsis patients.
XGBoost does not require data normalisation of input features and has the ability to cope with sparse data. It surpasses traditional tree-based models by introducing regularisation to avoid overfitting, by utilising gradient boosting to ensemble multiple tree models for better performance, and by mitigating biases. The objective function was utilised in minimising logistic loss, and we used the grid search method to tune the hyper-parameters of our model. During the training process, five-fold cross validation was applied to reduce sample bias. We selected the threshold that gave the highest F2-score. The equation for calculating the F2-score is given below:
F2 = (5 × Precision × Recall)/(4 × Precision + Recall)
Modelling was developed using the software Python version 3.6.3 and XGBoost Package version 1.2.1.

2.5. Evaluating Model Performance

We present the performance of the XGBoost model on the internal validation data for identifying sepsis using AUC. We calculated accuracy, sensitivity, specificity, negative predictive value (NPV), and positive predictive value (PPV). We compared the performance of model with traditional clinical tools, namely SIRS and qSOFA.

2.6. Feature Selection

To determine the major predictors of stratifying sepsis patients, feature selection was performed. We used XGBoost’s built-in function in Python, “feature importances”, and analysed the top ranking features. This provided the information of the relative contribution of the corresponding feature to the model calculated by taking each feature’s contribution for each tree in the model. A higher ranked feature on the chart implies that it is more important for generating the prediction.

2.7. External Validation

A separate cohort of 1744 unique adult patients admitted to the ED at Taoyuan General Hospital, a regional hospital located in Northern Taiwan, from January 2018 to March 2018 was used for external validation. We collected the following information from external validation datasets: (1) patient’s underlying comorbidities; (2) if antibiotics were prescribed, the triage-to-drug time; (3) whether the discharge summary, either from the ED or after hospitalisation, contained sepsis-related diagnosis codes; (4) documented infection focus for the sepsis.

2.8. Statistical Analysis

Statistical analyses were performed for top ranking predictors of the model between the development dataset and the external validation dataset. Additionally, in the external validation dataset, with the prediction results of the machine learning model, we compared the comorbidities and the infection source for sepsis between the true positive group and the false negative group, and between the true negative and the false positive group. All of the statistical analyses were performed using SAS Enterprise Guide 8.3. Student’s t-test was used for continuous variables and the chi-squared test was used for categorical variables to evaluate differences between the two groups.

2.9. Promoting Interoperability

This article followed TRIPOD (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis) guidelines [26] (Supplementary Table S1).

3. Results

3.1. Patient Characteristics

After applying the exclusion criteria, the final development cohort sizes were 8296 patients (6637 (80%) for training and 1659 (20%) for internal validation). However, the external validation cohort included 1744 patients, and 506 (29%) of them were sepsis patients. The mortality of sepsis patients in the development and the external validation dataset was 13.5% and 17%, respectively. In our external validation dataset, the average triage-to-antibiotic time for patients coded as having sepsis was 3.18 h, whereas the average triage-to-antibiotic time for true sepsis patients was 3.96 h (Table 2).

3.2. Model Performance for Identifying Sepsis Patients

When compared with the existing identification tools, the XGBoost model showed significantly greater discrimination of sepsis. In the development dataset, there were 1742 sepsis patients. The XGBoost model showed significantly greater discrimination (AUC: 0.86) in identifying sepsis patients. The area-under-the-curve (AUC) for external validation was 0.75. However, XGBoost exhibited a higher AUC compared with SIRS and qSOFA for identifying sepsis both in the internal and external validation sets (Figure 2).
Table 3 presents the performance comparison between the XGBoost model and traditional sepsis tools. In the internal validation, XGBoost had a sensitivity of 80% and specificity of 78%. For the identification of sepsis, SIRS had a sensitivity of 64% and specificity of 66%; qSOFA had a sensitivity of 35% and specificity of 96%. The predictive values of XGBoost (PPV: 0.47, NPV: 0.94) were higher than that of SIRS (PPV: 0.34, NPV: 0.77) and qSOFA (PPV: 0.76, NPV: 0.79). In the external validation, XGBoost had a sensitivity of 67% and specificity of 70%, which was higher than SIRS (sensitivity: 66%, specificity: 47%), and qSOFA (sensitivity: 36%, specificity: 89%). The PPV and NPV for XGBoost were 48% and 84%, respectively.

3.3. Most Important Predictors of Sepsis as Assessed with the XGBoost Model

Figure 3 shows the feature rankings of the XGBoost model and the statistical analysis of the 15 top-ranking features between septic patients in the development dataset and septic patients in the external validation dataset. These 15 top-ranking features accounted for 65% of the total weight.

3.4. Potential Clinical Confounders of Model Performance

Table 4 shows the statistical analysis of comorbidities and infection source for sepsis between the true positive (TP) group and the false negative (FN) group, and between the true negative (TN) and the false positive (FP) group in the external validation dataset after implementing the machine learning model. The result suggests that age, presence of coronary artery disease, chronic kidney disease, urinary tract infection, and pneumonia might interfere with the model performance.

4. Discussion

4.1. Main Findings

In the present study, we developed and externally validated an ML model to correctly identify sepsis in patients admitted to the ED. The XGBoost model demonstrated great performance with an AUROC of 0.85 and 0.75 in the internal and external validation, respectively. Our current model significantly outperformed the other clinically available stratifying tools. The findings of our study suggest that the XGBoost model has an important clinical role in identifying sepsis patients in the ED.

4.2. The Pivotal Role of the ED in Developing a Sepsis Identification Model

International consensus has continued to emphasise the benefit of the early recognition of sepsis, followed by timely treatment on patient outcomes [27]. Because the ED is generally the initial arrival site for septic patients [24,28], this recommendation underscores the pivotal role of optimising sepsis identification in the ED.
In order to minimise the risk of bias, our study followed the published recommendations [29,30] to develop and externally validate an ML model for sepsis identification in the ED. The study shows that ML, even when externally validated in a discrepant dataset, could demonstrate acceptable discriminative power in identifying sepsis patients and outperformed the existing SIRS and qSOFA criteria.
In contrast to the neural network, which is like a “black box”, we adopted the XGBoost model for its better clinical interpretability. According to Figure 2, the top-ranking features of the machine learning model, such as CRP, Na, Cr, BP, and platelets, correspond well with the key clinical features that physicians use to identify sepsis or to assess the severity of sepsis.

4.3. Machine Learning Might Help Shorten the Triage-to-Antibiotic Time

Early administration of antibiotics is crucial for improving outcomes in septic patients. A delay in starting antibiotics is associated with increased in-hospital mortality [30,31], especially in patients with septic shock. The survival rate can drop by 7.6% with each hour of delay after hypotension has developed [30]. However, the door-to-antibiotic time varies significantly among different attending physicians in the ED setting [32]; studies have shown that the median interval from the time of presentation to antibiotic administration can range from 71 to 359 min, with a median of 4 h [33,34]. Physician and hospital related factors that contribute to such variation in time to antibiotic treatment include diagnostic delays, computerised order entry systems and ED crowding [32]. In our external validation dataset, the average triage-to-antibiotic time for patients coded as having sepsis was 3.18 h, whereas the average triage-to-antibiotic time for true sepsis patients was 3.96 h. Previous prospective interventional studies have demonstrated that the implementation of computerised systems for assisting sepsis identification could shorten the time to intervention for patients who triggered the sepsis alert [35,36,37]. These results support the potential of ML for facilitating timely sepsis care and improving patient outcomes.

4.4. Factors Associated with the Heterogeneity of Model Performance

Variations in the predictive performance of a model across different patient cohorts have been well-documented for models developed using traditional statistical approaches [38,39,40,41]. However, the impact of heterogeneity in patient characteristics on model performance has rarely been assessed for sepsis prediction models in the previous literature [42,43,44,45]. Among published models targeting sepsis identification in the ED setting, only the one developed by Faisal et al. [46] was examined by external validation, which also showed discrepant performance between the development dataset and the external validation dataset. However, the authors did not elucidate the reason for the discrepancy. Heterogeneity between the two independent datasets can generally be categorised into “figure drifting” (i.e., differences in predictors) and “label drifting” (i.e., differences in outcomes) [45]. Figure drifting and label drifting can originate from the data itself (e.g., differences in prevalence or severity), or the criteria used for determining the predictor values or outcomes [41,47]. Different designs of external validation, such as temporal (i.e., same institute, different study period), institutional (i.e., two geographically adjacent institutes), and geographical (i.e., two institutes in different regions, or even countries), determine the extent of heterogeneity between the external validation dataset and the source dataset.
In the external validation dataset, we reaffirmed that ICD coding may not be an appropriate surrogate for the Sepsis-3 criteria as the reference standard to identify sepsis [27,48,49]. However, even when the determination of outcomes (i.e., presence of sepsis) was controlled by a standardised review process in this study, our results still show that the performance of the ML model declines when the prevalence of the outcome and the distribution of predictors differed between the two geographically and temporally independent datasets. Patient comorbidities and infection sources, although not predictors, also appeared to interfere with model performance. This finding indicates that the intended population should be specifically defined to ensure the clinical applicability of the model. Correspondingly, a thorough examination of data heterogeneity should be conducted to judge the efficacy of ML on the targeted clinical setting.

4.5. Study Limitations

There are several limitations of this study. First, the development dataset and validation dataset were both derived from a single institute, which suggested that selection bias might exists. However, the in-hospital mortality for patients with sepsis present on admission, using the Sepsis-3 criteria for sepsis identification was similar to the previous epidemiologic result (13.5% in our study vs. 13.4% in previous study) [24]. Second, heterogeneity in study designs has been shown to hinder the comparison of performance among different models [18,27]. Therefore, we could hardly compare the performance of our model with previous models [50,51,52,53] because considerable heterogeneity was found among our study and previous studies, including the patient characteristics (e.g., all ED visits versus only ED admissions, with differences in sepsis prevalence), selecting predictors (e.g., vitals, lab test results versus text data from the EHR), the reference standard for sepsis (ICD coding versus the Sepsis-3 criteria), and diagnostic outcomes (e.g., sepsis versus severe sepsis and septic shock versus mortality). Third, information on patient age, comorbidities, and infectious focus for sepsis was lacking in the development dataset for the comparison of the two datasets. However, the collected vital signs and lab tests results appeared to be sufficient to reflect the discrepancies in clinical manifestations between the two patient cohorts. Finally, because our study results suggested several patient characteristics (such as age, certain comorbidities, and infection focus) might interfere with the model performance, whether the implementation of a ML model can shorten the triage-to-antibiotics time requires future, prospective interventional trials.

5. Conclusions

Using commonly available clinical variables, we developed and externally validated a ML model to effectively identify sepsis patients in the ED. This study demonstrated that the XGBoost model outperformed the pre-existing conventional tools in identifying sepsis patients; however, we also revealed that differences in patient characteristics, while being key predictors, could reduce model performance. This finding reinforces the recommendation of performing external validation to ensure the generalisability of clinical decision support models. Because heterogeneity among patient cohorts seems inevitable, future studies are needed to solve this model adaptation problem.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/jpm11111055/s1, Table S1: TRIPOD checklist.

Author Contributions

Conceptualization, P.-C.L. and M.-C.L.; data curation, P.-C.L. and K.-T.C.; algorithm development, P.-C.L.; writing—original draft preparation, P.-C.L.; writing—review and editing, K.-T.C., H.-C.C., M.M.I. and M.-C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Human Research at both the Chi-Mei Medical Centre and the Taoyuan General Hospital (IRB No: TYGH107014).

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Singer, M.; Deutschman, C.S.; Seymour, C.C.; Shankar-Hari, M.; Annane, D.; Bauer, M.; Bellomo, R.; Bernard, G.R.; Chiche, J.-D.; Coopersmith, C.C.M.; et al. The Third International Consensus Definitions for Sepsis and Septic Shock (Sepsis-3). JAMA 2016, 315, 801–810. [Google Scholar] [CrossRef]
  2. Rodríguez, F.; Barrera, L.; De La Rosa, G.; Dennis, R.J.; Dueñas, C.; Granados, M.; Londoño, D.; Molina, F.; Ortiz, G.; Jaimes, F. The epidemiology of sepsis in Colombia: A prospective multicenter cohort study in ten university hospitals*. Crit. Care Med. 2011, 39, 1675–1682. [Google Scholar] [CrossRef]
  3. Silva, E.; Pedro, M.D.A.; Sogayar, A.C.B.; Mohovic, T.; Silva, C.L.D.O.; Janiszewski, M.; Cal, R.G.R.; de Sousa, E.F.; Abe, T.P.; de Andrade, J.; et al. Brazilian Sepsis Epidemiological Study (BASES study). Crit. Care 2004, 8, R251–R260. [Google Scholar] [CrossRef] [Green Version]
  4. Cheng, B.; Xie, G.; Yao, S.; Wu, X.; Guo, Q.; Gu, M.; Fang, Q.; Xu, Q.; Wang, D.; Jin, Y.; et al. Epidemiology of severe sepsis in critically ill surgical patients in ten university hospitals in China*. Crit. Care Med. 2007, 35, 2538–2546. [Google Scholar] [CrossRef] [PubMed]
  5. Fleischmann, C.; Scherag, A.; Adhikari, N.K. Assessment of global incidence and mortality of hospital-treated sepsis. Current estimates and limitations. Am. J. Respir. Crit. Care Med. 2016, 193, 259–272. [Google Scholar] [CrossRef]
  6. Gaieski, D.F.; Edwards, J.M.; Kallan, M.J.; Carr, B.G. Benchmarking the Incidence and Mortality of Severe Sepsis in the United States*. Crit. Care Med. 2013, 41, 1167–1174. [Google Scholar] [CrossRef] [PubMed]
  7. Torio, C.M.; Moore, B.J. National Inpatient Hospital Costs: The Most Expensive Conditions by Payer, 2013: Statistical Brief# 204; Agency for Healthcare Research and Quality: Rockville, MD, USA, 2016. [Google Scholar]
  8. Alsolamy, S.; Al Salamah, M.; Al Thagafi, M.; Al-Dorzi, H.M.; Marini, A.M.; Aljerian, N.; Al-Enezi, F.; Al-Hunaidi, F.; Mahmoud, A.M.; Alamry, A.; et al. Diagnostic accuracy of a screening electronic alert tool for severe sepsis and septic shock in the emergency department. BMC Med. Inform. Decis. Mak. 2014, 14, 105. [Google Scholar] [CrossRef] [Green Version]
  9. Nguyen, S.Q.; Mwakalindile, E.; Booth, J.S. Automated electronic medical record sepsis detection in the emergency de-partment. Peer J. 2014, 2, e343. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Tsai, C.; Patel, K.; Vincent, A.; Verzosa, N.; Norris, D.; Tillis, W.; Hafner, J. 253 Electronic Best Practice Advisories’ Effectiveness in Detecting Sepsis in the Emergency Department. Ann. Emerg. Med. 2015, 66, S91–S92. [Google Scholar] [CrossRef] [Green Version]
  11. Burdick, H.; Pino, E.; Gabel-Comeau, D.; Gu, C.; Roberts, J.; Le, S.; Slote, J.; Saber, N.; Pellegrini, E.; Green-Saxena, A.; et al. Validation of a machine learning algorithm for early severe sepsis prediction: A retrospective study predicting severe sepsis up to 48 h in advance using a diverse dataset from 461 US hospitals. BMC Med. Inform. Decis. Mak. 2020, 20, 1–10. [Google Scholar] [CrossRef]
  12. Usman, O.A.; Usman, A.; Ward, M.A. Comparison of SIRS, qSOFA, and NEWS for the early identification of sepsis in the Emergency Department. Am. J. Emerg. Med. 2019, 37, 1490–1497. [Google Scholar] [CrossRef] [PubMed]
  13. Petit, J.; Passerieux, J.; Maître, O. Impact of a qSOFA-based triage procedure on antibiotic timing in ED patients with sepsis: A prospective interventional study. Am. J. Emerg. Med. 2020, 38, 477–484. [Google Scholar] [CrossRef] [PubMed]
  14. Jouffroy, R.; Saade, A.; Ellouze, S.; Carpentier, A.; Michaloux, M.; Carli, P.; Vivien, B. Prehospital triage of septic patients at the SAMU regulation: Comparison of qSOFA, MRST, MEWS and PRESEP scores. Am. J. Emerg. Med. 2018, 36, 820–824. [Google Scholar] [CrossRef]
  15. Kalil, A.C.; Machado, F.R. Quick Sequential Organ Failure Assessment Is Not Good for Ruling Sepsis in or out. Chest 2019, 156, 197–199. [Google Scholar] [CrossRef]
  16. Amland, R.C.; Hahn-Cover, K.E. Clinical Decision Support for Early Recognition of Sepsis. Am. J. Med. Qual. 2016, 31, 103–110. [Google Scholar] [CrossRef] [Green Version]
  17. Islam, M.; Nasrin, T.; Walther, B.A.; Wu, C.C.; Yang, H.-C.; Li, Y.-C. Prediction of sepsis patients using machine learning approach: A meta-analysis. Comput. Methods Programs Biomed. 2019, 170, 1–9. [Google Scholar] [CrossRef]
  18. Hwang, M.I.; Bond, W.F.; Powell, E.S. Sepsis Alerts in Emergency Departments: A Systematic Review of Accuracy and Quality Measure Impact. West. J. Emerg. Med. 2020, 21, 1201–1210. [Google Scholar] [CrossRef]
  19. Kam, H.J.; Kim, H.Y. Learning representations for the early detection of sepsis with deep neural networks. Comput. Biol. Med. 2017, 89, 248–255. [Google Scholar] [CrossRef]
  20. Calvert, J.S.; Price, D.A.; Chettipally, U.K.; Barton, C.W.; Feldman, M.D.; Hoffman, J.L.; Jay, M.; Das, R. A computational approach to early sepsis detection. Comput. Biol. Med. 2016, 74, 69–73. [Google Scholar] [CrossRef] [Green Version]
  21. Nemati, S.; Holder, A.; Razmi, F.; Stanley, M.D.; Clifford, G.; Buchman, T.G. An Interpretable Machine Learning Model for Accurate Prediction of Sepsis in the ICU. Crit. Care Med. 2018, 46, 547–553. [Google Scholar] [CrossRef] [PubMed]
  22. Kong, G.; Lin, K.; Hu, Y. Using machine learning methods to predict in-hospital mortality of sepsis patients in the ICU. BMC Med. Inform. Decis. Mak. 2020, 20, 1–10. [Google Scholar] [CrossRef]
  23. Ibrahim, Z.M.; Wu, H.; Hamoud, A.; Stappen, L.; Dobson, R.; Agarossi, A. On classifying sepsis heterogeneity in the ICU: Insight using machine learning. J. Am. Med. Inform. Assoc. 2020, 27, 437–443. [Google Scholar] [CrossRef]
  24. Rhee, C.; Dantes, R.; Epstein, L.; Murphy, D.J.; Seymour, C.W.; Iwashyna, T.J.; Kadri, S.S.; Angus, D.C.; Danner, R.L.; Fiore, A.E.; et al. Incidence and Trends of Sepsis in US Hospitals Using Clinical vs Claims Data, 2009–2014. JAMA 2017, 318, 1241–1249. [Google Scholar] [CrossRef] [PubMed]
  25. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  26. Moons, K.G.; Altman, D.G.; Reitsma, J.B.; Ioannidis, J.P.; Macaskill, P.; Steyerberg, E.W.; Vickers, A.J.; Ransohoff, D.F.; Collins, G.S. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): Explanation and Elaboration. Ann. Intern. Med. 2015, 162, W1–W73. [Google Scholar] [CrossRef] [Green Version]
  27. Fleuren, L.M.; Klausch, T.L.; Zwager, C.L. Machine learning for the prediction of sepsis: A systematic review and me-ta-analysis of diagnostic test accuracy. Intensive Care Med. 2020, 46, 383–400. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Wang, H.E.; Jones, A.R.; Donnelly, J. Revised National Estimates of Emergency Department Visits for Sepsis in the United States*. Crit. Care Med. 2017, 45, 1443–1449. [Google Scholar] [CrossRef] [PubMed]
  29. Collins, G.S.; Reitsma, J.B.; Altman, D.G.; Moons, K.G.M. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): The TRIPOD Statement. BJS 2015, 102, 148–158. [Google Scholar] [CrossRef] [Green Version]
  30. Kumar, A.; Roberts, D.; Wood, K.E. Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic shock. Crit. Care Med. 2006, 34, 1589–1596. [Google Scholar] [CrossRef]
  31. Ferrer, R.; Martin-Loeches, I.; Phillips, G. Empiric antibiotic treatment reduces mortality in severe sepsis and septic shock from the first hour: Results from a guideline-based performance improvement program. Crit. Care Med. 2014, 42, 1749–1755. [Google Scholar] [CrossRef] [PubMed]
  32. Peltan, I.D.; Mitchell, K.H.; Rudd, K.; Mann, B.A.; Carlbom, D.J.; Hough, C.L.; Rea, T.D.; Brown, S. Physician Variation in Time to Antimicrobial Treatment for Septic Patients Presenting to the Emergency Department. Crit. Care Med. 2017, 45, 1011–1018. [Google Scholar] [CrossRef]
  33. Mok, K.; Christian, M.; Nelson, S.; Burry, L. Time to Administration of Antibiotics among Inpatients with Severe Sepsis or Septic Shock. Can. J. Hosp. Pharm. 2014, 67, 213–219. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Appelboam, R.; Tilley, R.; Blackburn, J. Time to antibiotics in sepsis. Crit. Care 2010, 14, P50. [Google Scholar] [CrossRef] [Green Version]
  35. Shimabukuro, D.W.; Barton, C.W.; Feldman, M.D.; Mataraso, S.; Das, R. Effect of a machine learning-based severe sepsis prediction algorithm on patient survival and hospital length of stay: A randomised clinical trial. BMJ Open Respir. Res. 2017, 4, e000234. [Google Scholar] [CrossRef]
  36. Sawyer, A.M.; Deal, E.N.; Labelle, A.J.; Witt, C.; Thiel, S.W.; Heard, K.; Reichley, R.M.; Micek, S.T.; Kollef, M.H. Implementation of a real-time computerized sepsis alert in nonintensive care unit patients*. Crit. Care Med. 2011, 39, 469–473. [Google Scholar] [CrossRef] [Green Version]
  37. Birnbaum, A.; Bijur, P.; Kuperman, G.; Gennis, P.; Berger, T. A Computerized Alert Screening for Severe Sepsis in Emergency Department Patients Increases Lactate Testing but Does Not Improve Inpatient Mortality. Appl. Clin. Inform. 2010, 01, 394–407. [Google Scholar] [CrossRef] [Green Version]
  38. Debray, T.P.; Vergouwe, Y.; Koffijberg, H.; Nieboer, D.; Steyerberg, E.W.; Moons, K.G. A new framework to enhance the interpretation of external validation studies of clinical prediction models. J. Clin. Epidemiol. 2015, 68, 279–289. [Google Scholar] [CrossRef] [Green Version]
  39. Ahmed, I.; Debray, T.P.; Moons, K.G. Developing and validating risk prediction models in an individual participant data meta-analysis. BMC Med. Res. Methodol. 2014, 14, 1–15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Pennells, L.; Kaptoge, S.; White, I.R. Assessing risk prediction models using individual participant data from multiple studies. Am. J. Epidemiol. 2014, 179, 621–632. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Riley, R.D.; Ensor, J.; Snell, K.; Debray, T.; Altman, D.G.; Moons, K.G.M.; Collins, G. External validation of clinical prediction models using big datasets from e-health records or IPD meta-analysis: Opportunities and challenges. BMJ 2016, 353, i3140. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Calvert, J.; Hoffman, J.; Barton, C.; Shimabukuro, D.; Ries, M.; Chettipally, U.; Kerem, Y.; Jay, M.; Mataraso, S.; Das, R. Cost and mortality impact of an algorithm-driven sepsis prediction system. J. Med. Econ. 2017, 20, 646–651. [Google Scholar] [CrossRef]
  43. McCoy, A.; Das, R. Reducing patient mortality, length of stay and readmissions through machine learning-based sepsis pre-diction in the emergency department, intensive care unit and hospital floor units. BMJ Open Qual. 2017, 6, e000158. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Mao, Q.; Jay, M.; Hoffman, J.L.; Calvert, J.; Barton, C.; Shimabukuro, D.; Shieh, L.; Chettipally, U.; Fletcher, G.; Kerem, Y.; et al. Multicentre validation of a sepsis prediction algorithm using only vital sign data in the emergency department, general ward and ICU. BMJ Open 2018, 8, e017833. [Google Scholar] [CrossRef] [Green Version]
  45. Giannini, H.M.; Ginestra, J.C.; Chivers, C. A machine learning algorithm to predict severe sepsis and septic shock: Development, implementation, and impact on clinical practice. Crit. Care Med. 2019, 47, 1485–1492. [Google Scholar] [CrossRef] [PubMed]
  46. Faisal, M.; Scally, A.; Richardson, D.; Beatson, K.; Howes, R.; Speed, K.; Mohammed, M.A. Development and External Validation of an Automated Computer-Aided Risk Score for Predicting Sepsis in Emergency Medical Admissions Using the Patient’s First Electronically Recorded Vital Signs and Blood Test Results*. Crit. Care Med. 2018, 46, 612–618. [Google Scholar] [CrossRef]
  47. Fang, W.; Chen, C.; Song, B.; Wang, L.; Zhou, J.; Zhu, K.Q. Adapted Tree Boosting for Transfer Learning. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 741–750. [Google Scholar]
  48. Fleischmann-Struzek, C.; Thomas-Rüddel, D.O.; Schettler, A.; Schwarzkopf, D.; Stacke, A.; Seymour, C.W.; Haas, C.; Dennler, U.; Reinhart, K. Comparing the validity of different ICD coding abstraction strategies for sepsis case identification in German claims data. PLoS ONE 2018, 13, e0198847. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Heldens, M.; Schout, M.; Hammond, N.E.; Bass, F.; Delaney, A.; Finfer, S.R. Sepsis incidence and mortality are underestimated in Australian intensive care unit administrative data. Med. J. Aust. 2018, 209, 255–260. [Google Scholar] [CrossRef] [PubMed]
  50. Brown, S.M.; Jones, J.; Kuttler, K.G.; Keddington, R.K.; Allen, T.L.; Haug, P. Prospective evaluation of an automated method to identify patients with severe sepsis or septic shock in the emergency department. BMC Emerg. Med. 2016, 16, 1–7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Horng, S.; Sontag, D.A.; Halpern, Y. Creating an automated trigger for sepsis clinical decision support at emergency de-partment triage using machine learning. PLoS ONE 2017, 12, e0174708. [Google Scholar] [CrossRef] [Green Version]
  52. Delahanty, R.J.; Alvarez, J.; Flynn, L.M.; Sherwin, R.L.; Jones, S.S. Development and Evaluation of a Machine Learning Model for the Early Identification of Patients at Risk for Sepsis. Ann. Emerg. Med. 2019, 73, 334–344. [Google Scholar] [CrossRef]
  53. Rico, P.M.; Pérez, A.V.; Martínez, J.M.L. Electronic alerting and decision support for early sepsis detection and management: Impact on clinical outcomes. Eur. J. Clin. Pharm. Atención Farm. 2017, 19, 33–40. [Google Scholar]
Figure 1. A general outline of the step-by-step approach to identify sepsis patients and compare the identification with traditional tools.
Figure 1. A general outline of the step-by-step approach to identify sepsis patients and compare the identification with traditional tools.
Jpm 11 01055 g001
Figure 2. The area under the receiver operating curve comparison of XGBoost model with SIRS and qSOFA; (A) internal validation and (B) external validation.
Figure 2. The area under the receiver operating curve comparison of XGBoost model with SIRS and qSOFA; (A) internal validation and (B) external validation.
Jpm 11 01055 g002
Figure 3. The upper bar plot shows the 15 top ranking features of the machine learning model. The lower table is the statistical analysis of these features between septic patients in the development dataset and septic patients in the external validation dataset. Most of the important predictors had a heterogeneous distribution between the two datasets. (Lympho = percentage of lymphocytes in the differential of the complete blood count).
Figure 3. The upper bar plot shows the 15 top ranking features of the machine learning model. The lower table is the statistical analysis of these features between septic patients in the development dataset and septic patients in the external validation dataset. Most of the important predictors had a heterogeneous distribution between the two datasets. (Lympho = percentage of lymphocytes in the differential of the complete blood count).
Jpm 11 01055 g003
Table 1. Sepsis-3 (revised) definition for sepsis compared with the traditional sepsis definition.
Table 1. Sepsis-3 (revised) definition for sepsis compared with the traditional sepsis definition.
Traditional DefinitionSepsis-3 Definition
Sepsis Suspicious/known infection + ≥2 SIRSSuspicious/known infection + rise in SOFA score ≥2
Severe sepsis Sepsis
+
SBP < 90 mmHg or MAP < 65 mmHg, lactate > 2 mmol/L (18 mg/dL)
INR > 1.5 or a PTT > 60 s
Bilirubin > 34 μmol/L
Urine output < 0.5 mL/kg/h for 2 h
Creatinine > 177 μmol/L
Platelets < 100 × 109/L
SpO2 < 90%on room air
Not a category
Septic ShockSepsis
+
hypotension
after adequate fluid resuscitation
Sepsis
+
Vasopressors needed for MAP >65 mmHg
+
Lactate >2 mmol/L
after adequate fluid resuscitation
Table 2. Description of the development dataset and the external validation dataset.
Table 2. Description of the development dataset and the external validation dataset.
Development DatasetValidation Dataset
Case number82961744
Geographical regionSouthern TaiwanNorthern Taiwan
Data collection period1 July 2016 to
31 October 2016
1 January 2018 to
31 March 2018
Study designRetrospectiveRetrospective
SettingA tertiary teaching hospitalA regional hospital
Inclusion criteriaAll the adult ED visits (≥20 years old) admitted as inpatient without further transferring during the whole hospitalisation
Reference standard for sepsisSepsis-3 definition
Prevalence of sepsis21%29%
Mortality for sepsis13.5%17%
Meanlength of stay (days)9.86.3
Model predictorsmeans.d.means.d.
Average of SIRS 1.221.021.711.04
Average of qSOFA0.330.620.570.76
Vital signs at triage
SBP 142.233.21132.536.94
DBP 84.5717.9476.5223.62
RR 18.243.4820.303.83
GCS 14.272.2313.652.99
BT 36.930.9636.463.42
HR 92.8021.7493.4425.54
Initial lab results at ED
WBC10.946.1411.455.74
Segment76.3015.7776.8413.07
Band0.452.280.151.26
Eosinophil1.382.381.211.95
Basophil0.340.360.340.41
Lymphocyte16.4511.1415.3211.07
Platelet232.696.85247.6121.9
Haemoglobin12.522.8712.042.91
Haematocrit36.777.5337.018.38
MCH29.453.329.243.46
MCHC33.161.6432.341.74
MCV88.768.4590.258.87
RBC4.160.904.1351.00
RDW14.242.3911.292.34
Lactate3.022.803.154.01
CRP60.0874.6567.6666.87
Glucose159.2104.2171.6130.8
Na142.329.381376.50
K4.081.374.140.73
BUN30.1727.1029.8027.36
Cr1.391.801.672.07
GOT87.2918761.68225.7
GPT42.1114.242.91104.4
T.bil2.954.981.431.91
hsTnI1043.821569.8811.820073.5
CK-MB 8.2432.536.1022.46
Note: MCH = mean corpuscular haemoglobin; MCHC = mean corpuscular haemoglobin concentration; MCV = mean corpuscular volume; RBC = red blood cell; RDW= red cell distribution width; SBP = systolic blood pressure; DBP = diastolic blood pressure; RR = respiratory rate; GCS = Glasgow Coma Scale; BT = blood temperature; HR = heart rate; CRP = C-reactive protein; Na = sodium; K = potassium; BUN = blood urea nitrogen; Cr = creatinine; GOT= aspartate aminotransferase; T.bil = bilirubin test; hsTnI = high sensitivity troponin; CK-MB = creatine kinase-MB.
Table 3. Diagnostic performance of the identification of sepsis.
Table 3. Diagnostic performance of the identification of sepsis.
Model PerformanceXGBoostSIRSqSOFA
Internal validation
Accuracy0.780.690.79
Sensitivity0.800.640.35
Specificity0.780.660.96
PPV0.470.370.53
NPV0.940.880.81
External validation
Accuracy0.700.340.75
Sensitivity0.670.660.36
Specificity0.700.470.89
PPV0.480.340.76
NPV0.840.770.79
Note: PPV = positive predictive value, NPV = negative predictive value.
Table 4. Statistical analysis of comorbidities and infection source among patients in the external validation dataset, divided according to the model output.
Table 4. Statistical analysis of comorbidities and infection source among patients in the external validation dataset, divided according to the model output.
TPFNp-ValueTNFPp-Value
Age (years)74.467.8<0.00161.859.90.1063
Presence of comorbidity (%)
Diabetes mellitus40.1836.360.409727.7529.230.5967
Hypertension49.5649.090.921241.6336.890.1204
Coronary artery disease14.3713.330.753115.69.020.0021
Chronic kidney disease12.029.70.43788.65.190.0388
End-stage renal disease7.625.450.36714.935.740.5586
Cerebrovascular accident19.3516.360.41545.057.380.1074
Congestive heart failure9.097.270.49175.163.550.2216
Malignancy11.1410.910.93178.3711.20.116
Presence of infection focus (%)
Urinary tract infection28.4520.610.05916.6519.13<0.0001
Cellulitis3.523.030.77496.777.10.8302
Pneumonia24.058.48<0.00014.9324.32<0.0001
Intra-abdominal infection2.936.060.09056.316.280.9878
Note: TP: True positive; FN: False negative; TN: True negative; FP: False positive.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, P.-C.; Chen, K.-T.; Chen, H.-C.; Islam, M.M.; Lin, M.-C. Machine Learning Model to Identify Sepsis Patients in the Emergency Department: Algorithm Development and Validation. J. Pers. Med. 2021, 11, 1055. https://doi.org/10.3390/jpm11111055

AMA Style

Lin P-C, Chen K-T, Chen H-C, Islam MM, Lin M-C. Machine Learning Model to Identify Sepsis Patients in the Emergency Department: Algorithm Development and Validation. Journal of Personalized Medicine. 2021; 11(11):1055. https://doi.org/10.3390/jpm11111055

Chicago/Turabian Style

Lin, Pei-Chen, Kuo-Tai Chen, Huan-Chieh Chen, Md. Mohaimenul Islam, and Ming-Chin Lin. 2021. "Machine Learning Model to Identify Sepsis Patients in the Emergency Department: Algorithm Development and Validation" Journal of Personalized Medicine 11, no. 11: 1055. https://doi.org/10.3390/jpm11111055

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop