Next Article in Journal
Symptomatic Intraosseous Lipoma of the Calcaneum
Previous Article in Journal
Importance of Visual Estimation of Coronary Artery Stenoses and Use of Functional Evaluation for Appropriate Guidance of Coronary Revascularization—Multiple Operator Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predicting Prolonged Length of ICU Stay through Machine Learning

1
National Institute of Health Data Science, Peking University, Beijing 100191, China
2
Advanced Institute of Information Technology, Peking University, Hangzhou 311215, China
3
Department of Medicine and Therapeutics, LKS Institute of Health Science, The Chinese University of Hong Kong, Hong Kong, China
4
Department of Epidemiology and Biostatistics, School of Public Health, Peking University, Beijing 100191, China
5
Medical Informatics Center, Peking University, Beijing 100191, China
6
Renal Division, Department of Medicine, Peking University First Hospital, Peking University Institute of Nephrology, Beijing 100034, China
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(12), 2242; https://doi.org/10.3390/diagnostics11122242
Submission received: 15 November 2021 / Revised: 22 November 2021 / Accepted: 24 November 2021 / Published: 30 November 2021
(This article belongs to the Topic Artificial Intelligence in Healthcare)

Abstract

:
This study aimed to construct machine learning (ML) models for predicting prolonged length of stay (pLOS) in intensive care units (ICU) among general ICU patients. A multicenter database called eICU (Collaborative Research Database) was used for model derivation and internal validation, and the Medical Information Mart for Intensive Care (MIMIC) III database was used for external validation. We used four different ML methods (random forest, support vector machine, deep learning, and gradient boosting decision tree (GBDT)) to develop prediction models. The prediction performance of the four models were compared with the customized simplified acute physiology score (SAPS) II. The area under the receiver operation characteristic curve (AUROC), area under the precision-recall curve (AUPRC), estimated calibration index (ECI), and Brier score were used to measure performance. In internal validation, the GBDT model achieved the best overall performance (Brier score, 0.164), discrimination (AUROC, 0.742; AUPRC, 0.537), and calibration (ECI, 8.224). In external validation, the GBDT model also achieved the best overall performance (Brier score, 0.166), discrimination (AUROC, 0.747; AUPRC, 0.536), and calibration (ECI, 8.294). External validation showed that the calibration curve of the GBDT model was an optimal fit, and four ML models outperformed the customized SAPS II model. The GBDT-based pLOS-ICU prediction model had the best prediction performance among the five models on both internal and external datasets. Furthermore, it has the potential to assist ICU physicians to identify patients with pLOS-ICU risk and provide appropriate clinical interventions to improve patient outcomes.

1. Introduction

Intensive care units (ICU) provide complex and resource-intensive treatment for the sickest hospitalized patients. The need for critical care medicine has grown substantially over the past decade [1] and has consumed a huge portion of the income in many countries worldwide [2]. In the US, critical care medicine costs account for approximately 13% of hospital costs and 4% of national health expenditures [3]. Despite the huge investment in critical care medicine, medical resources in ICU are usually insufficient to meet the demands of ICU patients, especially in developing countries. Hospitals are under pressure to improve the efficiency and reduce costs for critical care. Length of stay in ICU (LOS-ICU) is a key indicator for medical efficiency [4] and critical care quality in hospitals [5]; a prolonged LOS-ICU (pLOS-ICU) generally leads to additional use of resources and thus increased medical costs [6,7]. A small percentage of patients with pLOS-ICU could consume a large proportion (nearly 50%) of ICU resource use [8,9]. The early identification of pLOS-ICU risk for ICU patients can provide not only an important reference for patient and family counseling but also an important indicator for optimal clinical interventions. However, ICU physicians can hardly accurately predict pLOS-ICU at ICU admission [10]. Effective pLOS-ICU prediction tools are strongly needed by the ICU physicians [4].
The present study sought to construct machine learning (ML) based models to perform pLOS-ICU prediction for general ICU patients. Four different ML methods, namely random forest (RF), support vector machine (SVM), deep learning (DL), and gradient boosting decision tree (GBDT), were used for prediction model development in this study. The reasons that we employed the above four ML methods are as follows. Firstly, SVM was a frequently used single ML method to deal with complex ICU data, and it showed robust performance in handling noisy and nonlinearly classified data [11,12]. Secondly, the ensemble learning method could combine multiple ML models to achieve better performance and generalizability than a single one [13], while RF and GBDT are typical ensemble learning models with different ensemble mechanisms. Thirdly, the emerging ML algorithm, DL, also showed good performance in supporting clinical decision making in ICU [14,15]. Therefore, the above four different ML methods were used for model development. Their prediction performance was compared with the customized SAPS II, which was used as benchmark model.
This article is structured as follows. Section 1 introduces the background and purpose of this study. Section 2 carries out a review of related work in the literature. Section 3 describes the datasets, study subjects, model development and evaluation methods used in this study. Section 4 shows various results generated in this study, and the results are then further discussed in Section 5. Finally the conclusions of this study are summarized in Section 6.

2. Related Work

In the literature, some pLOS-ICU prediction models have been developed. Among them, some were customized from traditional severity scoring systems, such as the simplified acute physiology score (SAPS) II [16], acute physiology and chronic health evaluation (APACHE) III score [17], and the sequential organ failure assessment (SOFA) score [18]. Alternatively, some were developed via the logistic regression (LR) method. For example, Zoller et al. [16] customized the SAPS II to predict pLOS-ICU based on a prospective single-center dataset and found that the customized SAPS II showed limited accuracy and utility. Houthooft et al. [18] customized the SOFA score to predict pLOS-ICU using a single-center dataset in Belgium, and the corresponding sensitivity was low. Herman et al. [6] used the LR method to develop a pLOS-ICU prediction model for patients undergoing isolated coronary artery bypass grafting (CABG), but the model was constructed using a single-center dataset with a small sample size. Rotar et al. [19] also developed a pLOS-ICU prediction model for patients following CABG surgery based on least absolute shrinkage and selection operator (Lasso) algorithm. Their model was internally tested with an area under the receiver operation characteristic curve (AUROC) of 0.72. However, no external validation was performed for all the above-mentioned pLOS-ICU prediction studies.
In recent years, ML algorithms have been used to develop prediction models in complex clinical contexts, such as ICU [20,21,22,23]. Meiring et al. [24] used four ML methods to predict ICU mortality and compared their prediction performance with APACHE-II and the traditional LR model. They found that all the four ML models outperformed APACHE-II and LR. Lin et al. [20] constructed three ML models, namely, artificial neural network, SVM, and RF, together with a customized SAPS II model, to predict the mortality of acute kidney injury patients in ICU. They found that RF model outperformed the three ML models, and all ML models performed much better than the customized SAPS II model. DL as an emerging ML algorithm in the past decade, and it has been broadly studied in healthcare studies to support clinical decision-making [14,15]. Viton et al. [25] utilized a DL model based on convolutional neural networks (CNN) to predict the risk of in hospital mortality using the real time series records in ICU, and their experimental results showed a competitive accuracy. Despite its accuracy, as a black-box model, DL is hard to explain and thus has difficulties in gaining clinicians’ acceptance. Qian et al. [13] performed comparison study of six ML models for early prediction of acute kidney injury in ICU, and their results showed that the GBDT model outperformed both LR and CNN models.
However, limited studies have used ML methods to develop models for predicting the pLOS-ICU among general ICU patients. Navaz et al. [26] developed a decision tree-based pLOS-ICU prediction model, and the results showed that the prediction performance of the decision tree model is poor. Rocheteau et al. [27] developed a DL model by combining temporal convolution and pointwise convolution to predict numeric LOS-ICU, but their model showed a limited performance with a R2 of 0.40 in external validation. Ma et al. [28] constructed a model to predict whether an ICU patient can be discharged after 10 days by combining just-in-time learning and one-class extreme learning, but the dataset used for model training and test was a small sample with only 4000 records and no external validation was conducted. A rough comparison of the related work in the literature with our proposed model is shown in Table 1.
The existing pLOS-ICU prediction models for general ICU patients are not effective enough as required by ICU clinicians [18,26,27,29]. In addition, most of the existing pLOS-ICU prediction models have not been externally validated [4,19,28,30], and the generalizability of these models is unknown. Therefore, more accurate prediction models with external validation are needed for pLOS-ICU prediction. A well-developed pLOS-ICU prediction model has the potential to assist ICU physicians to identify patients at a high risk of prolonged ICU stay, and thus may help improve clinical decision making and family counseling quality.

3. Materials and Methods

3.1. Datasets and Study Subjects

A publicly available critical care database, eICU Collaborative Research Database (eICU-CRD) [31], was used for model development and internal validation. The eICU-CRD database is a multicenter database that is maintained by the Laboratory for Computational Physiology (LCP) at the Massachusetts Institute of Technology (MIT), which has partnered with the eICU Research Institute. The database contains medical records of 200,859 admissions for 139,367 patients admitted to 335 units in 208 hospitals from 2014 to 2015 across the United States. Data include vital sign measurements, laboratory tests, care plan documentation, diagnosis information, treatment information, and others. All protected health information was deidentified, and no patient privacy data can be identified.
Another publicly available critical care database, the Medical Information Mart for Intensive Care (MIMIC) III [32], was used for external validation of the developed models. The MIMIC-III database, which is also maintained by LCP at MIT, is a single-center database. MIMIC-III contains 53,423 medical records of 38,597 adult patients admitted to critical care units at the Beth Israel Deaconess Medical Center (BIDMC) in Boston, Massachusetts between 2001 and 2012. MIMIC-III is also deidentified; variables and data types in MIMIC-III are similar to eICU-CRD. The source hospital of MIMIC-III does not participate in the eICU program. Thus, the MIMIC-III database is a completely independent dataset. A brief comparison of eICU-CRD and MIMIC-III is shown in Table 2.
In this study, all ICU records of patients between 18 and 90 years old were extracted from the eICU-CRD. Patients were excluded if they met the following criteria: variable missing rate was larger than 30%; or LOS-ICU was missing or an outlier, defined as a LOS-ICU above the 99th percentile of LOS-ICU in the studied dataset [33,34]. In addition, patients who died within ICU were also excluded, as the LOS-ICU pattern of patients who died within ICU may be different from the patients who survived in the ICU [7]. For patients with twice or more ICU admissions during one hospitalization in eICU-CRD, we randomly selected one record for the corresponding patient to ensure that all observations were independent in model development. Compared with the approach of selecting the first admission record for a patient having multiple ICU admissions during a hospitalization, randomly selecting one ICU record for the patient may help include patients with varying severities [35]. In MIMIC-III, patient data were extracted following the same inclusion and exclusion criteria. The difference is that we kept all the ICU records for patients with multiple ICU admissions during a hospitalization in MIMIC-III for external validation. The flowchart of the process for the patients’ inclusion is shown in Figure 1.

3.2. Primary Outcome and Predictor Variables

The primary outcome of this study was having pLOS-ICU. pLOS-ICU is defined as a length of stay longer than the reported average LOS-ICU [16,19,36], which is three days for general ICU patients in the United States [37].
To compare the prediction performance of four ML-based models with the customized SAPS II in an objective way, we used all predictor variables in the customized SAPS II to construct the SVM, RF, DL, and GBDT models. A total of 17 variables are used in the SAPS II scoring system, consisting of age, chronic diseases (metastatic cancer, acquired immunodeficiency syndrome, and hematologic malignancy), type of admission (scheduled surgical, unscheduled surgical, or medical), and 12 routine physiological measurements during the first 24 h in ICU [38]. The 12 physiological measurements include body temperature, heart rate, Pao2/Fio2 ratio, systolic blood pressure, urinary output, white blood cell count, serum urea nitrogen level, serum sodium level, serum potassium level, bilirubin level, serum bicarbonate level, and Glasgow Coma Score. Some of the physiological variables are time-stamped variables, which have multiple time-variant measurements. Similar to the original SAPS II [38], the time-stamped variables in the customized SAPS II, were scored on the basis of the worst value during the first 24 h, either minimum or maximum. In the SVM, RF, DL, and GBDT models, we used both the minimum and maximum values during the first 24 h of time-stamped variables as parallel inputs since the minimum and maximum values may reflect two different extreme physical conditions of one patient [20]. The chronic diseases were identified using International Classification of Diseases codes. As to the ICU admission type, its variable and corresponding options recorded in the two databases are different. By referring to the original SAPS-II [38], we used three different admission types: scheduled surgical, unscheduled surgical, and medical, to categorize the ICU admission type. In eICU-CRD, a patient would be identified as: (1) scheduled surgical if the ADMISSION_TYPE is ELECTIVE and he or she had surgery during the stay; (2) unscheduled surgical if the ADMISSION_TYPE is not ELECTIVE and he or she had surgery during the stay; and (3) medical if the ADMISSION_TYPE is none of the above. In MIMIC-III, a patient would be identified as: (1) scheduled surgical if variable electiveSurgery equals 1; (2) unscheduled surgical if variable electiveSurgery equals 0; and (3) medical if variable electiveSurgery is blank. The distributions of admission type in the eICU-CRD and MIMIC-III databases were similar.

3.3. Model Development

3.3.1. Support Vector Machine (SVM)

SVM is a supervised ML algorithm that attempts to find an optimal separating hyperplane in the feature space for classification [39]. SVM has good prediction performance on either linearly or nonlinearly separable datasets, especially on the latter [40]. An SVM model can transform a nonlinearly separable dataset from the original feature space to a high-dimensional space and find a maximum-margin hyperplane to make classifications. Suppose a nonlinearly separable dataset D with N labeled cases is available; D = { ( x 1 ,   y 1 ) , ( x 2 ,   y 2 ) , ,   ( x i , y i ) ,   , ( x N , y N ) } , where y i { 1 ,   1 } . If ϕ ( x ) is the function for transforming dataset D to a high-dimensional space, then the classification hyperplane in the high-dimensional space can be defined if it satisfies the following equation:
w T ϕ ( x ) + b = 0 ,
where w and b are parameter vectors, w is a normal vector determining the direction, and b is the bias. The margin r i between a case ( x i , y i )   and the hyperplane in the high-dimensional space is defined as follows:
r i = | w T ϕ ( x i ) + b | w .
The cases nearest to the hyperplane are called support vectors, which satisfy
| w T ϕ ( x ) + b | = 1 .
The margin R between the support vector and the hyperplane is
R = 1 w .
The hyperplane that makes the margin R maximum is the optimal separating hyperplane (i.e., maximum-margin hyperplane). If we use w ^ and b ^ to denote the parameter vectors of the optimal hyperplane, then the optimal hyperplane can be expressed using the following equation:
w ^ T ϕ ( x ) + b ^ = 0 .
In the process of finding the optimal separating hyperplane, a kernel function is usually used to deal with the high computational cost. Commonly used kernel functions include the polynomial, linear, exponential and radial basis function kernels. A new instance x n e w is then classified by the trained SVM model with an optimal separating hyperplane as follows:
y n e w = {       1 ,                   i f   f ( x n e w ) > 0 1 ,                   i f   f ( x n e w ) < 0 .  
We used the sklearn.svm package in Python to construct the SVM model [41], and a set of optimal parameters of the SVM model were found using grid search, which is an exhaustive searching method that uses a manually specified subset of hyperparameter space to find the optimal parameters of a learning algorithm [42]. The SVM model obtained in this study had the following parameters: the kernel function was a radial basis function kernel. Gamma in the kernel function was 0.04, and the penalty parameter C was 1.

3.3.2. Random Forest (RF)

RF is an ensemble learning model consisting of a multitude of decision trees [43]. Compared with single basic classifiers, ensemble learning models can combine the outputs of multiple basic classifiers and achieve an improved prediction performance [44]. An RF model is trained as follows: First, K (a tunable parameter, K = 100 in our study) subsets of a training dataset D, { D 1 ,   D 2 ,   ,   D K } are generated using the bootstrap sampling method. The sampling proportion is 1 ( 1 1 N ) N , where N is the total number of cases in the entire training dataset. Second, K decision trees { T 1 ,   T 2 ,   ,   T K } are generated from the K subsets of training dataset separately. In decision tree induction, a total of M predictor variables is assumed, and F (a tunable parameter with F < M) out of M variables would be randomly selected for each node splitting based on the minimum impurity principle. Gini index is an indicator to measure information impurity, and it is frequently used in decision tree training [20]. For each tree, a variable or feature should not be used for node splitting any more if it has already been used for previous node splitting. For a dataset D containing samples with J classes, the Gini index of D is defined as follows [45]:
Gini ( D ) = 1 j = 1 J p j 2 ,
where p j is the frequency of the jth class in the dataset D. If a dataset D can be split into two subsets D1 and D2 by the variable V, then the decrease in Gini index S caused by this variable V is
S ( V ) = Gini ( D ) Gini ( D 1 ) Gini ( D 2 ) .
The variable with a maximum decrease in Gini index is then used for node splitting in a decision tree growth. After all the K trees have been generated, an RF model forms. In a RF-based inference or classification, the clinical data of a new patient are inputs to the model, the outputs of all the K decision trees are aggregated through a voting algorithm, and then the majority vote is declared as the final classification.
RF model training has two types of randomizations: randomization of training datasets and randomization of feature subsets in its basic decision tree growth. This classification helps reduce the scale and dimension of the training dataset in generating decision trees. These two randomizations enable an RF model to deal with high-dimensional and large-scale data.
We used the sklearn.RandomForestClassifier package in Python to construct the RF model in this study [41]. A set of optimal parameters of the RF model were found using grid search. The RF model obtained in this study had the following parameters: the number of decision trees K was 100; the number of variables selected for each node splitting F was the square root of the number of input variables, sqrt(M); and the minimum number of samples required to split an internal node was 2.

3.3.3. Gradient Boosting Decision Tree (GBDT)

GBDT is also a kind of ensemble learning model that uses decision trees as the basic classifier [46]. In contrast to parallel decision trees in an RF model, decision trees in a GBDT model are serially generated. A decision tree in a GBDT model is trained based on the bias of all the previous decision trees in the model. In its inference process, a GBDT model synthesizes outputs of the serial decision trees through an addition algorithm to make classifications.
The training of a GBDT model proceeds as follows: suppose a training dataset D with N labeled cases is available; D = { ( x 1 ,   y 1 ) , ( x 2 ,   y 2 ) , ,   ( x i , y i ) ,   , ( x N , y N ) } , where y i { 1 ,   1 } . The decision trees are generated iteratively in a GBDT model training, and each tree in the model is trained based on the bias between the observed outcomes and the predicted probabilities generated by all its previous trees. Therefore, we need to set an initial predicted probability f 0 ( x i ) for a training case ( x i , y i ) before the generation of the first decision tree in a GBDT model. The initial predicted probability of the case ( x i , y i ) is defined as
f 0 ( x i ) = 1 2 l n P ( y = 1 | x ) P ( y = 1 | x ) ,
where P ( y = 1 | x ) is the frequency of class y = 1 in the dataset D, and P ( y = 1 | x ) is the frequency of class y = 1 in the dataset D. The bias between the probability f 0 ( x i ) and y i , i.e., the observed outcome of x i , is defined as residual r 0 , i , which is calculated using the following equation:
r 0 , i = 2 y i 1 + e x p ( 2 y i f 0 ( x i ) )   .
In the first round of iteration, the training dataset for the first tree training is constructed as D 1 = { ( x 1 ,   r 0 , 1 ) , ( x 2 ,   r 0 , 2 ) , ,   ( x i , r 0 , i ) ,   , ( x N , r 0 , N ) } and a decision tree stops growing when no more decrease occurs in its prediction error or it has achieved a preset threshold of max depth. Then, the generated decision tree can be used to predict the bias between the initial probability and the actual outcome for all cases and obtain results T 1 = { t 1 , 1 ,   t 1 , 2 ,   , t 1 , i ,   ,   t 1 , N } Afterward, the predicted probability for x i , f 1 ( x i ) based on the first decision tree can be calculated as
f 1 ( x i ) = f 0 ( x i ) + t 1 , i   .
After prediction results have been generated using the first decision tree, a second residual r 1 , i between the predicted probability and the actual outcome can be generated, and then a second training dataset D 2 = { ( x 1 ,   r 1 , 1 ) , ( x 2 ,   r 1 , 2 ) , ,   ( x i , r 1 , i ) ,   , ( x N , r 1 , N ) } can be constructed for the second decision tree training.
Similarly, in the kth round of iteration, the kth ( k = 1 ,   2 ,     ,   K ; K is the total number of decision trees and is a tunable parameter; K = 100 in our study) decision tree is generated based on the bias of the previous (k − 1) decision trees.
r k 1 , i = 2 y i 1 + e x p ( 2 y i f k 1 ( x i ) ) .
Then, the training dataset for the kth tree D k = { ( x 1 ,   r k 1 , 1 ) , ( x 2 ,   r k 1 , 2 ) , ,   ( x i , r k 1 , i ) ,   , ( x N , r k 1 , N ) } can be constructed. The kth generated tree predicts the bias for all cases and obtains T k = { t k , 1 ,   t k , 2 ,   , t k , i ,   ,   t k , N } . The predicted probability for ( x i , y i ) after the kth round of iteration f k ( x i ) is calculated as
f k ( x i ) = f k 1 ( x i ) + t k , i .
After all the K decision trees have been generated using a forward stage-wise algorithm, a GBDT model forms. For a trained GBDT model, when a new instance x n e w is inputted into the model, each tree makes a prediction about the bias between the actual outcome and the predicted probability, given by all its previous trees for the new instance, and obtains { t 1 , n e w ,   t 2 , n e w ,   ,   t k , n e w ,   ,   t K , n e w   } . Then, the model combines the predictions of all K decision trees using an addition algorithm, as follows:
F ( x n e w ) = f 0 ( x ) + k = 1 K t k , n e w .
The final predicted probability for x n e w generated by the GBDT model is
P n e w = 1 1 + e x p ( 2 F ( x n e w ) ) .
In addition, a GBDT model can provide a variable importance ranking based on the variable importance weight generated by all its basic decision trees.
We used the sklearn.GradientBoostingClassifier package in Python to construct the GBDT model in this study [41]. A set of optimal parameters of the GBDT model were found using grid search. The GBDT model obtained in this study had the following parameters: the fraction of samples used for each decision tree was 0.7, the number of decision trees K was 100, the number of variables selected for each node splitting was the square root of the number of input variables, the maximal depth of each decision tree was 6, and the minimum number of samples required to split an internal node was 200.

3.3.4. Deep Learning (DL)

DL models [47] were developed from artificial neural networks. A DL model was constructed with a greedy layer-by-layer method, where multiple layers were used to progressively extract higher-level features from the raw input and pick out features with high predictive value. DL models have been broadly used in healthcare studies to support clinical decision-making, such as diagnosis [48], prognosis prediction [49], and resource allocation [50].
General architecture of a DL model consists of an input layer, multiple hidden layers, and an output layer. Each layer contains a set of neurons, and is fully connected with its adjacent layers. A neuron receives a signal, processes it, and then signals neurons connected to it. Signals travel from the first (input), to the last (output) layer. In the input layer, the number of neurons was determined by the number of input features. Each neuron in the input layer represents an input feature. In the hidden layer, neurons transform the signals from the input layer (or the previous hidden layer) with a weighted summation followed by a non-linear activation function. Starting from initial random weights, the weights between neurons were repeatedly updated using optimization algorithm to minimize the loss function. The process of weight training stops when it reaches a preset maximum number of iterations, or when the improvement in loss is below a certain number. At last, the output layer receives values from the last hidden layer and transforms them into outcome values.
In this study, a multilayer perceptron algorithm (MLP) was employed to construct a DL model. MLP is the most typical DL algorithm. Compared to complex DL architecture such as CNN, MLP has a relatively small number of parameters and is less complex. Prediction models based on MLP is expected to be more acceptable in clinical practice than CNN [51]. The sklearn.MLPClassifier package in Python was used for model development [41]. A set of optimal parameters of the DL model were found using grid search. The DL model obtained in this study had the following parameters: the optimization algorithm was Limited-memory BFGS (L-BFGS); the number of hidden layers was 2, the number of neurons in each hidden layer was (100, 100); and the penalty parameter was 10.

3.3.5. Customized SAPS II

The SAPS II model is a commonly used scoring tool in ICU to assess the severity of illness, and it is frequently used as a benchmark model for performance comparison in prognosis prediction model development [20,52,53]. SAPS II was developed by LeGall et al. [38] in 1993 on the basis of the clinical dataset of 12,997 ICU patients. As the original SAPS II tool was developed from a clinical dataset of 1993 via traditional logistic regression, clinicians and researchers usually use newly collected dataset to customize the coefficients for target population [54,55]. In the literature, some studies have carried out the work of customizing SAPS II for pLOS-ICU prediction [16,18]. By referring to those studies, we developed a customized SAPS II model for pLOS-ICU prediction using the eICU-CRD database, and then used the customized SAPS II as the benchmark to evaluate the prediction performance of ML models developed in this study.
The algorithm behind the SAPS II model is the LR algorithm. A total of 17 predictor variables are used in the SAPS II scoring system, and each variable is assigned a different score between 0 and 26 according to each patient’s condition. The coefficient of each variable obtained from a multivariate LR analysis, is used as a criterion for assigning a score to the variable. Then, the assigned score can be used to rank variable importance. In SAPS II, the in-hospital mortality probability of an ICU patient can be calculated based on the overall score using the following formula:
z = β 0 + β 1 × S c o r e + β 2 × ln ( S c o r e + 1 ) ,
P m o r = 1 1 + e z ,
where Score is the overall SAPS II score of a specific ICU patient; β 0 , β 1 , and β 2 are the coefficients generated via the LR algorithm; and P m o r is the mortality probability of the ICU patient.
We developed a customized SAPS II model for pLOS-ICU prediction using the eICU-CRD database. In model customization, the coefficients β 0 , β 1 , and β 2 were re-estimated based on the eICU-CRD training dataset, and a new set of coefficients β 0 ,   β 1 ,   and   β 2 were generated. In our customized SAPS II model, the probability of pLOS-ICU for a patient P p L O S I C U can be calculated as follows:
z n e w = β 0 + β 1 × S c o r e + β 2 × ln ( S c o r e + 1 ) ,
P I C U L O S = 1 1 + e z n e w .
The customized SAPS II model used in this study is as follows.
z n e w = 1.88 + 0.06 × S c o r e 0.27 × ln ( S c o r e + 1 ) ,
P I C U L O S = 1 1 + e z n e w .

3.4. Model Validation

We randomly split the eICU-CRD dataset into two parts: 70% as training dataset and 30% as test dataset for internal validation. The entire MIMIC-III dataset was used for external validation. In model training, we used five-fold cross validation to find optimal parameters for the four ML-based models.

3.5. Analysis

We used PostgreSQL 10.5 (The PostgreSQL Global Development Group, Berkeley, California, United States) to extract data from the eICU-CRD and MIMIC-III databases. In the two extracted datasets, the missing value of each predictor variable was filled up using the median value after excluding unqualified patient records. Descriptive data are presented either as mean ± standard deviation or actual numbers (percentages). The prediction performance of the five models were measured using AUROC [56], area under the precision-recall curve (AUPRC) [25], estimated calibration index (ECI) [57], and Brier score [56]. AUROC measures the discrimination power of a prediction model, representing the ability of distinguishing between the positive and negative samples. A high AUROC value represents a strong discrimination power. AUPRC also measures the discrimination power of a model, while AUPRC pays more attention to the ability of identifying positive samples. Compared with AUROC, AUPRC is more sensitive to data imbalance. It should be noted that the baseline value of AUPRC for a model is equal to the fraction of positives in a classification task [58]. This means that the higher the AUPRC is (compared to the fraction of positives), the better performance a model can achieve. In this study, the fraction of positives (pLOS-ICU) is around 25%, and thus the baseline AUPRC is 0.25. Therefore, obtaining an AUPRC of more than 0.50 means a good pLOS-ICU prediction. ECI measures the calibration power of a model, representing the average difference between the predicted probability and the observed probability of each ICU patient. A low ECI suggests a strong calibration power. The Brier score is an overall performance measure, and a low Brier score suggests a superior overall performance. Another issue worth mentioning here is that the training dataset is imbalanced as patients with pLOS-ICU are the minority in eICU-CRD. To remove the effect that an imbalanced dataset may have on trained prediction models, a comprehensive performance measure considering both sensitivity and specificity instead of just prediction accuracy should be used as criteria for model evaluation. The AUROC value meets this need. A calibration plot was used to illustrate the calibration power of a model visually. The ideal calibration curve for a perfect model is a diagonal, which indicates that the predicted probabilities are consistent with the observed probabilities. The model with the best prediction performance was used to generate variable importance ranking, and the top five important predictor variables were presented.

4. Results

Overall, 117,306 ICU patients in eICU-CRD and 42,932 ICU patients in MIMIC-III were included for model derivation and validation. The characteristics of the ICU patients were similar in both databases (Table 3). The proportion of ICU patients with pLOS-ICU was 26.7% in eICU-CRD and 34.8% in MIMIC-III. In eICU-CRD, the proportion of male patients was 54.8%, and the age of all ICU patients was 61.6 ± 16.6 years. In MIMIC-III, the proportion of male patients was 57.6% and the age of all ICU patients was 62.0 ± 16.5 years.
The prediction performance of the five models on the internal and external validation datasets are compared in Table 4. On eICU-CRD (internal validation dataset), the GBDT model achieved the best overall performance (Brier score, 0.164), discrimination (AUROC, 0.742; AUPRC, 0.537), and second-best calibration (ECI, 8.224). On MIMIC-III, the external validation dataset, the GBDT model also achieved the best overall performance (Brier score, 0.166), discrimination (AUROC, 0.747; AUPRC, 0.536), and calibration (ECI, 8.294). The prediction performance of all the five models on eICU-CRD (internal validation dataset) was superior to that on MIMIC-III. On the internal dataset, the RF and GBDT models performed better than the customized SAPS II, but the SVM model performed slightly worse than the customized SAPS II in Brier score and ECI though it had better discrimination. Meanwhile, all the four ML models performed better than customized SAPS II on the external validation dataset.
Figure 2 shows the calibration plots of the five models on MIMIC-III (external validation dataset). The calibration curve of the GBDT model was an optimal fitting. The customized SAPS II and SVM model tended to overestimate probabilities of pLOS-ICU in most ICU patients, whereas the DL model tended to underestimate probabilities of pLOS in most ICU patients, the RF model tended to underestimate the probabilities of pLOS-ICU in low-risk patients and overestimate the probabilities of pLOS-ICU in high-risk patients.
Top five predictive variables identified by GBDT and SAPS II models are listed in Table 5. Three variables, namely Glasgow Coma Score, systolic blood pressure, and white blood cell count, were ranked among top five important variables by both models.

5. Discussion

In this study, we used four ML methods, namely, SVM, RF, DL, and GBDT, to construct pLOS-ICU prediction models on the basis of eICU-CRD. Furthermore, we used MIMIC-III to validate the developed models externally. The four ML-based models were compared with the customized SAPS II, which is based on the traditional LR algorithm. The comparison results showed that the GBDT outperformed the other four models in terms of discrimination, calibration, and overall performance in either internal or external validation. The main contribution of this study was an optimal data-driven ML model for predicting pLOS-ICU risk, and the model had the following characteristics. First, although some pLOS-ICU prediction models have been developed in the literature [6,16,26], the GBDT-based model developed in this study showed better prediction performance than the state-of-the-art pLOS-ICU prediction models for general ICU patients. Second, most published pLOS-ICU prediction models have not been externally validated, while the GBDT-based prediction model developed in this study was externally validated and the validation results showed a satisfied prediction performance. Third, if clinical application is taken into consideration, a pragmatic pLOS-ICU prediction model could help physicians identify patients at high risk and thus may provide timely individualized interventions, and finally, patients’ prognosis may be improved. Therefore, from application perspective, the pLOS-ICU prediction model developed in this study is an innovational tool though it has limited contributions from the perspective of the ML method.
The RF model has the second-best prediction performance on both datasets. The good prediction performance of the GBDT and RF models may be due to the fact that both models are ensemble learning models, which make predictions by combining the outputs of corresponding basic classifiers, and thus can help reduce the bias that occurs in a single classifier. Our results also verified that ensemble models are usually superior to single models [44,59,60]. The GBDT model is slightly superior to the RF model in terms of all performance measures, probably since the basic decision trees in the two models were trained by two different approaches. In an RF model, decision trees are grown in a parallel way. In a GBDT model, trees are trained iteratively, and each decision tree is trained to correct the discrepancy of all its preceding decision trees. Thus, it helps generate a growing forest with decreasing prediction errors. Our results are consistent with previous studies [61,62,63], which indicated that the GBDT models outperform RF models in terms of prediction. All the four ML models yield superior performance over the customized SAPS II in external validation, although the performance of the customized SAPS II and SVM models slightly differ in internal validation. The possible reason is that ML algorithms have better generalizability compared with traditional LR models [64,65].
Application of ML techniques in healthcare has led to an increased emphasis for ML explainability. For most ML systems, an improved predictive accuracy may often be achieved through increased model complexity [66]. The prime example is the DL paradigm. However, explainability is highly associated with acceptance and promotability of a ML system in clinical practice. There is usually a trade-off between performance and explainability of a ML model. Compared to DL or SVM, GBDT based pLOS-ICU prediction models are less complex and more explainable. The decision trees in the GBDT model can be transformed to understandable decision rules, and thus GBDT may be more acceptable for clinicians. In addition, GBDT has the mechanism to rank the importance of predictable variables from a population perspective, and the importance ranking is intuitionistic for physicians to understand the association of clinical signs and symptoms with pLOS-ICU risk. However, explainability of GBDT is still limited in terms of individual prediction as it is hard to identify predictable variables for each individual prediction. In the literature, some studies have been carried out to improve ML explainability. For example, Ribeiro et al. [67] proposed a local interpretable model-agnostic explanations (LIME) method to generate explainability for individual prediction given by any black-box model. In LIME, an explainable model (such as a decision tree) can be developed based on neighboring instances (identified by the black-box model) of an individual instance. Lundberg et al. [68] proposed a Shapley additive explanations (SHAP) method to enhance explainability by computing the importance value of each feature based on the average marginal contribution to individual predictions. Therefore, ML explainability can be taken into consideration in future pLOS-ICU prediction model development.
The overall prediction performance of all five models on eICU-CRD (internal validation dataset) is superior to that of MIMIC-III. This result suggests that a possible degradation of prediction performance occurs when the models are applied to a new dataset. The internal and external validation datasets exhibit a slight difference in performance in our study, demonstrating a strong generalizability of the developed pLOS-ICU prediction models. Such strong generalizability may be attributed to the fact that the pLOS-ICU models were trained on the basis of a multicenter dataset, which is more population representative than a single-center dataset.
The top five important variables identified by the GBDT model include Pao2/Fio2 ratio, Glasgow Coma Score, serum urea nitrogen level, systolic blood pressure, and white blood cell count. Three variables (i.e., Glasgow Coma Score, systolic blood pressure, and white blood cell count) are also ranked among the top five by the SAPS II model. The Glasgow Coma Score is used to assess the level of consciousness in patients [69], and patients with decreased levels of consciousness tend to have poor prognosis [70,71]. The independent capabilities of systolic blood pressure and white blood cell count in predicting the prognosis of ICU patients have also been verified by existing studies [72,73]. The top five important variables identified by the two models only show a slight difference. As identified by the GBDT model, Pao2/Fio2 ratio and serum urea nitrogen level, in addition to Glasgow Coma Score, systolic blood pressure, and white blood cell count, may have potential influence on the prognosis of ICU patients. This finding may provide a clue for future research. Studies focusing on the detailed association between the prognosis of ICU patients and the Pao2/Fio2 ratio or serum urea nitrogen level are limited [74,75,76].
This study has several strengths. First, the database used to derive the prediction models is a large, multicenter database with a relatively representative population. Second, a large, single-center database (i.e., MIMIC-III) was used for external validation and helped assess the generalizability of the developed prediction models. Third, all the predictor variables used to construct prediction models are routinely collected during the first 24 h in ICU, thereby ensuring the feasibility of applying the prediction models in clinical practice to assist physicians in decision making.
However, this study has limitations. First, the eICU-CRD database contains only data of ICU patients admitted between 2014 and 2015 in the US, and the MIMIC-III dataset contains only data of ICU patients admitted to the BIDMC from 2001 to 2012. No data from other countries were used for model validation. Therefore, the clinical utility of the pLOS-ICU prediction models needs further assessment before application in other regions. Second, selection bias may exist since we excluded patients who died in the ICU. Accordingly, the pLOS-ICU prediction models developed in this study may not apply to patients who die in ICU. Third, to compare the prediction performance of four ML-based models with the customized SAPS II in an objective manner, we only included the predictor variables used in the customized SAPS II for model training. Other potential predictor variables may have been neglected in our study.

6. Conclusions

In summary, this study demonstrates that the GBDT model outperforms the other four developed models in pLOS-ICU prediction. As all the predictor variables can be available during the first 24 h in ICU, the GBDT-based pLOS-ICU prediction model has potential to assist ICU physicians in identifying patients with pLOS-ICU risk and thus make optimal clinical intervention decisions. This study lays a foundation for the future application of a GBDT-based pLOS-ICU prediction model in ICU clinical practice.

Author Contributions

Conceptualization, J.W., P.L., L.Z., and G.K.; data curation, Y.H. and G.K.; formal analysis, J.W. and Y.L.; investigation, J.W. and Y.L.; methodology, J.W. and Y.L.; project administration, G.K.; resources, P.L., Y.H., L.Z., and G.K.; supervision, L.Z. and G.K.; validation, J.W. and G.K.; writing—original draft, J.W. and G.K.; writing—review and editing, J.W. and G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by grants from the National Natural Science Foundation of China (Grant Nos. 81771938 and 91846101), Beijing Municipal Science & Technology Commission (Grant No. 7212201), The National Key Research and Development Program of China (Grant No. 2018AAA0102100), the University of Michigan Health System-Peking University Health Science Center Joint Institute for Translational and Clinical Research (Grant No. BMU2020JI011), PKU-Baidu Fund (Grant Nos. 2020BD005, 2019BD017), and Zhejiang Provincial Natural Science Foundation of China (Grant No.Z22F026106).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Health Insurance Portability and Accountability Act (HIPAA) safe harbor provision. The re-identification risk was certified as meeting safe harbor standards by Privacert (Cambridge, MA, USA) (HIPAA Certification no. 1031219-2).

Informed Consent Statement

Not applicable.

Data Availability Statement

Data supporting reported results in this study can be found at https://eicu-crd.mit.edu/ (accessed on 20 December 2018) and https://mimic.mit.edu/ (accessed on 20 December 2018).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Meyer, A.; Zverinski, D.; Pfahringer, B.; Kempfert, J.; Kuehne, T.; Sündermann, S.H.; Stamm, C.; Hofmann, T.; Falk, V.; Eickhoff, C. Machine learning for real-time prediction of complications in critical care: A retrospective study. Lancet Respir. Med. 2018, 6, 905–914. [Google Scholar] [CrossRef]
  2. Vincent, J.L.; Marshall, J.C.; Namendys-Silva, S.A.; Francois, B.; Martin-Loeches, I.; Lipman, J.; Reinhart, K.; Antonelli, M.; Pickkers, P.; Njimi, H.; et al. Assessment of the worldwide burden of critical illness: The Intensive Care Over Nations (ICON) audit. Lancet Resp. Med. 2014, 2, 380–386. [Google Scholar] [CrossRef]
  3. Halpern, N.A.; Pastores, S.M. Critical care medicine in the United States 2000-2005: An analysis of bed numbers, occupancy rates, payer mix, and costs. Crit. Care Med. 2010, 38, 65–71. [Google Scholar] [CrossRef]
  4. Verburg, I.W.M.; Atashi, A.; Eslami, S.; Holman, R.; Abu-Hanna, A.; de Jonge, E.; Peek, N.; de Keizer, N.F. Which models can I use to predict adult ICU length of stay? A systematic review. Crit. Care Med. 2017, 45, e222–e231. [Google Scholar] [CrossRef] [PubMed]
  5. Kılıç, M.; Yüzkat, N.; Soyalp, C.; Gülhaş, N. Cost analysis on Intensive Care Unit costs based on the length of stay. Turk. J. Anaesthesiol. Reanim. 2019, 47, 142. [Google Scholar] [CrossRef]
  6. Herman, C.; Karolak, W.; Yip, A.M.; Buth, K.J.; Hassan, A.; Legare, J.-F. Predicting prolonged intensive care unit length of stay in patients undergoing coronary artery bypass surgery—Development of an entirely preoperative scorecard. Interact. Cardiovasc. Thorac. Surg. 2009, 9, 654–658. [Google Scholar] [CrossRef]
  7. Verburg, I.W.M.; de Keizer, N.F.; de Jonge, E.; Peek, N. Comparison of Regression Methods for Modeling Intensive Care Length of Stay. PLoS ONE 2014, 9, e109684. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Evans, J.; Kobewka, D.; Thavorn, K.; D’Egidio, G.; Rosenberg, E.; Kyeremanteng, K. The impact of reducing intensive care unit length of stay on hospital costs: Evidence from a tertiary care hospital in Canada. Can. J. Anesth. J. Can. d’anesthésie 2018, 65, 627–635. [Google Scholar] [CrossRef] [Green Version]
  9. Stricker, K.; Rothen, H.U.; Takala, J. Resource use in the ICU: Short- vs. long-term patients. Acta Anaesthesiol. Scand. 2003, 47, 508–515. [Google Scholar] [CrossRef]
  10. Nassar, A.P.; Caruso, P. ICU physicians are unable to accurately predict length of stay at admission: A prospective study. Int. J. Qual. Health Care 2016, 28, 99–103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Monteiro, F.; Meloni, F.; Baranauskas, J.A.; Macedo, A.A. Prediction of mortality in Intensive Care Units: A multivariate feature selection. J. Biomed. Inform. 2020, 107, 103456. [Google Scholar] [CrossRef]
  12. Pereira, T.; Gadhoumi, K.; Ma, M.; Liu, X.; Xiao, R.; Colorado, R.A.; Keenan, K.J.; Meisel, K.; Hu, X. A Supervised Approach to Robust Photoplethysmography Quality Assessment. IEEE J. Biomed. Health Inform. 2020, 24, 649–657. [Google Scholar] [CrossRef]
  13. Qian, Q.; Wu, J.; Wang, J.; Sun, H.; Yang, L. Prediction Models for AKI in ICU: A Comparative Study. Int. J. Gen. Med. 2021, 14, 623–632. [Google Scholar] [CrossRef]
  14. Li, X.; Ge, P.; Zhu, J.; Li, H.; Graham, J.; Singer, A.; Richman, P.S.; Duong, T.Q. Deep learning prediction of likelihood of ICU admission and mortality in COVID-19 patients using clinical variables. PeerJ 2020, 8, e10337. [Google Scholar] [CrossRef]
  15. Theis, J.; Galanter, W.; Boyd, A.; Darabi, H. Improving the In-Hospital Mortality Prediction of Diabetes ICU Patients Using a Process Mining/Deep Learning Architecture. IEEE J. Biomed. Health Inform. 2021. [Google Scholar] [CrossRef] [PubMed]
  16. Zoller, B.; Spanaus, K.; Gerster, R.; Fasshauer, M.; Stehberger, P.A.; Klinzing, S.; Vergopoulos, A.; von Eckardstein, A.; Bechir, M. ICG-liver test versus new biomarkers as prognostic markers for prolonged length of stay in critically ill patients—A prospective study of accuracy for prediction of length of stay in the ICU. Ann. Intensive Care 2014, 4, 5. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Eachempati, S.R.; Hydo, L.J.; Barie, P.S. Severity scoring for prognostication in patients with severe acute pancreatitis—Comparative analysis of the ranson score and the APACHE III score. Arch. Surg. 2002, 137, 730–736. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Houthooft, R.; Ruyssinck, J.; van der Herten, J.; Stijven, S.; Couckuyt, I.; Gadeyne, B.; Ongenae, F.; Colpaert, K.; Decruyenaere, J.; Dhaene, T.; et al. Predictive modelling of survival and length of stay in critically ill patients using sequential organ failure scores. Artif. Intell. Med. 2015, 63, 191–207. [Google Scholar] [CrossRef] [Green Version]
  19. Rotar, E.P.; Beller, J.P.; Smolkin, M.E.; Chancellor, W.Z.; Ailawadi, G.; Yarboro, L.T.; Hulse, M.; Ratcliffe, S.J.; Teman, N.R. Prediction of prolonged intensive care unit length of stay following cardiac surgery. In Seminars in Thoracic and Cardiovascular Surgery; Elsevier: Amsterdam, The Netherlands, 2021. [Google Scholar]
  20. Lin, K.; Hu, Y.; Kong, G. Predicting in-hospital mortality of patients with acute kidney injury in the ICU using random forest model. Int. J. Med Inform. 2019, 125, 55–61. [Google Scholar] [CrossRef]
  21. Wellner, B.; Grand, J.; Canzone, E.; Coarr, M.; Brady, P.W.; Simmons, J.; Kirkendall, E.; Dean, N.; Kleinman, M.; Sylvester, P. Predicting Unplanned Transfers to the Intensive Care Unit: A Machine Learning Approach Leveraging Diverse Clinical Elements. JMIR Med. Inf. 2017, 5, 16. [Google Scholar] [CrossRef]
  22. Pirracchio, R.; Petersen, M.L.; Carone, M.; Rigon, M.R.; Chevret, S.; van der Laan, M.J. Mortality prediction in intensive care units with the Super ICU Learner Algorithm (SICULA): A population-based study. Lancet Resp. Med. 2015, 3, 42–52. [Google Scholar] [CrossRef] [Green Version]
  23. Gonzalez-Robledo, J.; Martin-Gonzalez, F.; Sanchez-Barba, M.; Sanchez-Hernandez, F.; Moreno-Garcia, M.N. Multiclassifier Systems for Predicting Neurological Outcome of Patients with Severe Trauma and Polytrauma in Intensive Care Units. J. Med. Syst. 2017, 41, 8. [Google Scholar] [CrossRef]
  24. Meiring, C.; Dixit, A.; Harris, S.; MacCallum, N.S.; Brealey, D.A.; Watkinson, P.J.; Jones, A.; Ashworth, S.; Beale, R.; Brett, S.J.; et al. Optimal intensive care outcome prediction over time using machine learning. PLoS ONE 2018, 13, e0206862. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Viton, F.; Elbattah, M.; Guérin, J.-L.; Dequen, G. Multi-channel ConvNet Approach to Predict the Risk of in-Hospital Mortality for ICU Patients. DeLTA 2020, 98–102. [Google Scholar]
  26. Navaz, A.N.; Mohammed, E.; Serhani, M.A.; Zaki, N. The Use of Data Mining Techniques to Predict Mortality and Length of Stay in an ICU. In Proceedings of the 2016 12th International Conference on Innovations in Information Technology, Abu Dhabi, United Arab Emirates, 28–30 November 2016; IEEE: New York, NY, USA, 2016; pp. 164–168. [Google Scholar]
  27. Rocheteau, E.; Liò, P.; Hyland, S. Temporal pointwise convolutional networks for length of stay prediction in the intensive care unit. In Proceedings of the Conference on Health, Inference, and Learning, Virtual Event, Association for Computing Machinery, New York, NY, USA, 8–10 April 2021; pp. 58–68. [Google Scholar]
  28. Ma, X.; Si, Y.; Wang, Z.; Wang, Y. Length of stay prediction for ICU patients using individualized single classification algorithm. Comput. Methods Programs Biomed. 2020, 186, 105224. [Google Scholar] [CrossRef] [PubMed]
  29. Vasilevskis, E.E.; Kuzniewicz, M.W.; Cason, B.A.; Lane, R.K.; Dean, M.L.; Clay, T.; Rennie, D.J.; Vittinghoff, E.; Dudley, R.A. Mortality Probability Model III and Simplified Acute Physiology Score II Assessing Their Value in Predicting Length of Stay and Comparison to APACHE IV. Chest 2009, 136, 89–101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Ettema, R.G.A.; Peelen, L.M.; Schuurmans, M.J.; Nierich, A.P.; Kalkman, C.J.; Moons, K.G.M. Prediction Models for Prolonged Intensive Care Unit Stay After Cardiac Surgery Systematic Review and Validation Study. Circulation 2010, 122, 682–689. [Google Scholar] [CrossRef] [Green Version]
  31. Pollard, T.J.; Johnson, A.E.W.; Raffa, J.D.; Celi, L.A.; Mark, R.G.; Badawi, O. The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Sci. Data 2018, 5, 13. [Google Scholar] [CrossRef]
  32. Johnson, A.E.W.; Pollard, T.J.; Shen, L.; Lehman, L.W.H.; Feng, M.L.; Ghassemi, M.; Moody, B.; Szolovits, P.; Celi, L.A.; Mark, R.G. MIMIC-III, a freely accessible critical care database. Sci. Data 2016, 3, 9. [Google Scholar] [CrossRef] [Green Version]
  33. Soveri, I.; Snyder, J.; Holdaas, H.; Holme, I.; Jardine, A.G.; L’Italien, G.J.; Fellstrom, B. The External Validation of the Cardiovascular Risk Equation for Renal Transplant Recipients: Applications to BENEFIT and BENEFIT-EXT Trials. Transplantation 2013, 95, 142–147. [Google Scholar] [CrossRef]
  34. Grooms, K.N.; Ommerborn, M.J.; Pham, D.Q.; Djousse, L.; Clark, C.R. Dietary Fiber Intake and Cardiometabolic Risks among US Adults, NHANES 1999-2010. Am. J. Med. 2013, 126, 1059–1067.e4. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Harel, Z.; Wald, R.; McArthur, E.; Chertow, G.M.; Harel, S.; Gruneir, A.; Fischer, N.D.; Garg, A.X.; Perl, J.; Nash, D.M.; et al. Rehospitalizations and Emergency Department Visits after Hospital Discharge in Patients Receiving Maintenance Hemodialysis. J. Am. Soc. Nephrol. 2015, 26, 3141–3150. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Song, X.; Xia, C.; Li, Q.; Yao, C.; Yao, Y.; Chen, D.; Jiang, Q. Perioperative predictors of prolonged length of hospital stay following total knee arthroplasty: A retrospective study from a single center in China. BMC Musculoskelet. Disord. 2020, 21, 62. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Lilly, C.M.; Zuckerman, I.H.; Badawi, O.; Riker, R.R. Benchmark Data From More Than 240,000 Adults That Reflect the Current Practice of Critical Care in the United States. Chest 2011, 140, 1232–1242. [Google Scholar] [CrossRef] [PubMed]
  38. Legall, J.R.; Lemeshow, S.; Saulnier, F. A new simplified acute physiology score (SAPS-II) based on a European North-American multicenter study. JAMA 1993, 270, 2957–2963. [Google Scholar] [CrossRef]
  39. Chang, C.-C.; Lin, C.-J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 27. [Google Scholar] [CrossRef]
  40. Zoppis, I.; Mauri, G.; Dondi, R. Kernel methods: Support vector machines. Encycl. Bioinform. Comput. Biol. 2019, 1, 503–510. [Google Scholar]
  41. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  42. Fayed, H.A.; Atiya, A.F. Speed up grid-search for parameter selection of support vector machines. Appl. Soft Comput. 2019, 80, 202–210. [Google Scholar] [CrossRef]
  43. Probst, P.; Wright, M.N.; Boulesteix, A.L. Hyperparameters and tuning strategies for random forest. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2019, 9, e1301. [Google Scholar] [CrossRef] [Green Version]
  44. Supatcha, L.; Chinae, T.; Chakarida, N.; Boonserm, K.; Marasri, R. Heterogeneous ensemble approach with discriminative features and modified-SMOTEbagging for pre-miRNA classification. Nucleic Acids Res. 2013, 41, e21. [Google Scholar]
  45. Kulkarni, V.Y.; Sinha, P.K.; Petare, M.C. Weighted Hybrid Decision Tree Model for Random Forest Classifier. J. Inst. Eng. 2016, 97, 209–217. [Google Scholar] [CrossRef]
  46. Luo, R.; Tan, X.; Wang, R.; Qin, T.; Chen, E.; Liu, T.-Y. Accuracy Prediction with Non-neural Model for Neural Architecture Search. arXiv 2020, arXiv:2007.04785. [Google Scholar]
  47. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Montreal, QC, Canada, 31 March 2010; pp. 249–256. [Google Scholar]
  48. Yu, H.; Liu, J.; Chen, C.; Heidari, A.A.; Zhang, Q.; Chen, H.; Mafarja, M.; Turabieh, H. Corn leaf diseases diagnosis based on K-means clustering and deep learning. IEEE Access 2021, 9, 143824–143835. [Google Scholar] [CrossRef]
  49. Gu, J.; Tong, T.; He, C.; Xu, M.; Yang, X.; Tian, J.; Jiang, T.; Wang, K. Deep learning radiomics of ultrasonography can predict response to neoadjuvant chemotherapy in breast cancer at an early stage of treatment: A prospective study. Eur. Radiol. 2021, 1–11. [Google Scholar] [CrossRef]
  50. Yu, K.; Yang, Z.; Wu, C.; Huang, Y.; Xie, X. In-hospital resource utilization prediction from electronic medical records with deep learning. Knowl.-Based Syst. 2021, 223, 107052. [Google Scholar] [CrossRef]
  51. Imai, S.; Takekuma, Y.; Kashiwagi, H.; Miyai, T.; Kobayashi, M.; Iseki, K.; Sugawara, M. Validation of the usefulness of artificial neural networks for risk prediction of adverse drug reactions used for individual patients in clinical practice. PLoS ONE 2020, 15, e0236789. [Google Scholar] [CrossRef]
  52. Zeng, Z.; Yao, S.; Zheng, J.; Gong, X. Development and validation of a novel blending machine learning model for hospital mortality prediction in ICU patients with Sepsis. BioData Min. 2021, 14, 40. [Google Scholar] [CrossRef] [PubMed]
  53. Dzaharudin, F.; Ralib, A.; Jamaludin, U.; Nor, M.; Tumian, A.; Har, L.; Ceng, T. Mortality Prediction in Critically Ill Patients Using Machine Learning Score; IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2020. [Google Scholar]
  54. Fallenius, M.; Skrifvars, M.B.; Reinikainen, M.; Bendel, S.; Raj, R. Common intensive care scoring systems do not outperform age and glasgow coma scale score in predicting mid-term mortality in patients with spontaneous intracerebral hemorrhage treated in the intensive care unit. Scand. J. Trauma Resusc. Emerg. Med. 2017, 25, 102. [Google Scholar] [CrossRef] [Green Version]
  55. Jahn, M.; Rekowski, J.; Gerken, G.; Kribben, A.; Canbay, A.; Katsounas, A. The predictive performance of SAPS 2 and SAPS 3 in an intermediate care unit for internal medicine at a German university transplant center; A retrospective analysis. PLoS ONE 2019, 14, e0222164. [Google Scholar] [CrossRef] [PubMed]
  56. Steyerberg, E.W.; Vickers, A.J.; Cook, N.R.; Gerds, T.; Gonen, M.; Obuchowski, N.; Pencina, M.J.; Kattan, M.W. Assessing the Performance of Prediction Models A Framework for Traditional and Novel Measures. Epidemiology 2010, 21, 128–138. [Google Scholar] [CrossRef] [Green Version]
  57. Van Hoorde, K.; Van Huffel, S.; Timmerman, D.; Bourne, T.; Van Calster, B. A spline-based tool to assess and visualize the calibration of multiclass risk predictions. J. Biomed. Inform. 2015, 54, 283–293. [Google Scholar] [CrossRef] [Green Version]
  58. Saito, T.; Rehmsmeier, M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar]
  59. Wang, S.Q.; Yang, J.; Chou, K.C. Using stacked generalization to predict membrane protein types based on pseudo-amino acid composition. J. Theor. Biol. 2006, 242, 941–946. [Google Scholar] [CrossRef]
  60. Phan, J.H.; Hoffman, R.; Kothari, S.; Wu, P.Y.; Wang, M.D. Integration of Multi-Modal Biomedical Data to Predict Cancer Grade and Patient Survival; IEEE: New York, NY, USA, 2016; pp. 577–580. [Google Scholar]
  61. Zhang, C.S.; Liu, C.C.; Zhang, X.L.; Almpanidis, G. An up-to-date comparison of state-of-the-art classification algorithms. Expert Syst. Appl. 2017, 82, 128–150. [Google Scholar] [CrossRef]
  62. Ichikawa, D.; Saito, T.; Ujita, W.; Oyama, H. How can machine-learning methods assist in virtual screening for hyperuricemia? A healthcare machine-learning approach. J. Biomed. Inform. 2016, 64, 20–24. [Google Scholar] [CrossRef]
  63. Zhou, C.; Yu, H.; Ding, Y.J.; Guo, F.; Gong, X.J. Multi-scale encoding of amino acid sequences for predicting protein interactions using gradient boosting decision tree. PLoS ONE 2017, 12, 18. [Google Scholar] [CrossRef] [PubMed]
  64. Hotzy, F.; Theodoridou, A.; Hoff, P.; Schneeberger, A.R.; Seifritz, E.; Olbrich, S.; Jager, M. Machine Learning: An Approach in Identifying Risk Factors for Coercion Compared to Binary Logistic Regression. Front. Psychiatry 2018, 9, 11. [Google Scholar] [CrossRef]
  65. Xiong, J. Radial Distance Weighted Discrimination. Ph.D. Thesis, The University of North Carolina at Chapel Hill, Chapel Hill, NC, USA; 120p.
  66. Linardatos, P.; Papastefanopoulos, V.; Kotsiantis, S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy 2021, 23, 18. [Google Scholar] [CrossRef] [PubMed]
  67. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  68. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4768–4777. [Google Scholar]
  69. Mehta, R.; Chinthapalli, K. Glasgow coma scale explained. BMJ 2019, 365, l1296. [Google Scholar] [CrossRef] [PubMed]
  70. Al-Mufti, F.; Amuluru, K.; Lander, M.; Mathew, M.; El-Ghanem, M.; Nuoman, R.; Park, S.; Patel, V.; Singh, I.P.; Gupta, G. Low glasgow coma score in traumatic intracranial hemorrhage predicts development of cerebral vasospasm. World Neurosurg. 2018, 120, e68–e71. [Google Scholar] [CrossRef] [PubMed]
  71. Ghelichkhani, P.; Esmaeili, M.; Hosseini, M.; Seylani, K. Glasgow coma scale and four score in predicting the mortality of trauma patients; a diagnostic accuracy study. Emergency 2018, 6, e42. [Google Scholar]
  72. Gayat, E.; Cariou, A.; Deye, N.; Vieillard-Baron, A.; Jaber, S.; Damoisel, C.; Lu, Q.; Monnet, X.; Rennuit, I.; Azoulay, E. Determinants of long-term outcome in ICU survivors: Results from the FROG-ICU study. Crit. Care 2018, 22, 8. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Sahin, S.; Dogan, U.; Ozdemir, K.; Gok, H. Evaluation of clinical and demographic characteristics and their association with length of hospital stay in patients admitted to cardiac intensive care unit with the diagnosis of acute heart failure. Anat. J. Cardiol. 2012, 12, 123–131. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  74. Esteve, F.; Lopez-Delgado, J.C.; Javierre, C.; Skaltsa, K.; Carrio, M.L.L.; Rodriguez-Castro, D.; Torrado, H.; Farrero, E.; Diaz-Prieto, A.; Ventura, J.L.L.; et al. Evaluation of the PaO2/FiO2 ratio after cardiac surgery as a predictor of outcome during hospital stay. BMC Anesthesiol. 2014, 14, 9. [Google Scholar] [CrossRef] [Green Version]
  75. Piriyapatsom, A.; Williams, E.C.; Waak, K.; Ladha, K.S.; Eikermann, M.; Schmidt, U.H. Prospective Observational Study of Predictors of Re-Intubation Following Extubation in the Surgical ICU. Respir. Care 2016, 61, 306–315. [Google Scholar] [CrossRef] [Green Version]
  76. Faisst, M.; Wellner, U.F.; Utzolino, S.; Hopt, U.T.; Keck, T. Elevated blood urea nitrogen is an independent risk factor of prolonged intensive care unit stay due to acute necrotizing pancreatitis. J. Crit. Care 2010, 25, 105–111. [Google Scholar] [CrossRef]
Figure 1. The procedure of study population selection.
Figure 1. The procedure of study population selection.
Diagnostics 11 02242 g001
Figure 2. Calibration plots of the five models on MIMIC-III.
Figure 2. Calibration plots of the five models on MIMIC-III.
Diagnostics 11 02242 g002
Table 1. Characteristics of related work.
Table 1. Characteristics of related work.
CategoryStudyPopulationSample SizeDatasetOutcomeModelsExternal ValidationPerformance
Traditional regression based pLOS-ICU prediction modelsZoller et al. [16]General ICU patients110Single-centerpLOS-ICUCustomized SAPS II×AUROC: 0.70
Houthooft et al. [18]General ICU patients14,480Single-centerpLOS-ICUCustomized SOFA score×Sensitivity: 0.71
Herman et al. [6]Patients undergoing CABG3483Single-centerpLOS-ICULR×AUROC: 0.78
Rotar et al. [19] Patients following CABG3283Single-centerpLOS-ICULASSO×AUROC: 0.72
ML-based models in ICUMeiring et al. [24]General ICU patients22,514MulticenterICU mortalityAdaBoost, RF, SVM, DL, LR, and customized APACHE-II×AUROC: 0.88 (DL)
Lin et al. [20]Acute kidney injury patients19,044Single-centerICU mortalityANN, SVM, RF, and customized SAPS II×AUROC: 0.87 (RF)
Viton et al. [25]General ICU patients13,000Single-centerICU mortalityDL×AUROC: 0.85
Qian et al. [13]General ICU patients17,205Single-centerAcute kidney injuryXGBoost, RF, SVM, GBDT, DL, and LR×AUROC: 0.91 (GBDT)
ML-based pLOS-ICU prediction modelsNavaz et al. [26]General ICU patients40,426Single-centerpLOS-ICUDecision tree×Accuracy: 0.59
Rocheteau et al. [27]General ICU patients168,577MulticenterLOS-ICUDLR2: 0.40
Ma et al. [28]General ICU patients4000Single-centerpLOS-ICUCombining just-in-time learning and one-class extreme learning×AUROC: 0.85
Our studyGeneral ICU patients160,238MulticenterpLOS-ICURF, SVM, DL, GBDT, and customized SAPS II-
Table 2. Characteristics of eICU-CRD and MIMIC-III.
Table 2. Characteristics of eICU-CRD and MIMIC-III.
ItemseICU-CRDMIMIC-III
CountryUnited StatesUnited States
DataMulticenterSingle-center
Year2014–20152001–2012
Number of units3351
Number of hospitals2081
Number of patients139,36738,597
Number of admissions200,85953,423
DeidentificationAll protected health information was deidentified, and no patient privacy data can be identified.
Data contentVital sign measurements, laboratory tests, care plan documentation, diagnosis information, treatment information, and others.
Table 3. Characteristics of ICU patients in eICU-CRD and MIMIC-III.
Table 3. Characteristics of ICU patients in eICU-CRD and MIMIC-III.
ItemseICU-CRDMIMIC-III
Total number117,30642,932
Age/years61.6 ± 16.662.0 ± 16.5
Gender, n (%)
Male64,244 (54.8%)24,740 (57.6%)
Female53,049 (45.2%)18,192 (42.4%)
SAPS II score30.0 ± 13.332.7 ± 12.7
LOS-ICU (IQR1)/day1.8 (1.0–3.2)2.1 (1.2–4.0)
PLOS-ICU, n (%)31,296 (26.7%)14,951 (34.8%)
IQR1, interquartile range.
Table 4. Prediction performance of the five models on eICU-CRD (internal) and MIMIC-III (external).
Table 4. Prediction performance of the five models on eICU-CRD (internal) and MIMIC-III (external).
ModelseICU-CRDMIMIC-III
Brier ScoreAUROCAUPRCECIBrier ScoreAUROCAUPRCECI
Customized SAPS II0.1810.6670.4399.0280.1750.6690.4028.742
RF0.1660.7350.5308.3170.1690.7450.5308.469
SVM0.1830.6900.4809.1370.1720.7160.4828.577
DL0.1640.7420.5368.2230.1710.7430.5278.551
GBDT0.1640.7420.5378.2240.1660.7470.5368.294
Table 5. Top five important variables identified by GBDT and SAPS II.
Table 5. Top five important variables identified by GBDT and SAPS II.
RanksGBTDSAPS II
1Pao2/Fio2 ratioGlasgow Coma Score
2Glasgow Coma ScoreAge
3Serum urea nitrogen levelChronic diseases
4Systolic blood pressureSystolic blood pressure
5White blood cell countWhite blood cell count
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, J.; Lin, Y.; Li, P.; Hu, Y.; Zhang, L.; Kong, G. Predicting Prolonged Length of ICU Stay through Machine Learning. Diagnostics 2021, 11, 2242. https://doi.org/10.3390/diagnostics11122242

AMA Style

Wu J, Lin Y, Li P, Hu Y, Zhang L, Kong G. Predicting Prolonged Length of ICU Stay through Machine Learning. Diagnostics. 2021; 11(12):2242. https://doi.org/10.3390/diagnostics11122242

Chicago/Turabian Style

Wu, Jingyi, Yu Lin, Pengfei Li, Yonghua Hu, Luxia Zhang, and Guilan Kong. 2021. "Predicting Prolonged Length of ICU Stay through Machine Learning" Diagnostics 11, no. 12: 2242. https://doi.org/10.3390/diagnostics11122242

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop