Next Article in Journal
The Effects of Intensive Residential Treatment for Feeding and Eating Disorders (FEDs) in Adolescence: The Case of an Italian Facility
Previous Article in Journal
Selenium and Skeletal Muscle Health in Sports Nutrition
Previous Article in Special Issue
Mapping Research Trends on Intestinal Permeability in Irritable Bowel Syndrome with a Focus on Nutrition: A Bibliometric Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Methodological Review of Classification Trees for Risk Stratification: An Application Example in the Obesity Paradox

by
Javier Trujillano
1,*,†,
Luis Serviá
1,†,
Mariona Badia
1,
José C. E. Serrano
2,
María Luisa Bordejé-Laguna
3,
Carol Lorencio
4,
Clara Vaquerizo
5,
José Luis Flordelis-Lasierra
6,7,
Itziar Martínez de Lagrán
8,
Esther Portugal-Rodríguez
9 and
Juan Carlos López-Delgado
10
1
IRBLLeida (Institut de Recerca Biomèdica de Lleida Fundació Dr. Pifarré), Av. Alcalde Rovira Roure, 80, 25198 Lleida, Spain
2
NUTREN-Nutrigenomics, Department of Experimental Medicine, University of Lleida, 25198 Lleida, Spain
3
Intensive Care Department, Hospital Universitario Germans Trias i Pujol, Carretera de Canyet, s/n, 08916 Badalona, Spain
4
Intensive Care Department, Hospital Universitari Josep Trueta, Av. de França, s/n, 17007 Girona, Spain
5
Intensive Care Department, Hospital Universitario de Fuenlabrada, Cam. del Molino, 2, 28942 Fuenlabrada, Spain
6
Intensive Care Department, Hospital Universitario 12 de Octubre, Av. de Córdoba s/n, 28041 Madrid, Spain
7
4i+12 (Instituto de Investigación Sanitaria Hospital 12 de Octubre, Research Institute Hospital 12 de Octubre), Av. de Córdoba s/n, 28041 Madrid, Spain
8
Intensive Care Department, Hospital de Mataró, 08304 Mataró, Spain
9
Intensive Care Department, Hospital Clínico Universitario de Valladolid, Av. Ramón y Cajal, 3, 47003 Valladolid, Spain
10
Area de Vigilancia Intensiva, Clinical Institute of Internal Medicine & Dermatology (ICMiD), Hospital Clínic de Barcelona, C/Villarroel, 170, 08036 Barcelona, Spain
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Nutrients 2025, 17(11), 1903; https://doi.org/10.3390/nu17111903
Submission received: 2 May 2025 / Revised: 22 May 2025 / Accepted: 28 May 2025 / Published: 31 May 2025
(This article belongs to the Special Issue Biostatistics Methods in Nutritional Research)

Abstract

:
Background: Classification trees (CTs) are widely used machine learning algorithms with growing applications in clinical research, especially for risk stratification. Their ability to generate interpretable decision rules makes them attractive to healthcare professionals. This review provides an accessible yet rigorous overview of CT methodology for clinicians, highlighting their utility through a case study addressing the “obesity paradox” in critically ill patients. Methods: We describe key methodological aspects of CTs, including model development, pruning, validation, and classification types (simple, ensemble, and hybrid). Using data from the ENPIC (Evaluation of Practical Nutrition Practices in the Critical Care Patient) study, which assessed artificial nutrition in ICU (intensive care unit) patients, we applied various CT approaches—CART (classification and regression trees), CHAID (chi-square automatic interaction detection), and XGBoost (extreme gradient boosting)—and compared them with logistic regression. SHAP (SHapley Additive exPlanation) values were used to interpret ensemble models. Results: CTs allowed for identification of optimal cut-off points in continuous variables and revealed complex, non-linear interactions among predictors. Although the obesity paradox was not confirmed in the full cohort, CTs uncovered a specific subgroup in which obesity was associated with reduced mortality. The ensemble model (XGBoost) achieved the best predictive performance (highest area under the ROC curve), though at the expense of interpretability. Conclusions: CTs are valuable tools in clinical epidemiology, complementing traditional models by uncovering hidden patterns and enhancing risk stratification. While ensemble models offer superior predictive accuracy, their complexity necessitates interpretability techniques such as SHAP. CT-based approaches can guide personalized medicine but require cautious interpretation and external validation.

1. Introduction

In medical decision-making, tools are needed to assist in risk stratification for the occurrence of adverse events such as death or complications, as well as for making accurate diagnoses [1]. The development of such tools requires the application of various methodologies for risk model construction. Numerous approaches can be employed, ranging from classical statistical methods to advanced machine learning techniques [1]. Among these, classification tree (CT)-based methodologies remain an appealing option due to their specific advantages [2].
This work aims to provide an overview of the methodology of classification trees (CT), offering a perspective directed at clinical professionals interested in risk models. A technical yet accessible language has been used so that the content can be understood without requiring extensive methodological knowledge. For readers interested in a deeper understanding of the topic, we provide bibliographic references that we consider most suitable for further exploration, including examples related to nutrition-related problems.
Our focus is limited to classification trees that generate classification rules essential for establishing relationships between variables and identifying groups of patients with specific characteristics. CTs belong to a family of machine learning algorithms that use tree-like structures to support decision-making. This review is accompanied by a real-data application example to illustrate the utility and features of CTs.

1.1. Concept of a Classification Tree

A CT is the graphical representation of a series of classification rules. Starting from a root node, which includes all cases, the tree branches into different “child” nodes containing subgroups of cases. The splitting criterion, also known as branching criterion, is optimally determined after examining the values of all included predictor variables [2]. CTs are a form of supervised machine learning in which the algorithm is provided with records that include predictor variables and the outcome variable. These algorithms function by reducing classification error until the optimal CT is found [3].
CT methodology has been in use for quite some time. The earliest references to CTs are attributed to Quinlan in 1986 [4]. Regression trees, which use continuous outcome variables, had already been in use for over 50 years by then [3].

1.2. Phases in the Construction of a Classification Tree

To illustrate the process of constructing a CT, we use the CART (classification and regression tree) model as a reference. This process can be divided into several phases [5]:
  • Phase 1—Tree Development: From the root node, the most appropriate variable is identified to split the node into two child nodes by establishing an optimal cut-off point if the variable is continuous. Each child node is subsequently split following the same methodology. A supervised machine learning model is used, with all records including predictor variables and the outcome variable submitted to the algorithm.
  • Phase 2—Tree Growth Stopping Criteria: Tree development can continue until terminal nodes contain only a single case, or when the value of the dependent variable is the same for all cases within a node. Additional criteria, such as a minimum number of cases per node, can be defined to prevent excessive branching.
  • Phase 3—Tree Pruning: A CT developed using the aforementioned method tends to be overly complex and branched, which may lead to overfitting the training dataset. Removing superfluous branches results in a simpler tree with better generalizability. The pruning process uses predefined cost–complexity criteria to eliminate branches that add more complexity than effectiveness. Supervised learning aims to reduce classification error.
  • Phase 4—Selection of the Optimal CT: Selecting the optimal CT requires an internal validation system. This can be achieved by randomly splitting the sample into a training set and a validation set, or by applying cross-validation techniques. Cross-validation divides the dataset into subsets—e.g., 10 partitions, using 9 for training and 1 for validation in a recursive process.
A final CT includes the classification rules that generate a probability for the event of interest, such as mortality or disease diagnosis.

1.3. Use of Classification Trees in Medicine

CTs have been used in medicine since their inception. The main tasks assigned to them include generating classification rules for diagnosis, selecting variables based on their importance, determining cut-off points for continuous variables, and identifying clinical relationships among variables [6]. CT algorithms select the most relevant variables, their order of appearance in tree branching, and the optimal cut-off points [3]. The interpretability of classification rules makes CTs attractive for use in clinical settings [7].
A review of bibliographic databases reveals an exponential increase in publications using CT methodology, supporting their ongoing relevance in medical problems [7]. In the past two decades, the widespread adoption of machine learning techniques, including CTs, has further promoted their use [8].
We reference several studies that develop risk models using CTs and provide clear explanations of their methodological construction, such as a model for serious fall injury in older adults [9], or risk stratification in critically ill patients [10]. CTs have also been applied in nutrition, such as in malnutrition detection [11], identifying the relationship between frailty and diet quality indicators [12], or predicting dropout from psychological treatment in bariatric surgery candidates [13].

1.4. Types of Classification Trees

There are many types of CTs [2,3]. Broadly, they can be divided into three main types (see Table 1): simple models that generate a single CT, ensemble models that use multiple CTs to improve accuracy, and hybrid models that combine CTs with other machine learning techniques, such as fuzzy logic or artificial neural networks [14,15].
Different types of simple CTs vary based on their stopping criteria, pruning methods, and procedures for selecting the optimal tree [10]. The most commonly used simple models include CART, CHAID, C4.5, and ctree [16,17,18,19]. Table 1 shows their specifications and the available software for implementation [20,21,22,23]. Notably, CART-type trees perform well with small datasets, which explains their continued use in limited data scenarios [10].
Other types of simple CTs, such as FACT (Fast and Accurate Classification Tree), QUEST (Quick Unbiased and Efficient Statistical Tree), CRUISE (Classification Rule with Unbiased Interaction Selection and Estimation), GUIDE (Generalized Unbiased Interaction Detection and Estimation), and Bayesian trees, have shown specific advantages in particular patient groups [3].
Table 1. Types of classification trees.
Table 1. Types of classification trees.
1. Simple Classification Trees
CARTCHAIDC4.5ctree
DescriptionClassification and regression treeChi-square automatic interaction detectionConcept Learning Systems
Version 4.5
Conditional inference trees
DeveloperBreiman (1984)
[16]
Kass (1980)
[17]
Quinlan (1993) [18]Hothorm (2006)
[19]
Primary UseMany disciplines with few dataApplied statisticiansData minersApplied statisticians
Splitting MethodEntropy
Gini index
Chi-square tests
F test
Gain ratioAsymptotic approximations
Branch LimitationsBest binary splitNumber of values of the inputBest binary splitBonferroni-adjusted p-values
PruningCross-validationBest binary split
p-value
Misclassification ratesNo pruning
Software *AnswerTree
WEKA
R-Python
AnswerTree
R-Python
WEKA
R-Python
R-Python
2. Ensembled Classification Trees
Random ForestAdaBoostXG-Boost
DescriptionUncorrelated forestAdaptive boostingExtreme gradient boosting
DeveloperBreiman and Cutler (2001)
[24]
Freund and Schapire (1995) [25]Chen and Guestrin (2016)
[26]
Ensembled MethodParallel
bagging
Adaptive
boosting
Boosting
gradient descent
Software *R/Python/JavaR/Python/JavaR/Python/Java
3. Hybrid Models
Fuzzy Random ForestRandom Forest Neural Network
Other methodFuzzy logicNeural network
DeveloperOlaru (2003)
[15]
Khozeimeh (2022)
[14]
Software *C languagePython
Modified from [2,3,5,7]. * References of available software [20,21,22,23].
In general, model fitting parameters—called hyperparameters—include the tree’s maximum depth (branching levels), the minimum number of cases per terminal node, and the internal validation method used (e.g., training/validation split or cross-validation). Each simple CT model may require specific additional hyperparameters for proper functioning [2,3,9].
Hybrid models that combine CTs with other machine learning techniques are also an expanding area. A separate, dedicated review would be necessary to explore their methodologies in depth. In Table 1, we included only two examples of hybrid models: one that combines fuzzy logic with CTs [15], and another that incorporates artificial neural networks [14].

1.5. Advantages and Disadvantages of Classification Trees

Although we have already mentioned several advantages and disadvantages of CTs, they can be summarized as follows [2,3,4,5,6]:

1.5.1. Advantages

  • Non-parametric models.
  • Can handle all variable types (continuous, ordinal, categorical).
  • Easy to interpret, with clinically meaningful classification rules.
  • No additional calculations required to determine individual patient risk.
  • Perform variable selection and establish variable hierarchy.
  • Identify optimal cut-off points for continuous variables.
  • Detect relationships among variables without assuming independence.
  • Less affected by outliers or missing values.

1.5.2. Disadvantages

  • Risk of overfitting and limited generalizability.
  • High sensitivity to data, leading to model instability.
  • Complex trees may lose interpretability.
  • Require specific software and development methodology.
  • Many CT types exist, and the most suitable one for a specific problem may not be obvious in advance.
In summary, it is necessary to strike a balance between the advantage of interpretability through classification rules and the methodological rigor required for their use.

2. ENPIC Study and the Obesity Paradox

We next describe the dataset used in this study, conducting a post hoc analysis of the ENPIC (Evaluation of Practical Nutrition Practices in the Critical Care Patient) study, applying classification tree (CT) methodology to explore the obesity paradox [27].

2.1. The ENPIC Study

Patients admitted to intensive care units (ICUs) often require artificial nutrition. The goal of artificial nutrition is to achieve optimal caloric and protein intake. Studies such as ENPIC aim to address questions related to improving this process [28,29,30].
The objective of the ENPIC study was to evaluate compliance with recommendations for specialized nutritional–metabolic therapy in ICU patients and to assess the impact of nutritional therapy on mortality in this population [29]. This was a multicenter study involving a sample of 525 ICU patients who required artificial nutrition, either enteral or parenteral.
One of this study’s conclusions was that we are far from achieving caloric and protein goals in these patients, particularly in those with obesity, which negatively influences outcomes, including increased mortality [28,29,30].

2.2. The Obesity Paradox

Obesity is generally considered a risk factor for numerous medical conditions. However, the idea of a paradoxical protective effect of obesity in certain diseases has been discussed for over 20 years. The concept of the “obesity paradox” was first described in patients undergoing percutaneous coronary interventions, where improved survival was observed in overweight and obese individuals [31]. The obesity paradox suggests that obesity may confer a protective effect in some clinical contexts, leading to better survival outcomes [31]. Even a meta-analysis in patients with heart failure found some evidence supporting this paradox [32].
Its presence in ICU patients remains controversial [33]. Some authors argue that the phenomenon may be attributable to selection bias and confounding, arising from inadequate adjustment for variables influencing the obesity–mortality relationship [34,35,36].
A study published in this journal using the ENPIC dataset—Nutrition Therapy in Critically Ill Patients with Obesity—reported differences among BMI groups in some of the analyzed characteristics, as shown in Table 2 [27].
This table presents the general characteristics, nutritional therapy, and outcomes of ICU patients stratified by body mass index (BMI) categories. It reveals demographic differences across BMI groups (normal weight, overweight, and obese). There is a non-significant trend toward lower 28-day mortality in obese patients, raising the question of whether the paradox might be present in this cohort.
The findings suggest that building a mortality risk model requires adjustment for patient-related factors (age, sex, BMI group), nutritional status based on subjective global assessment (SGA), disease severity according to the APACHE II (Acute Physiology and Chronic Health disease Classification System II) score [37], and daily caloric (Kcal/kg/day) and protein (g/kg/day) intake. If the paradox is present, obesity would emerge as a protective factor.
We performed a multiple logistic regression (LR) model using the selected variables and calculated odds ratios (95% CI). Table 3 shows the model, which indicates that factors independently associated with 28-day mortality are older age, higher APACHE II score, malnutrition, and lower protein intake.
The obesity paradox was not detected in this model, as the odds ratio for obesity was not statistically significant. Given the limited sample size and the LR model used, we cannot confirm the presence of the paradox. Classification trees may uncover interactions or relationships not easily detected through traditional regression methods.
In the following sections, we apply CTs to the ENPIC dataset to illustrate methodological aspects and explore the potential presence of the obesity paradox.

3. Use of Classification Trees to Determine Cut-Off Points

Continuous variables cannot be arbitrarily categorized. Typically, literature-based criteria or statistical methods are used to justify selected cut-off points.
Classification trees can be used to identify cut-off points in continuous variables and convert them into categorical ones—for example, to define score thresholds for diagnosing heart failure in patients with pleural effusion [38].
Other examples in nutrition-related contexts include determining a TNF-α (Tumor Necrosis Factor-alpha) cut-off related to HbA1c levels using a CHAID tree [39] or defining walking speed thresholds indicative of severe mobility limitations in sarcopenic patients using a CART tree [40].
In our paradox example, we employed a CHAID tree to determine cut-off points for age and APACHE II score. Figure 1 shows that the selected age and APACHE II groups for categorization can be justified using classification trees.

4. Use of Classification Trees to Identify Relationships Between Variables

As mentioned earlier, one advantage of classification trees is their ability to detect relationships between variables. These relationships are often non-linear. Identifying such associations without assuming independence—and grouping patients based on these patterns—is essential for understanding how risk factors operate differently across subgroups [33].
In our example, previous research has suggested that exploring the obesity paradox requires stratifying patients by nutritional status [33]. In Figure 2, using a CART model, we observe that the risk of mortality is influenced by malnutrition. Among patients without malnutrition, the obese subgroup shows lower mortality, which may suggest that the paradoxical effect is detectable only in non-malnourished patients.

5. Multivariable Risk Models Using Classification Trees

A multivariable classification tree (CT) model can select only the most informative variables, establish a hierarchy, determine optimal cut-off points to categorize continuous variables, and indicate relationships between variables, resulting in a set of classification rules [7,10].
In the obesity paradox example, we included in the CT the same variables used in the logistic regression (LR) model (see Table 2). The CART-type classification tree model discarded caloric intake and sex. Figure 3 shows the classification tree with its classification rules. The first variable selected by the algorithm is age; the subsequent branches include malnutrition status, disease severity, protein intake, and BMI groups. Although not directly related to the paradox, it is interesting to note that in the subgroup of older patients with malnutrition, higher protein intake is associated with lower mortality.
Also highlighted in a red circle is a subgroup of patients in which a paradoxical effect appears to be present. An analysis of this specific group of 62 patients—23 obese and 39 non-obese—found statistically significant differences only in mortality (21.7% vs. 48.7%, p = 0.035) and the prevalence of hypertension (78.3% vs. 51.3%, p = 0.035). No significant differences were observed in age, sex, patient type, time to initiation of artificial nutrition, severity level, type of nutritional support (enteral or parenteral), or caloric and protein intake. This subgroup shows a significant reduction in mortality among obese patients compared to non-obese, suggesting the obesity paradox may exist in specific contexts. The conclusion of this CT model is that the obesity paradox does not exist in general but may be observed in specific patient groups, which can be identified using CT-based methodologies.

6. Ensemble Classification Tree Models

As previously noted, one disadvantage of simple CT models is their tendency to overfit, which limits their generalizability. To mitigate this issue, the idea arose of using not just one CT but rather an ensemble of trees to improve both the precision and generalizability of risk models [3].
Ensembles of CTs can be built in various ways (see Table 1), with the two most common being bagging and boosting [3,41].
Bagging (bootstrap aggregating) uses parallel ensemble learning to reduce model variance, averaging the results of individual CTs to produce a final prediction [42]. Each individual tree operates independently, allowing for fast performance. A representative of this method is the random forest algorithm [24,43].
In contrast, boosting involves sequential ensemble learning, where each CT is dependent on the previous ones. Boosting reduces bias and is considered “slow learning” compared to bagging. By adding one CT after another, the algorithm progressively improves classification. Popular boosting-based ensemble models include AdaBoost and gradient-based methods like XGBoost [44,45]. Boosting’s flexibility and strong performance have made it one of the most widely used techniques in recent years [45].
Although ensemble models outperform single-tree models in prediction accuracy, they come at the cost of reduced interpretability. This “black-box” nature has led to the development of methods to explain how these models work, providing clinicians with insights into variable importance and relationships [45].
One of the most used explanatory techniques is SHAP (SHapley Additive exPlanation) values. Based on cooperative game theory, SHAP values determine the contribution of each “player” (variable) to the final model outcome. A SHAP value reflects each variable’s contribution to the prediction, and results are visualized using importance plots. SHAP also generates partial dependence plots showing the relationship between each variable and the model outcome [46].
In the obesity paradox example, we use an XGBoost and SHAP model for explanation. Using ensemble methods like XGBoost involves tuning many hyperparameters, which increases complexity [47]. Some of the hyperparameters used in our XGBoost model included learning rate (eta) set at 0.3, number of trees (n_estimators) set at 20, gamma (loss reduction threshold) at 0, and maximum tree depth (max_depth) at 6.
The same variables used in the LR and CART models were included. Figure 4 shows the SHAP values.
The graph provides a visual explanation of the model, showing the influence of each variable on the studied outcome, which in this case is mortality. It displays the relative importance of variables based on their SHAP scores. The association with mortality is also indicated, with values on the right linked to higher mortality and those on the left to lower mortality. The variable categories are color-coded, enabling the identification of clear groupings associated with increased or decreased risk of the event.
The XGBoost model identified age as the most important variable, followed by caloric intake (higher intake = worse outcome), protein intake (higher intake = better outcome), and BMI group. Within the BMI analysis, the obese group showed an association with improved survival.
Figure 5 shows the partial dependence plots. These graphs illustrate the individual impact of each factor on the studied outcome. For example, the age groups demonstrate that higher age corresponds to higher SHAP values, indicating a stronger association with mortality. In contrast, the BMI groups reveal that obese individuals (highlighted with a red circle) have lower SHAP values, indicating a weaker association with mortality.
Both figures indicate that being obese is associated with a protective effect on mortality.

7. Model Evaluation

The methodology for evaluating a risk model is also very extensive and could, by itself, constitute an independent review. However, it is important to note that the development, validation, and assessment of the predictive accuracy of a risk model should follow standardized procedures, such as the TRIPOD (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis) guidelines, which provide guidance on all the necessary steps from model construction to evaluation [48].
Our LR models, the CART-type decision tree, and the XGBoost ensemble model generate a probability of death for each patient. We can assess model accuracy using global performance metrics (Brier score, accuracy, recall, f-measure), discrimination (AUC ROC), and calibration (calibration curve with slope and intercept values) [10,43,49].
For example, Figure 6 illustrates the discriminatory ability of the models through the ROC curve and the corresponding area under the curve (AUC). We observe that the XGBoost ensemble model achieves the highest AUC, while the LR and CART models yield similar values.

8. Discussion

In this article, we describe how classification trees (CT) can be used in clinical research as a complementary tool to traditional statistical techniques.
Using the obesity paradox as an example, we explored the methodology based on classification trees. One of the main advantages of simple CTs is their interpretability [10]. This feature, combined with the ability to identify patient subgroups—defined by classification rules—with distinct characteristics, makes them appealing to clinical practitioners [2,3].
However, a key disadvantage of CTs is the risk of overfitting and their strong dependence on the specific dataset used for model construction, which can lead to instability issues [6]. To address these limitations, ensemble CTs have been developed, achieving higher accuracy and better generalization performance [24,25,26]. Nevertheless, this improved performance often comes at the cost of reduced interpretability—an issue also common in more complex machine learning methods and other artificial intelligence tools [49,50].
Although complex models tend to lack transparency, several techniques have emerged to help interpret their inner workings. In our work, we applied SHAP (SHapley Additive exPlanations), which provides insights into variable importance and reveals partial relationships between individual risk factors and the predicted outcome [51,52].
Our work is not without limitations. The depth of the methodological review could have been greater, although we believe that the references provided may help interested readers to achieve that objective. The example used to illustrate the methodological review involved a simple selection of variables to facilitate the generation, interpretation, and explanation of the classification tree (CT) models. In model construction, the inclusion of more variables or the use of alternative methodologies—both simple and ensemble-based—could have led to different results and conclusions. Moreover, the models we have developed should undergo external validation to demonstrate their generalizability.
This suggests that CTs are especially useful in uncovering clinically relevant patterns that could guide personalized medicine approaches. Nevertheless, caution should be taken when interpreting these findings, as tree-based models are sensitive to data characteristics and may require validation in independent cohorts to confirm their generalizability.
We can state that classification tree–based methodologies remain an important area of experimentation and are constantly evolving technologically [53,54]. Nutrition professionals working with risk models should consider the use of CTs. Simple trees remain appealing due to their interpretability, while more sophisticated models must continue to evolve to become more user-friendly [55]. We also offer our collaboration, sharing our experience with other research groups interested in developing CT-based models.

9. Conclusions

The application of CT methodology in our example, based on the ENPIC study database, has allowed for us to illustrate several important aspects:
  • How CTs can be used to establish cut-off points for continuous variables.
  • How they can identify interactions between variables that might go unnoticed in traditional regression models.
  • How multivariable CT models generate decision rules and stratify patient risk based on the most influential predictors.
  • How ensemble methods such as random forest and XGBoost improve predictive accuracy.
  • And finally, how explanatory tools like SHAP values can provide insight into the structure and predictions of complex models.
Although we did not find strong evidence to confirm the existence of the obesity paradox in critically ill patients in a general sense, CT-based analysis helped us identify a specific subgroup of patients in which obesity appeared to be associated with reduced mortality.
In conclusion, CTs—especially when combined with ensemble methods and model explanation techniques—are valuable tools in clinical epidemiology. They can support both predictive modelling and the exploration of complex relationships among clinical variables, offering new perspectives to improve patient care.

Author Contributions

Conceptualization, J.T., L.S., M.B., J.C.E.S. and J.C.L.-D.; methodology, J.T., L.S. and J.C.E.S.; software, J.T. and J.C.E.S.; analysis of results, M.L.B.-L., C.L., C.V., J.L.F.-L., I.M.d.L., E.P.-R. and J.C.L.-D.; writing—review and editing, J.T., L.S., M.B., J.C.E.S., J.C.L.-D., M.L.B.-L., C.L., C.V., J.L.F.-L., I.M.d.L. and E.P.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data in this study are available on request from the author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AdaBoostadaptive boosting
APACHE IIAcute Physiology and Chronic Health disease Classification System II
AUC ROCarea under the ROC curve
BMIbody mass index
C4.5Concept Learning Systems Version 4.5
CARTclassification and regression tree
CHAIDchi-square automatic interaction detection tree
CRUISEClassification Rule with Unbiased Interaction Selection and Estimation
CTclassification tree
ctreeconditional inference trees
ENPICEvaluation of Practical Nutrition Practices in the Critical Care Patient
FACTFast and Accurate Classification Tree
GUIDEGeneralized Unbiased Interaction Detection and Estimation
ICUintensive care unit
LRlogistic regression
PythonPython software
QUESTQuick Unbiased and Efficient Statistical Tree
RThe R Project for Statistical Computing
SHAPSHapley Additive exPlanations
TNF-αTumor Necrosis Factor-alpha
TRIPODTransparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis
WEKAWaikato Environment for Knowledge Analysis
XGBoostextreme gradient boosting

References

  1. Masic, I. Medical Decision Making—An Overview. Acta Inform. Med. 2022, 30, 230–235. [Google Scholar] [CrossRef] [PubMed]
  2. Podgorelec, V.; Kokol, P.; Stiglic, B.; Rozman, I. Decision trees: An overview and their use in medicine. J. Med. Syst. 2002, 26, 445–463. [Google Scholar] [CrossRef]
  3. Wei-Yin, L. Fifty Years of Classification and Regression Trees. Int. Stat. Rev. 2014, 82, 329–348. [Google Scholar] [CrossRef]
  4. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  5. Trujillano, J.; Sarria-Santamera, A.; Esquerda, A.; Badia, M.; Palma, M.; March, J. Approach to the methodology of classification and regression trees. Gac. Sanit. 2008, 22, 65–72. [Google Scholar] [CrossRef] [PubMed]
  6. Mitchell, T. (Ed.) Decision tree learning. In Machine Learning; McGraw Hill: New York, NY, USA, 1997. [Google Scholar]
  7. Lemon, S.C.; Roy, J.; Clark, M.A.; Friedmann, P.D.; Rakowski, W. Classification and regression tree analysis in public health: Methodological review and comparison with logistic regression. Ann. Behav. Med. 2003, 26, 172–181. [Google Scholar] [CrossRef]
  8. Shang, H.; Ji, Y.; Cao, W.; Yi, J. A predictive model for depression in elderly people with arthritis based on the TRIPOD guidelines. Geriatr. Nurs. 2025, 29, 85–93. [Google Scholar] [CrossRef]
  9. Speiser, J.L.; Callahan, K.E.; Houston, D.K.; Fanning, J.; Gill, T.M.; Guralnik, J.M.; Newman, A.B.; Pahor, M.; Rejeski, W.J.; Miller, M.E. Machine Learning in Aging: An Example of Developing Prediction Models for Serious Fall Injury in Older Adults. J. Gerontol. A Biol. Sci. Med. Sci. 2021, 76, 647–654. [Google Scholar] [CrossRef] [PubMed]
  10. Trujillano, J.; Badia, M.; Serviá, L.; March, J.; Rodriguez-Pozo, A. Stratification of the severity of critically ill patients with classification trees. BMC Med. Res. Methodol. 2009, 9, 83. [Google Scholar] [CrossRef]
  11. Yin, L.; Lin, X.; Liu, J.; Li, N.; He, X.; Zhang, M.; Guo, J.; Yang, J.; Deng, L.; Wang, Y.; et al. Investigation on Nutrition Status and Clinical Outcome of Common Cancers (INSCOC) Group. Classification Tree-Based Machine Learning to Visualize and Validate a Decision Tool for Identifying Malnutrition in Cancer Patients. JPEN J. Parenter. Enteral Nutr. 2021, 45, 736–1748. [Google Scholar] [CrossRef]
  12. Tay, E.; Barnett, D.; Rowland, M.; Kerse, N.; Edlin, R.; Waters, D.L.; Connolly, M.; Pillai, A.; Tupou, E.; The, R. Sociodemographic and Health Indicators of Diet Quality in Pre-Frail Older Adults in New Zealand. Nutrients. 2023, 15, 4416. [Google Scholar] [CrossRef] [PubMed]
  13. Marchitelli, S.; Mazza, C.; Ricci, E.; Faia, V.; Biondi, S.; Colasanti, M.; Cardinale, A.; Roma, P.; Tambelli, R. Identification of Psychological Treatment Dropout Predictors Using Machine Learning Models on Italian Patients Living with Overweight and Obesity Ineligible for Bariatric Surgery. Nutrients 2024, 16, 2605. [Google Scholar] [CrossRef]
  14. Khozeimeh, F.; Sharifrazi, D.; Izadi, N.H.; Joloudari, J.H.; Shoeibi, A.; Alizadehsani, R.; Tartibi, M.; Hussain, S.; Sani, Z.A.; Khodatars, M.; et al. RF-CNN-F: Random forest with convolutional neural network features for coronary artery disease diagnosis based on cardiac magnetic resonance. Sci. Rep. 2022, 12, 11178. [Google Scholar] [CrossRef] [PubMed]
  15. Olaru, C.; Wehenkel, L. A complete fuzzy decision tree technique. Fuzzy Sets Syst. 2003, 138, 221–254. [Google Scholar] [CrossRef]
  16. Breiman, L.; Friedman, J.H.; Olshen, R.A.; Stone, C.J. Classification and Regression Trees; Wadsworth Publishing Co.: Belmont, CA, USA, 1984. [Google Scholar]
  17. Kass, G.V. An exploratory for investigating large quantities of categorical data. Ann. Appl. Stat. 1980, 29, 119–127. [Google Scholar] [CrossRef]
  18. Quinlan, J.R. C4.5: Programs for Machine Learning; Morgan Kaufmann: San Mateo, CA, USA, 1993. [Google Scholar]
  19. Hothorn, T.; Hornik, K.; Zeileis, A. Unbiased Recursive Partitioning: A Conditional Inference Framework. J. Comput. Graph. Stat. 2006, 15, 651–674. [Google Scholar] [CrossRef]
  20. SPSS Inc. AnswerTree User’s Guide; SPSS Inc.: Chicago, IL, USA, 2001. [Google Scholar]
  21. Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H. The WEKA data mining software. ACM SIGKDD Explor. Newslett. 2009, 11, 10–18. [Google Scholar] [CrossRef]
  22. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef]
  23. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2021; Available online: https://www.R-project.org/ (accessed on 1 January 2023).
  24. Breiman, L. Random Forest. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  25. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  26. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ‘16), Association for Computing Machinery, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  27. Lopez-Delgado, J.C.; Sanchez-Ales, L.; Flordelis-Lasierra, J.L.; Mor-Marco, E.; Bordeje-Laguna, M.L.; Portugal-Rodriguez, E.; Lorencio-Cardenas, C.; Vera-Artazcoz, P.; Aldunate-Calvo, S.; Llorente-Ruiz, B.; et al. Nutrition Therapy in Critically Ill Patients with Obesity: An Observational Study. Nutrients 2025, 17, 732. [Google Scholar] [CrossRef]
  28. Yébenes, J.C.; Bordeje-Laguna, M.L.; Lopez-Delgado, J.C.; Lorencio-Cardenas, C.; Martinez De Lagran Zurbano, I.; Navas-Moya, E.; Servia-Goixart, L. Smartfeeding: A Dynamic Strategy to Increase Nutritional Efficiency in Critically Ill Patients-Positioning Document of the Metabolism and Nutrition Working Group and the Early Mobilization Working Group of the Catalan Society of Intensive and Critical Care Medicine (SOCMiC). Nutrients 2024, 16, 1157. [Google Scholar] [CrossRef]
  29. Servia-Goixart, L.; Lopez-Delgado, J.C.; Grau-Carmona, T.; Trujillano-Cabello, J.; Bordeje-Laguna, M.L.; Mor-Marco, E.; Portugal-Rodriguez, E.; Lorencio-Cardenas, C.; Montejo-Gonzalez, J.C.; Vera-Artazcoz, P.; et al. Evaluation of Nutritional Practices in the Critical Care patient (The ENPIC study): Does nutrition really affect ICU mortality? Clin. Nutr. ESPEN 2022, 47, 325–332. [Google Scholar] [CrossRef] [PubMed]
  30. Cahill, N.E.; Heyland, D.K. Bridging the guideline-practice gap in critical care nutrition: A review of guideline implementation studies. J. Parenter Enteral Nutr. 2010, 34, 653–659. [Google Scholar] [CrossRef] [PubMed]
  31. Gruberg, L.; Weissman, N.J.; Waksman, R.; Fuchs, S.; Deible, R.; Pinnow, E.E.; Ahmed, L.M.; Kent, K.M.; Pichard, A.D.; Suddath, W.O.; et al. The impact of obesity on the short-term and long-term outcomes after percutaneous coronary intervention: The obesity paradox? J. Am. Coll. Cardiol. 2002, 39, 578–584. [Google Scholar] [CrossRef]
  32. Oreopoulos, A.; Padwal, R.; Kalantar-Zadeh, K.; Fonarow, G.C.; Norris, C.M.; McAlister, F.A. Body mass index and mortality in heart failure: A meta-analysis. Am. Heart J. 2008, 156, 13–22. [Google Scholar] [CrossRef]
  33. Robinson, M.K.; Mogensen, K.M.; Casey, J.D.; McKane, C.K.; Moromizato, T.; Rawn, J.D.; Christopher, K.B. The relationship among obesity, nutritional status, and mortality in the critically ill. Crit. Care Med. 2015, 43, 87–100. [Google Scholar] [CrossRef]
  34. Zhou, D.; Wang, C.; Lin, Q.; Li, T. The obesity paradox for survivors of critically ill patients. Crit. Care 2022, 26, 198. [Google Scholar] [CrossRef]
  35. Dickerson, R.N. The obesity paradox in the ICU: Real or not? Crit. Care 2013, 17, 154. [Google Scholar] [CrossRef]
  36. Ripoll, J.G.; Bittner, E.A. Obesity and Critical Illness-Associated Mortality: Paradox, Persistence and Progress. Crit. Care Med. 2023, 51, 551–554. [Google Scholar] [CrossRef]
  37. Knaus, W.A.; Zimmerman, J.E.; Wagner, D.P.; Draper, E.A.; Lawrence, D.E. APACHE-acute physiology and chronic health evaluation: A physiologically based classification system. Crit. Care Med. 1981, 9, 591–597. [Google Scholar] [CrossRef] [PubMed]
  38. Porcel, J.M.; Trujillano, J.; Porcel, L.; Esquerda, A.; Bielsa, S. Development and validation of a diagnostic prediction model for heart failure-related pleural effusions: The BANCA score. ERJ Open Res. 2025, in press. [CrossRef]
  39. Díaz-Prieto, L.E.; Gómez-Martínez, S.; Vicente-Castro, I.; Heredia, C.; González-Romero, E.A.; Martín-Ridaura, M.D.C.; Ceinos, M.; Picón, M.J.; Marcos, A.; Nova, E. Effects of Moringa oleifera Lam. Supplementation on Inflammatory and Cardiometabolic Markers in Subjects with Prediabetes. Nutrients 2022, 14, 1937. [Google Scholar] [CrossRef]
  40. Cawthon, P.M.; Patel, S.M.; Kritchevsky, S.B.; Newman, A.B.; Santanasto, A.; Kiel, D.P.; Travison, T.G.; Lane, N.; Cummings, S.R.; Orwoll, E.S.; et al. What Cut-Point in Gait Speed Best Discriminates Community-Dwelling Older Adults with Mobility Complaints from Those Without? A Pooled Analysis from the Sarcopenia Definitions and Outcomes Consortium. J. Gerontol. A Biol. Sci. Med. Sci. 2021, 76, e321–e327. [Google Scholar] [CrossRef] [PubMed]
  41. Tamanna, T.; Mahmud, S.; Salma, N.; Hossain, M.M.; Karim, M.R. Identifying determinants of malnutrition in under-five children in Bangladesh: Insights from the BDHS-2022 cross-sectional study. Sci. Rep. 2025, 15, 14336. [Google Scholar] [CrossRef]
  42. Turjo, E.A.; Rahman, M.H. Assessing risk factors for malnutrition among women in Bangladesh and forecasting malnutrition using machine learning approaches. BMC Nutr. 2024, 10, 22. [Google Scholar] [CrossRef]
  43. Reusken, M.; Coffey, C.; Cruijssen, F.; Melenberg, B.; van Wanrooij, C. Identification of factors associated with acute malnutrition in children under 5 years and forecasting future prevalence: Assessing the potential of statistical and machine learning methods. BMJ Public Health 2025, 3, e001460. [Google Scholar] [CrossRef]
  44. Li, N.; Peng, E.; Liu, F. Prediction of lymph node metastasis in cervical cancer patients using AdaBoost machine learning model: Analysis of risk factors. Am. J. Cancer Res. 2025, 5, 1158–1173. [Google Scholar] [CrossRef]
  45. Wong, J.E.; Yamaguchi, M.; Nishi, N.; Araki, M.; Wee, L.H. Predicting Overweight and Obesity Status Among Malaysian Working Adults with Machine Learning or Logistic Regression: Retrospective Comparison Study. JMIR Form. Res. 2022, 6, e40404. [Google Scholar] [CrossRef]
  46. Lian, R.; Tang, H.; Chen, Z.; Chen, X.; Luo, S.; Jiang, W.; Jiang, J.; Yang, M. Development and multi-center cross-setting validation of an explainable prediction model for sarcopenic obesity: A machine learning approach based on readily available clinical features. Aging Clin. Exp. Res. 2025, 37, 63. [Google Scholar] [CrossRef]
  47. Gu, Y.; Su, S.; Wang, X.; Mao, J.; Ni, X.; Li, A.; Liang, Y.; Zeng, X. Comparative study of XGBoost and logistic regression for predicting sarcopenia in postsurgical gastric cancer patients. Sci. Rep. 2025, 5, 12808. [Google Scholar] [CrossRef] [PubMed]
  48. Collins, G.S.; Reitsma, J.B.; Altman, D.G.; Moons, K.G. Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): The TRIPOD statement. Ann. Intern. Med. 2015, 62, 55–63. [Google Scholar] [CrossRef] [PubMed]
  49. Van Dort, B.A.; Engelsma, T.; Medlock, S.; Dusseljee-Peute, L. User-Centered Methods in Explainable AI Development for Hospital Clinical Decision Support: A Scoping Review. Stud. Health Technol. Inform. 2025, 326, 17–21. [Google Scholar] [CrossRef]
  50. Moser, D.; Sariyar, M. Explainable Versus Interpretable AI in Healthcare: How to Achieve Understanding. Stud. Health Technol. Inform. 2025, 327, 1433–1437. [Google Scholar] [CrossRef] [PubMed]
  51. Alkhanbouli, R.; Matar Abdulla Almadhaani, H.; Alhosani, F.; Simsekler, M.C.E. The role of explainable artificial intelligence in disease prediction: A systematic literature review and future research directions. BMC Med. Inform. Decis. Mak. 2025, 25, 110. [Google Scholar] [CrossRef]
  52. Sandri, E.; Cerdá Olmedo, G.; Piredda, M.; Werner, L.U.; Dentamaro, V. Explanatory AI Predicts the Diet Adopted Based on Nutritional and Lifestyle Habits in the Spanish Population. Eur. J. Investig. Health Psychol. Educ. 2025, 15, 11. [Google Scholar] [CrossRef]
  53. Harper, P.R. A review and comparison of classification algorithms for medical decision making. Health Policy. 2005, 71, 315–331. [Google Scholar] [CrossRef]
  54. El-Latif, E.I.A.; El-Dosuky, M.; Darwish, A.; Hassanien, A.E. A deep learning approach for ovarian cancer detection and classification based on fuzzy deep learning. Sci. Rep. 2024, 14, 26463. [Google Scholar] [CrossRef]
  55. Lundberg, S.M.; Nair, B.; Vavilala, M.S.; Horibe, M.; Eisses, M.J.; Adams, T.; Liston, D.E.; Low, D.K.; Newman, S.F.; Kim, J.; et al. Explainable machine-learning predictions for the prevention of hypoxaemia during surgery. Nat. Biomed. Eng. 2018, 2, 749–760. [Google Scholar] [CrossRef]
Figure 1. Using CHAID-type AC trees to establish cut-off points for continuous variables. (A) Cut-off points for age. (B) Cut-off points for the APACHE II severity score variable. R software (CHAID library) was used [23].
Figure 1. Using CHAID-type AC trees to establish cut-off points for continuous variables. (A) Cut-off points for age. (B) Cut-off points for the APACHE II severity score variable. R software (CHAID library) was used [23].
Nutrients 17 01903 g001
Figure 2. CART-type AC model to establish the relationship between two variables. Values in bold indicate 28-day mortality. SGA: subjective global assessment; BMI: body mass index group. It is observed that the non-malnutrition group of obese patients has a lower mortality rate. R software (rpart library) was used [23].
Figure 2. CART-type AC model to establish the relationship between two variables. Values in bold indicate 28-day mortality. SGA: subjective global assessment; BMI: body mass index group. It is observed that the non-malnutrition group of obese patients has a lower mortality rate. R software (rpart library) was used [23].
Nutrients 17 01903 g002
Figure 3. Multivariate model based on CART type. AGER: age groups; SGA: subjective global assessment; APA25: APACHE II score greater than 25; BMIG: body mass index group; PROTEIN: protein intake (g/kg/day). The hierarchy of variables is shown, and a special group in which obese patients have a lower mortality rate than those with overweight or a normal BMI is indicated in the red circle. AnswerTree 3.0 software was used [24].
Figure 3. Multivariate model based on CART type. AGER: age groups; SGA: subjective global assessment; APA25: APACHE II score greater than 25; BMIG: body mass index group; PROTEIN: protein intake (g/kg/day). The hierarchy of variables is shown, and a special group in which obese patients have a lower mortality rate than those with overweight or a normal BMI is indicated in the red circle. AnswerTree 3.0 software was used [24].
Nutrients 17 01903 g003
Figure 4. Graph showing SHAP scores generated by the XGBoost model. AGER: age groups; KCAL: calorie intake (Kcal/kg/day); PROTEIN: protein intake (g/kg/day); SGA: subjective global assessment; APA25: APACHE II score greater than 25; BMIG: body mass index group. The importance of each variable can be observed, with age being the greatest. It is also observed that the obese patient group has a negative impact on mortality, suggesting a paradoxical effect of obesity. R software (xgboost and SHAPforxgboost libraries) was used [23].
Figure 4. Graph showing SHAP scores generated by the XGBoost model. AGER: age groups; KCAL: calorie intake (Kcal/kg/day); PROTEIN: protein intake (g/kg/day); SGA: subjective global assessment; APA25: APACHE II score greater than 25; BMIG: body mass index group. The importance of each variable can be observed, with age being the greatest. It is also observed that the obese patient group has a negative impact on mortality, suggesting a paradoxical effect of obesity. R software (xgboost and SHAPforxgboost libraries) was used [23].
Nutrients 17 01903 g004
Figure 5. Partial dependency graphs between each variable and mortality. AGER: age groups; KCAL: calorie intake (kcal/kg/day); PROTEIN: protein intake (g/kg/day); SGA: subjective global assessment; APA25: APACHE II score greater than 25; BMIG: body mass index groups. We observed different behavior between calorie and protein intake. Within the BMI groups, mortality was lower in the obese group. R software (SHAPforxgboost library) was used [23].
Figure 5. Partial dependency graphs between each variable and mortality. AGER: age groups; KCAL: calorie intake (kcal/kg/day); PROTEIN: protein intake (g/kg/day); SGA: subjective global assessment; APA25: APACHE II score greater than 25; BMIG: body mass index groups. We observed different behavior between calorie and protein intake. Within the BMI groups, mortality was lower in the obese group. R software (SHAPforxgboost library) was used [23].
Nutrients 17 01903 g005
Figure 6. ROC analysis of the developed models. The area under the ROC curve values are shown. LR: logistic regression model; CART: CART-type classification tree; XGBoost: XGBoost-type ensemble classification tree model; Prob: probability of death at 28 days. The LR and CART-type AC models had similar AUCs. The XGBoost model achieved better discrimination (Long’s test with p < 0.001).
Figure 6. ROC analysis of the developed models. The area under the ROC curve values are shown. LR: logistic regression model; CART: CART-type classification tree; XGBoost: XGBoost-type ensemble classification tree model; Prob: probability of death at 28 days. The LR and CART-type AC models had similar AUCs. The XGBoost model achieved better discrimination (Long’s test with p < 0.001).
Nutrients 17 01903 g006
Table 2. General characteristics, nutritional therapy, and outcomes of patients admitted to the ICU based on body mass index subgroup. Results of the ENPIC study [20].
Table 2. General characteristics, nutritional therapy, and outcomes of patients admitted to the ICU based on body mass index subgroup. Results of the ENPIC study [20].
All Patients
n = 525
Normal
n = 165
Overweight
n = 210
Obese
n = 150
p-Value
Baseline characteristics and comorbidities
Age, years, mean ± SD61.5 ± 1558.8 ± 16.562.8 ± 14.762.7 ± 13.50.05
Sex, male patients, n (%)67.2% (353)64.8% (107)74.8% (157)59.3% (89)0.003 B
Hypertension, n (%)43.6% (229)33.9% (56)41.9% (88)56.7% (85)0.01 A,B
Diabetes mellitus, n (%)25% (131)21.2% (35)20% (42)36% (54)0.001 A,B
AMI, n (%)14.1% (74)8.5% (14)16.7% (35)16.7% (25)0.04 B
Neoplasia, n (%)20.6% (108)24.2% (40)19.5% (41)18% (27)0.11
Type of patientMedical, n (%)63.8% (335)65.5% (108)62.9% (132)63.3% (95)0.81
Trauma, n (%)12.6% (66)10.9% (18)15.2% (32)10.7% (16)0.75
Surgery, n (%)23.6% (124)23.6% (39)21.9% (46)26% (39)0.67
Prognosis ICU scores and nutrition status on ICU admission
APACHE II, mean ± SD20.3 ± 7.919.7 ± 7.620.1 ± 7.521.2 ± 8.50.18
Malnutrition (based on SGA), n (%)41% (215)52.7% (87)37.1% (78)33.3% (50)0.01 B
Characteristics of Medical Nutrition Therapy
Early nutrition, <48 h, n (%)74.9% (393)77.6% (128)75.2% (158)71.3% (107)0.43
Kcal/kg/day, mean ± SD19 ± 5.623.1 ± 618.6 ± 3.715.27 ± 4.240.001 A
Protein, g/kg/day, mean ± SD1 ± 0.41.2 ± 0.41 ± 0.30.8 ± 0.20.01 A,B
EN63.2% (332)59.4% (98)64.3% (135)66% (99)0.34
PN15.4% (81)13.3% (22)16.2% (34)16.7% (25)0.85
EN-PN7.8% (41)8.5% (14)7.6% (16)7.3% (11)0.92
PN-EN13.5% (71)18.8% (31)11.9% (25)10% (15)0.27
Outcomes
Mechanical ventilation, n (%)92.8% (487)89.1% (147)93.8% (197)95.3% (143)0.08
Vasoactive drug support, n (%)77% (404)73.9% (122)79.5% (167)76.7% (115)0.44
Renal replacement therapy, n (%)16.6% (87)16.4% (27)12.9% (27)22% (33)0.07
ICU stay, days, mean ± SD20.3 ± 1818.2 ± 13.821.1 ± 17.121.6 ± 22.50.08
28-day mortality, n (%)26.7% (140)29.1% (48)27.1% (57)23.3% (35)0.51
AMI: acute myocardial infarction; PN: parenteral nutrition; EN: enteral nutrition; SD: standard deviation; APACHE II: Acute Physiology and Chronic Health Disease Classification System II; SGA: subjective global assessment; ICU: intensive care unit. Statistically significant p-values are written in bold. Statistical results correspond to ANOVA p values. Bonferroni post hoc testing with statistically significant differences A between normal weight and obese subgroup; B between overweight and obese subgroup.
Table 3. Multiple logistic regression model (LR) with 28-day mortality as the outcome variable.
Table 3. Multiple logistic regression model (LR) with 28-day mortality as the outcome variable.
VariableOR (95% CI)p-Value
Age (years)
<501
50–753.3 (1.7–6.5)0.001
>757.0 (3.3–14.9)<0.001
Sex (Male)1.0 (0.6–1.6)0.998
APACHE II score
≤251
>252.2 (1.4–3.5)<0.001
BMI groups
Normal1
Overweight0.9 (0.5–1.5)0.623
Obese0.7 (0.4–1.4)0.651
SGA
Non-malnutrition1
Malnutrition2.6 (1.74.0)<0.001
Median Kcal/Kg/day1.1 (0.9–1.1)0.315
Median g protein/Kg/day0.3 (0.1–0.9)0.022
OR: odds ratio; CI: confidence interval; APACHE II: Acute Physiology and Chronic Health Disease Classification System II; BMI: body mass index; SGA: subjective global assessment.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Trujillano, J.; Serviá, L.; Badia, M.; Serrano, J.C.E.; Bordejé-Laguna, M.L.; Lorencio, C.; Vaquerizo, C.; Flordelis-Lasierra, J.L.; Martínez de Lagrán, I.; Portugal-Rodríguez, E.; et al. Methodological Review of Classification Trees for Risk Stratification: An Application Example in the Obesity Paradox. Nutrients 2025, 17, 1903. https://doi.org/10.3390/nu17111903

AMA Style

Trujillano J, Serviá L, Badia M, Serrano JCE, Bordejé-Laguna ML, Lorencio C, Vaquerizo C, Flordelis-Lasierra JL, Martínez de Lagrán I, Portugal-Rodríguez E, et al. Methodological Review of Classification Trees for Risk Stratification: An Application Example in the Obesity Paradox. Nutrients. 2025; 17(11):1903. https://doi.org/10.3390/nu17111903

Chicago/Turabian Style

Trujillano, Javier, Luis Serviá, Mariona Badia, José C. E. Serrano, María Luisa Bordejé-Laguna, Carol Lorencio, Clara Vaquerizo, José Luis Flordelis-Lasierra, Itziar Martínez de Lagrán, Esther Portugal-Rodríguez, and et al. 2025. "Methodological Review of Classification Trees for Risk Stratification: An Application Example in the Obesity Paradox" Nutrients 17, no. 11: 1903. https://doi.org/10.3390/nu17111903

APA Style

Trujillano, J., Serviá, L., Badia, M., Serrano, J. C. E., Bordejé-Laguna, M. L., Lorencio, C., Vaquerizo, C., Flordelis-Lasierra, J. L., Martínez de Lagrán, I., Portugal-Rodríguez, E., & López-Delgado, J. C. (2025). Methodological Review of Classification Trees for Risk Stratification: An Application Example in the Obesity Paradox. Nutrients, 17(11), 1903. https://doi.org/10.3390/nu17111903

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop