Previous Article in Journal
Rigorous Asymptotic Perturbation Bounds for Hermitian Matrix Eigendecompositions
Previous Article in Special Issue
Successful Management of Public Health Projects Driven by AI in a BANI Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Management of Severe COVID-19 Diagnosis Using Machine Learning

1
Department of Family Medicine, Bukovinian State Medical University, 58012 Chernivtsi, Ukraine
2
Department of Infectious Diseases and Epidemiology, Bukovinian State Medical University, 58012 Chernivtsi, Ukraine
3
Department of Management and Accounting, DTI University, 533/20, 018 41 Dubnica nad Váhom, Slovakia
4
Department of Artificial Intelligence, Lviv Polytechnic National University, 79005 Lviv, Ukraine
5
Department of General Surgery №2, Bukovinian State Medical University, 58012 Chernivtsi, Ukraine
*
Author to whom correspondence should be addressed.
Computation 2025, 13(10), 238; https://doi.org/10.3390/computation13100238
Submission received: 14 September 2025 / Revised: 3 October 2025 / Accepted: 4 October 2025 / Published: 9 October 2025
(This article belongs to the Special Issue Artificial Intelligence Applications in Public Health: 2nd Edition)

Abstract

COVID-19 remains a global health challenge, with severe cases often leading to complications and fatalities. The objective of this study was to assess supervised machine learning algorithms for predicting severe COVID-19 based on demographic, clinical, biochemical, and genetic variables, with the aim of identifying the most informative prognostic markers. For Machine Learning (ML) analysis, we utilized a dataset comprising 226 observations with 68 clinical, biochemical, and genetic features collected from 226 patients with confirmed COVID-19 (54—moderate, 142—severe and 30 with mild disease). The target variable was disease severity (mild, moderate, severe). The feature set included demographic variables (age, sex), genetic markers (single-nucleotide polymorphisms (SNPs) in FGB (rs1800790), NOS3 (rs2070744), and TMPRSS2 (rs12329760)), biochemical indicators (IL-6, endothelin-1, D-dimer, fibrinogen, among others), and clinical parameters (blood pressure, body mass index, comorbidities). The target variable was disease severity. To identify the most effective predictive models for COVID-19 severity, we systematically evaluated multiple supervised learning algorithms, including logistic regression, k-nearest neighbors, decision trees, random forest, gradient boosting, bagging, naïve Bayes, and support vector machines. Model performance was assessed using accuracy and the area under the receiver operating characteristic curve (AUC-ROC). Among the predictors, IL-6, presence of depression/pneumonia, LDL cholesterol, AST, platelet count, lymphocyte count, and ALT showed the strongest correlations with severity. The highest predictive accuracy, with negligible error rates, was achieved by ensemble-based models such as ExtraTreesClassifier, HistGradientBoostingClassifier, BaggingClassifier, and GradientBoostingClassifier. Notably, decision tree models demonstrated high classification precision at terminal nodes, many of which yielded a 100% probability for a specific severity class.

1. Introduction

The coronavirus disease 2019 (COVID-19) pandemic, caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), has resulted in substantial morbidity, mortality, and profound socioeconomic disruption worldwide. COVID-19 elicits an immune–inflammatory response of variable severity, ranging from asymptomatic infection to acute respiratory failure and death [1,2]. While the respiratory system remains the most frequent site of early manifestations, accumulating evidence indicates that SARS-CoV-2 can affect multiple organ systems. Consequently, the heterogeneity in symptom severity translates into diverse health, economic, and societal consequences [3,4].
One of the most pressing areas of contemporary research in infectious diseases is the prediction of infection risk and mortality associated with COVID-19, as well as the development of classification models to stratify patients on the basis of their clinical profiles [5,6]. Accurate prediction of disease severity is critical for optimizing clinical care, ensuring appropriate allocation of hospital resources, prioritizing patients at highest risk, and supporting physicians in making timely, evidence-based decisions that improve outcomes [7]. Recent studies have highlighted numerous patient-related factors—including demographic, clinical, immunological, haematological, biochemical, and radiological variables—that may assist clinicians in predicting disease severity and mortality [8]. Increasingly, research has focused on designing novel screening methods for infected patients across different stages of disease to uncover associations between clinical features and the likelihood of adverse outcomes [9,10].
Machine learning (ML), as an emerging paradigm, has the potential to enhance policy development, forecasting, screening, drug discovery, and risk stratification. Artificial intelligence (AI) can support equitable decision-making by minimizing inter-observer variability and bridging the gap between healthcare demands and human workload [11]. In medical diagnostics, AI-driven approaches such as fuzzy logic, genetic algorithms, decision trees, support vector machines, artificial neural networks, and deep learning have gained popularity for identifying, classifying, and assessing a wide spectrum of health conditions [12,13]. ML has already demonstrated clinical utility in diverse fields, including diabetes [14], cardiovascular disease [15], cancer [16,17], sepsis [18], and depression [19], yielding favourable results. Moreover, ML has been employed for diagnostic and prognostic modelling in COVID-19 with encouraging findings [20,21,22].
Current evidence underscores that AI- and ML-based methods may play a pivotal role in mitigating the impact of SARS-CoV-2 on population health and life expectancy. Traditional manual risk assessments of disease severity and mortality, based on initial screening and laboratory testing, remain challenging and subject to human bias [23,24,25,26,27,28]. Although several prognostic models have been proposed for early identification of high-risk COVID-19 patients, a critical gap persists in the development of interpretable ML algorithms and high-performance quantitative systems capable of isolating the most selective prognostic biomarkers of mortality. Identifying and prioritizing individuals at elevated risk is crucial both for effective resource planning and for optimizing therapeutic strategies [29]. Accordingly, the objective of this study was to formally evaluate and compare a range of supervised machine learning algorithms for their ability to predict severe COVID-19, using an integrated dataset comprising clinical, biochemical, and genetic variables. This approach was designed to identify the most informative predictors and provide a robust methodological framework for diagnostic management of severe disease.

2. Materials and Methods

The study methodology comprised the following sequential stages:
  • Data acquisition and preprocessing—construction of the feature set, imputation of missing values, and reduction of redundancy through the elimination of highly correlated variables.
  • Correlation analysis—application of statistical methods to assess associations among variables and preliminary selection of the most relevant predictors of COVID-19 severity.
  • Model development and testing—implementation of multiple machine learning classification algorithms, including ensemble methods, naïve Bayes, logistic regression, and others, to address the multiclass classification task of COVID-19 severity.
  • Model optimization and feature importance assessment—hyperparameter tuning of classifiers and estimation of each variable’s contribution to the final predictive decision.
  • Interpretation of results—visualization of model performance using receiver operating characteristic (ROC) curves and explanation of classification logic derived from decision tree structures.
The overall methodology is illustrated in Figure 1, which depicts the workflow and interrelationships among the analytical stages.
The adoption of this workflow not only ensured high predictive accuracy but also provided transparency of the modelling process, which is essential in clinical practice.
1. Data acquisition and preprocessing.
This cohort study enrolled 257 patients with confirmed COVID-19, of whom 197 had moderate-to-severe disease (including 55 with moderate and 142 with severe disease) and 60 presented with mild disease, resulting in three severity classes: mild (n = 60), moderate (n = 55), and severe (n = 142). The target variable was disease severity (mild, moderate, severe). The diagnosis of COVID-19 in all study participants was confirmed by polymerase chain reaction (PCR) testing. After excluding 31 incomplete records, the final dataset included 226 observations with class/severity distribution: mild (n = 30, class 0), moderate (n = 54, class 1), and severe (n = 142, class 2) (Table 1). For machine learning analysis, a dataset of 226 observations comprising 68 features was generated; data from 31 patients were excluded due to incompleteness identified by automated classifiers. Incompleteness was defined as the simultaneous absence of several key clinical or laboratory measurements required for downstream modeling (e.g., IL-6, ALT, AST, lymphocyte count, and platelet count). Automated classifiers were used at the preprocessing stage to flag records in which more than 20% of essential features were missing, or when critical biomarkers (IL-6 and ALT) were absent. Records that failed to meet these minimum completeness thresholds (n = 31) were excluded from modeling. In contrast, isolated missing values (<1% across several categorical variables such as family history or comorbidities) were imputed using median substitution and therefore did not result in case exclusion.
For variables with <1% missingness (e.g., family history of pulmonary, cardiovascular, and connective tissue diseases, comorbid nephrological or gastroenterological pathology, and peripheral artery disease), values were binary indicators coded as 0 = absent and 1 = present. In these cases, median substitution is numerically equivalent to mode imputation and thus maintains the most frequent category. Model-based imputation approaches were considered, but given the extremely low level of missingness (<1%), median substitution was deemed sufficient and minimized the risk of introducing spurious correlations.
In contrast, the 31 excluded records lacked multiple essential continuous variables (IL-6, ALT, AST, lymphocyte count, platelet count), which are among the most important predictors of severity and could not be reliably imputed without biasing the results. Because these features carried high predictive value, exclusion rather than imputation was chosen to avoid introducing synthetic patterns into the dataset. This conservative approach ensured that models were trained only on records with complete measurements of the key prognostic biomarkers.
The mild COVID-19 group (n = 60) included non-hospitalized patients managed by primary care physicians, all with diagnoses confirmed by PCR according to established protocols. Mild COVID-19 was defined according to the following clinical criteria: body temperature not exceeding 38 °C, presence of rhinitis, and dry cough without clinical signs of respiratory insufficiency (including dyspnea, increased respiratory rate, or hemoptysis); absence of gastrointestinal symptoms (nausea, vomiting, and/or diarrhea); and absence of neurological or mental status changes (such as impaired consciousness or lethargy). Within this group, 40.0% (n = 24) were male and 60.0% (n = 36) female, with a mean age of 56.77 ± 5.67 years. A total of 66.7% (n = 40) were vaccinated and 33.3% (n = 20) unvaccinated; 20 patients were active smokers. Age and sex distribution did not significantly differ from those in the moderate-to-severe COVID-19 group (p > 0.05).
The moderate-to-severe group (n = 197) comprised hospitalized patients at the Infectious Disease Department of Uman Central City Hospital with community-acquired COVID-19 pneumonia. Of these, 21.97% (n = 55) had moderate disease, while 78.03% (n = 142) had severe disease. Moderate COVID-19 was defined by the presence of fever, dry cough, dyspnea, and tachypnea (20–30 breaths/min) without clinical signs of severe pneumonia, including oxygen saturation (SpO2) ≥ 90%. Severe disease was characterized by the above-mentioned clinical manifestations in combination with at least one of the following: respiratory rate ≥ 30 breaths/min, blood oxygen saturation ≤ 90%, PaO2/FiO2 ratio < 300, or pulmonary infiltrates involving >50% of lung fields. In all cases, pneumonia was confirmed by instrumental diagnostic methods, including computed tomography (CT) or chest radiography.
Diagnostic evaluation, laboratory testing, and treatment adhered to the Ukrainian Ministry of Health protocols on COVID-19 management [30,31], as well as WHO, CDC, and international standards for COVID-19 diagnosis, treatment, and prevention [32].
Across both groups, patients with moderate disease did not differ significantly from those with severe disease in terms of age, systolic and diastolic blood pressure, or body mass index, although trends toward worse values were noted in the severe group. Sex distribution was balanced (97 men, 100 women); however, women were more frequently represented in the severe group, while men predominated in the moderate group by 17.45% (χ2 = 4.83; p = 0.028). The proportion of unvaccinated individuals was nearly double that of vaccinated patients in both groups, with no statistically significant difference between them (χ2 = 0.4; p = 0.527).
The dataset included demographic features (age, sex), genetic markers [polymorphisms in FGB, NOS3, and TMPRSS2], biochemical indicators (IL-6, endothelin-1, D-dimer, fibrinogen, among others), and clinical parameters (blood pressure, BMI, comorbidities). The target variable was disease severity, coded as a categorical variable (0 = mild, 1 = moderate, 2 = severe). The use of routine laboratory markers for COVID-19 risk stratification has been previously demonstrated [33].
The study complied with international standards of biomedical ethics, including ICH/GCP guidelines, the Declaration of Helsinki (1964), and the Council of Europe Convention on Human Rights and Biomedicine (1997), as well as national legislation. Approval was granted by the Ethics Committee of Bukovinian State Medical University (Protocol No. 7, April 2025). Written informed consent was obtained from all participants prior to enrollment.
Categorical variables included sex, vaccine type, FGB genotype, NOS3 genotype, TMPRSS2 genotype, and severity status. Missing values (<1%) in the following variables—family history of pulmonary disease (Fam. Pulmo), family history of cardiovascular disease (Fam. Cardio), family history of collagenosis (Fam. Collagenosis), myocardial infarction (MI), comorbid nephrological pathology (Compl. Nephro), comorbid gastroenterological pathology (Compl. Gastro), and peripheral artery disease (PAD)—were imputed using median substitution.
2. Correlation analysis.
To identify interdependencies among variables and mitigate multicollinearity, thereby reducing data redundancy prior to the application of machine learning models, we performed a comprehensive correlation analysis. For continuous variables, we employed Pearson’s correlation coefficient; for categorical variables, we applied Cramér’s V coefficient, which is derived from the chi-square statistic. For mixed pairs of continuous and categorical variables, we used the correlation ratio (η), which quantifies the extent to which a continuous variable is explained by a categorical factor.
η 2 = S S b e t w e e n S S t o t a l
Here, SSbetween denotes the sum of squares between groups (defined by categorical variables), and SStotal represents the total sum of squares. The value of η ranges between 0 and 1, with higher values indicating stronger dependence.
3. Model development and evaluation.
We addressed a multiclass classification task with three severity levels of COVID-19 (mild, moderate, severe). To identify the most effective classifiers for predicting COVID-19 severity, we conducted an extensive benchmarking of all available classifiers from the Scikit-learn library under default hyperparameter settings. All models were implemented in Python (version 3.9.12) using the scikit-learn library (version 1.0.2). Model performance was assessed using two primary metrics: overall accuracy (the proportion of correctly classified samples) and the area under the receiver operating characteristic curve (AUC-ROC), which reflects the model’s discriminative ability. To ensure robust estimates, we applied stratified five-fold cross-validation, whereby the dataset was partitioned into five folds; in each iteration, the model was trained on four folds and tested on the remaining one. This process was repeated until every fold served as a test set once. Stratification preserved class proportions across folds, which was particularly critical given our imbalanced dataset, where the outcome variable Severity comprised three classes (0, 1, and 2) with 142, 54, and 30 samples, respectively.
For multiclass classification of COVID-19 severity, we constructed ROC curves using the one-vs-rest (OvR) strategy. Specifically, the target variable Severity encompassed three categories, making the task inherently multiclass. In OvR, a binary classifier is trained for each class against all others, while the alternative one-vs-one (OvO) approach trains classifiers for every pair of classes. In this study, we adopted OvR, as it is the standard approach in most of the algorithms tested and scales effectively when the number of classes is limited.
All models were evaluated under default hyperparameter settings to provide an unbiased estimate of their baseline performance. Given the pronounced class imbalance, accuracy was complemented with AUC-ROC, a more informative metric under uneven class distributions. Stratified cross-validation further ensured that all classes were proportionally represented in both training and test sets, enhancing the reliability of performance estimates.

3. Results

Correlation analysis revealed several pairs of features with high correlation coefficients (absolute values). Certain variables were excluded from further analysis for the following reasons:
Between Vaccine type and Vaccination (r = 0.98): The strong correlation arises because vaccine type is directly linked to the vaccination status. The variable Vaccination was excluded, as Vaccine type provided more detailed information.
Between Lymphocytes and Segmented Neutrophils (r = −0.98): The strong negative correlation reflects a well-known biological relationship, where an increase in segmented neutrophils is typically accompanied by a decrease in lymphocytes. Segmented Neutrophils was removed due to its lower clinical relevance in the context of COVID-19 severity.
Between LDL-C and Total Cholesterol (r = 0.95): The strong correlation is expected, as LDL-C constitutes a major component of total cholesterol. LDL-C was excluded to avoid redundancy.
Between DBP and SBP (r = 0.92): The strong correlation between diastolic (DBP) and systolic (SBP) blood pressure reflects their physiological dependence. DBP was removed because SBP carries greater prognostic value in COVID-19 outcomes.
Between GFR.1 and GFR (r = 0.90): The strong correlation between estimated glomerular filtration rate (GFR) by Cockcroft–Gault (GFR.1) and by CKD-EPI is attributable to their near identity or duplication in the dataset. GFR.1 was removed as redundant.
Between Urea and Creatinine (r = 0.87): Both are markers of renal function. Creatinine was retained due to its broader clinical application in estimating kidney injury (GFR).
Between Atherogenic Index (AI) and LDL-C (r = 0.86): AI largely depends on LDL-C levels; therefore, AI was removed.
Between AI and Total Cholesterol (r = 0.84): Similarly, AI was excluded due to dependence on total cholesterol.
Between Depression and Pneumonia (r = 0.84): This correlation may reflect shared clinical or psychosomatic factors; however, both features were retained due to their distinct clinical significance.
Between LDL-C and HDL-C (r = 0.80).
After excluding highly correlated variables (Vaccination, Segmented Neutrophils, LDL-C, DBP, GFR.1, AI), the dataset was optimized for machine learning analysis. The final dataset comprised 226 records and the following variables: Age (years), Sex (0 = female, 1 = male), Vaccine type (0 = unvaccinated, 1 = vaccinated, 2 = partially vaccinated, 3 = booster dose), Genetic variants: FGB (rs1800790; AA, AG, GG), NOS3 (rs2070744; CC, CT, TT), TMPRSS2 (rs12329760; CC, CT, TT), Protein concentrations: TMPRSS2 (ng/mL), Endothelin-1 (ET-1, pg/mL), Interleukin-6 (IL-6, pg/mL), BMI (kg/m2), Smoking status (0 = no, 1 = yes), Depression (0 = no, 1 = yes), Prothrombin index (PI, %), Fibrinogen (g/L), D-dimer (µg/mL), Triglycerides (mmol/L), ALT (U/L), AST (U/L), GFR (mL/min/1.73 m2), Lymphocytes (%), Platelet count (×109/L), Severity of COVID-19 (0 = mild, 1 = moderate, 2 = severe).
To examine the associations between the target variable Severity and the predictor variables, an additional association analysis was performed. The lower-triangular correlation matrix, summarizing relationships across all variables using mixed correlation coefficients (Pearson for continuous, Cramér’s V for categorical, η for mixed), is presented in Figure 2. A correlation threshold of 0.80 was applied to identify strongly collinear pairs. The analysis revealed no variable pairs exceeding this threshold, indicating the absence of substantial multicollinearity among the predictors.
Associations between the target variable Severity and each predictor were analysed separately and are summarized in Table 2. The table is ranked by the absolute value of the correlation coefficient, enabling the identification of variables exerting the strongest influence on disease severity.
The strongest correlations with the target variable Severity were observed for IL-6 (0.838), Depression (0.739), LDL-H (0.663), AST (0.663), Platelets (−0.628), Lymphocytes (0.627), and ALT (0.612). These findings highlight the relevance of inflammatory markers (IL-6), mental health status (depression), lipid profile (LDL-H, triglycerides), liver enzymes (AST, ALT), and haematological parameters (lymphocytes, platelets) in shaping the severity of COVID-19.
Moderate correlations were identified for Fibrinogen (0.329), BMI (0.324), D-dimer (0.319), and TMPRSS2 (0.253), indicating a less pronounced yet still notable association with disease severity.
By contrast, weak correlations were observed for Smoking (−0.206), Age (0.176), Vaccination (−0.169), ET-1 (0.142), as well as for the genetic markers (FGB, TMPRSS2, NOS3) and Sex, suggesting only minor contributions to severity variance. The variable PI, % showed virtually no correlation with the target (−0.002).
To systematically evaluate predictive performance, the entire spectrum of classifiers was tested, including logistic regression, k-nearest neighbours, decision trees, ensemble methods (random forest, gradient boosting, bagging), naïve Bayes, and support vector machines. Model performance was assessed using 5-fold stratified cross-validation, and the top 10 classifiers ranked by accuracy are presented below.
The predictive performance of the models was systematically assessed using 5-fold stratified cross-validation. The results for the top ten classifiers ranked by overall accuracy are summarized in Table 3.
To validate the impact of excluding highly correlated variables on model performance, we compared the top classifiers on the full dataset (including all original features) versus the optimized dataset (after exclusion of Vaccination, Segmented Neutrophils, LDL-C, DBP, GFR.1, and AI).
Overall, models on the optimized dataset demonstrated comparable or slightly higher accuracy and stability (lower standard deviation), confirming that the exclusion reduced redundancy without compromising predictive power. For instance, the ExtraTreesClassifier achieved 0.974 ± 0.022 on the optimized dataset versus 0.965 ± 0.028 on the full dataset, suggesting improved generalization.
In addition to overall accuracy and multiclass AUC, we further evaluated classifier performance using per-class metrics (precision, recall, and F1-score) under both macro and micro averaging schemes. These additional metrics allowed us to account for class imbalance across the severity categories (Supplementary Materials). Confusion matrices for the top-performing models (ExtraTreesClassifier, RandomForestClassifier, and DecisionTreeClassifier) are provided in the Supplementary Materials (Figures S1–S10). The results confirmed that while ensemble methods achieved the highest overall accuracy, they also demonstrated consistently high recall and precision for the severe class, which was the majority class, and competitive performance for moderate and mild cases. The Decision Tree, although slightly less accurate overall, maintained balanced recall across all three severity levels.
Best-performing models: The highest predictive accuracy was achieved with the ExtraTreesClassifier (Accuracy: 0.974 ± 0.022), an ensemble method based on decision trees with additional randomization. Its superior performance can be attributed to its capacity to efficiently capture complex, non-linear relationships within the dataset. This finding is consistent with the correlation analysis, in which variables such as IL-6, depression, and LDL-C demonstrated strong associations with disease severity. The Random Forest Classifier and Histogram-based Gradient Boosting Classifier also yielded high accuracies (0.960), underscoring the utility of ensemble methods in this predictive context.
Model stability: Algorithms with lower standard deviations, including Logistic RegressionCV (0.934 ± 0.020) and Linear Discriminant Analysis (0.929 ± 0.029), exhibited more stable performance across folds. This consistency is particularly advantageous when working with relatively small or imbalanced clinical datasets.
AUC-ROC performance: All of the top 10 models achieved an apparently perfect AUC-ROC value (1.000), suggesting excellent class discrimination. However, such uniformly high values may also reflect potential overfitting or limited variability within the cross-validation folds. Independent external validation is therefore essential to confirm the robustness of these findings.
Simpler models: Notably, more parsimonious algorithms such as Bernoulli Naïve Bayes (0.956 ± 0.037) and the Decision Tree Classifier (0.938 ± 0.043) performed well despite their relative simplicity. Naïve Bayes classifiers are particularly effective in scenarios where predictor variables are approximately independent, whereas decision trees capture non-linear dependencies but may be less stable due to sensitivity to noise.
Recommendations: Ensemble-based approaches, particularly the ExtraTreesClassifier, RandomForestClassifier, and HistGradientBoostingClassifier, emerge as the most promising candidates for further hyperparameter optimization, given their combination of high accuracy and resilience to data noise. Logistic RegressionCV also warrants consideration owing to its stability and interpretability, both of which are highly relevant in clinical decision-making contexts.
Feature importance analysis, derived via permutation importance, enabled the identification of key predictors of COVID-19 severity. The normalized importances for each model are visualized in Figure 3 as a heatmap, illustrating the relative contribution of each variable to model performance.
Although ensemble classifiers such as ExtraTrees and RandomForest achieved the highest predictive accuracy (up to 97.4%), we additionally present a single decision tree model (Figure 4). The decision tree reached high accuracy (93.8%) and demonstrated 100% class probability in several terminal nodes, indicating robust classification despite its relative simplicity. Importantly, unlike ensemble approaches, the decision tree provides a transparent, rule-based structure that allows clinicians to trace the decision process through explicit thresholds (e.g., IL-6, ALT, lymphocyte count, depression status). This interpretability makes the model particularly relevant in clinical settings, where trust and transparency are as critical as predictive performance. For these reasons, the decision tree was included as a complementary visualization to highlight both predictive power and practical applicability.
Analysis of normalized feature importance (Figure 3) revealed the following:
  • Most influential predictors:
  • IL-6: Consistently emerged as the single most important feature across all classifiers, with normalized importance values of ~1.0. This is concordant with its strong correlation with severity (r = 0.838), reinforcing its central role as a systemic inflammatory marker in COVID-19 pathogenesis.
  • Depression: Assigned high importance by several models, including BernoulliNB (1.0), CalibratedClassifierCV (1.0), BaggingClassifier (0.632), ExtraTreesClassifier (0.683), and DecisionTreeClassifier (0.555). This finding mirrors its correlation coefficient (r = 0.739) and underscores the psychosomatic contribution of depression to disease progression.
  • Lymphocytes: Particularly important for the HistGradientBoostingClassifier (1.0), RandomForestClassifier (0.581), CalibratedClassifierCV (0.623), and ExtraTreesClassifier (0.554), aligning with its correlation (r = 0.627) with severity.
  • LDL-C, AST, ALT, and triglycerides: These biochemical indicators demonstrated moderate-to-high importance in ensemble methods such as RandomForestClassifier (0.480–0.687) and ExtraTreesClassifier (0.308–0.340), consistent with their correlations (0.663, 0.663, 0.612, and 0.602, respectively).
  • Platelets: Retained predictive value in RandomForestClassifier (0.197), ExtraTreesClassifier (0.219), and CalibratedClassifierCV (0.038), in agreement with its negative correlation with severity (r = −0.628).
2.
Less influential predictors:
  • Sex, vaccination status, and genetic markers (FGB, NOS3, TMPRSS2), PI%: These variables had minimal or no importance in most models, reflecting their weak correlations with severity (<0.1). For instance, sex contributed marginal importance (0.008–0.025) in only a few classifiers, while genetic variants often showed zero contribution.
  • Age and BMI: Despite a moderate correlation for BMI (r = 0.324), both features demonstrated low importance (0.0–0.067), suggesting that their effects may be mediated through other covariates.
  • Fibrinogen, D-dimer, ET-1, and GFR: These markers displayed modest importance in some ensemble models (e.g., RandomForestClassifier, ExtraTreesClassifier), but their contributions were consistently lower than those of IL-6 or depression.
3.
Model-specific patterns:
  • Ensemble methods (ExtraTreesClassifier, RandomForestClassifier, BaggingClassifier) captured a broad spectrum of influential features, reflecting their ability to model non-linear interactions. For example, RandomForestClassifier assigned substantial weight to LDL-C (0.480), AST (0.510), and triglycerides (0.298).
  • HistGradientBoostingClassifier focused almost exclusively on IL-6 (0.828) and lymphocytes (1.0), which may reflect either strong regularization or sensitivity to class imbalance.
  • BernoulliNB and CalibratedClassifierCV prioritized depression (1.0), likely due to their sensitivity to categorical or binary predictors.
  • LogisticRegressionCV and Linear Discriminant Analysis emphasized a narrower set of variables (IL-6, lymphocytes, depression), consistent with their linear structure and more limited capacity to account for complex feature interactions.
Negative feature importance. Some models, such as BernoulliNB (Sex: −0.0098) and Linear Discriminant Analysis (Age, yo: −0.0077, TMPRSS2: −0.0077), demonstrated negative feature importance for certain variables. This phenomenon may indicate that shuffling these predictors slightly improves model performance, which is uncommon and may be related to noise in the dataset or the weak contribution of these variables.
Figure 4 presents a decision tree constructed to predict the probability of COVID-19 outcomes. The primary predictor at the root node is the interleukin-6 (IL-6) level. Subsequent branching is determined by clinical and biochemical parameters such as alanine aminotransferase (ALT), lymphocyte count (Lympho), depressive status (Depress), fibrinogen concentration, glomerular filtration rate (GFR), and the genetic marker (Gene N). Each node of the tree displays the splitting condition, the target class (0—mild course disease, 1—moderate severity course disease, 2—severe disease) (Table 1), probability, and the number of samples within the node. For example, when IL-6 ≤ 54.72 pg/mL, the presence of depression emerges as the next critical factor, followed by lymphocyte levels. Conversely, when IL-6 > 54.72 pg/mL, the decisive variable becomes ALT, suggesting hepatic involvement in the disease process.
The tree demonstrates a high degree of classification accuracy at terminal nodes, with the majority showing 100% probability of a specific class. This underscores the importance of multiparametric analysis in predicting the course of COVID-19.

4. Discussion

In this study, a comprehensive approach was applied to analyse the interplay between genetic, biochemical, clinical, and demographic features of COVID-19 patients and disease severity. The primary objective was to identify the most informative predictors, construct accurate machine learning models for severity prediction, and develop interpretable solutions applicable in clinical practice. A review of recent publications [34] indicates the increasing adoption of ML approaches in COVID-19 research since 2020.
Strong positive correlations identified in our analysis (IL-6, Depression, LDL-H, AST) suggest that elevated levels of these markers are associated with increased severity, while negative correlations for Platelets and GFR indicate reductions in these parameters during severe COVID-19 [35]. The observed role of depression as a splitting variable in the decision tree can be clinically interpreted through the interaction between mental health and immune function. Depression is known to be associated with chronic systemic inflammation and dysregulation of cytokine pathways, including elevated IL-6 levels reported in prior studies [36]. Therefore, patients without depression and with relatively low IL-6 values (≤54.72 pg/mL) are more likely to follow a mild disease course, as indicated by the leaf node predicting class 0. This phenomenon does not imply a causal effect of depression alone but rather reflects its correlation with immunological and biochemical alterations that contribute to a more severe course of COVID-19 when present. The impact of IL-6, -reactive protein (CRP), and procalcitonin on clinical status has also been investigated in previous studies [37].
Moderate correlations observed for Fibrinogen, BMI, D-dimer, and TMPRSS2 may highlight their secondary role in COVID-19 progression, particularly through coagulation pathways (Fibrinogen, D-dimer) and metabolic processes (BMI).
By contrast, weak correlations for Smoking, Age, Vaccination, ET-1, as well as genetic markers (FGB, TMPRSS2, NOS3) and Sex, suggest a limited impact on disease severity in this dataset. Notably, the negative correlation of Vaccination may indicate a potential protective effect, albeit weak. The absence of contribution for PI, % (zero correlation) further supports its minimal relevance in severity prediction. These findings emphasize the importance of inflammatory biomarkers, liver enzymes, and haematological indicators for severity assessment, whereas genetic markers and demographic factors, such as sex and vaccination status, appear less influential in the current cohort.
The critical role of immunological markers is consistent with findings from other studies [37], where IL-6, LDL-H, and lymphocyte count were highlighted as key predictors. The influence of genetic markers in COVID-19 modelling has been systematically reviewed in [38]. Furthermore, laboratory markers such as ALT, AST, and platelets have demonstrated strong prognostic significance [39]. IL-6 has been identified not only as a biomarker but also as a therapeutic target [40]. AI-driven models further validate the predictive value of LDH, IL-6, and D-dimer [41]. Confirmation of our modelling strategy is supported by findings from [42], where similar biomarkers enabled robust severity prediction.
Among the ten classifiers evaluated, the decision tree represents the only model capable of providing direct visual interpretability through a hierarchical set of if–then rules. Unlike “black-box” models—such as ensemble methods that aggregate multiple trees or rely on complex computations, thereby obscuring interpretability—the decision tree offers transparent logic, allowing each step of the classification process to be traced. This feature facilitates the identification of the most influential predictors, such as IL-6, D-dimer, and Age, yo, which play a critical role in clinical decision-making. The value of explainable AI in this context is underscored by prior work [43], which successfully applied transparent interpretability frameworks in biomedical modelling.
A limitation of the present study is the absence of a demographic stratification of communicative and behavioural signals that may interact in parallel with clinical markers, and the lack of external validation on an independent dataset. While our models were optimized and evaluated using stratified cross-validation, the absence of independent external data restricts the ability to fully assess their generalizability. Future studies should therefore prioritize validation in larger, independent patient cohorts to confirm the robustness and applicability of the proposed models.
Interpretability is particularly vital in medicine, where clinicians and researchers require not only accurate predictions but also comprehensible explanations to support informed decision-making. For example, a simple rule such as “If IL-6 > 5.0 pg/mL and Age > 50 years, then the probability of severe disease is high” provides an intuitive rationale for identifying high-risk patients and tailoring therapeutic strategies accordingly. While ensemble-based classifiers (e.g., ExtraTreesClassifier) achieved near-perfect accuracy (up to 100%), they inherently lack this explanatory transparency, rendering them less suitable for medical settings in which trust and interpretability are paramount. The decision tree, in contrast, strikes a balance between predictive accuracy (99.12%, with 224 correct classifications out of 226) and the capacity to deliver interpretable outputs, making it especially valuable for clinical applications. Importantly, our findings align with international evidence on the use of biomarkers to predict COVID-19 severity [44].

5. Conclusions

Our analysis demonstrated that the variables most strongly correlated with the target outcome Severity included IL-6, Depression/Pneumonia, LDL-H, AST, Platelets, Lymphocytes, and ALT. This highlights the central role of inflammatory markers (IL-6), clinical status (Depression/Pneumonia), lipid metabolism (LDL-H, Triglycerides), hepatic enzymes (AST, ALT), and haematological indices (Lymphocytes, Platelets) in disease progression. Positive correlations indicate that higher levels of these parameters are associated with more severe disease, whereas negative correlations for Platelets and GFR suggest a reduction in these values during severe COVID-19.
Machine learning models such as ExtraTreesClassifier, HistGradientBoostingClassifier, BaggingClassifier, and GradientBoostingClassifier achieved the highest accuracy with no classification errors, underscoring the strength of ensemble methods in capturing nonlinear dependencies and mitigating potential class imbalance. Nevertheless, among all classifiers, the decision tree remains unique in offering an interpretable, rule-based framework that complements predictive performance with clinical transparency.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/computation13100238/s1, Figure S1: Confusion matrix of the ExtraTreesClassifier on the optimized dataset (5-fold cross-validation predictions). Rows correspond to true class labels (0 = mild, 1 = moderate, 2 = severe), and col-umns correspond to predicted labels; Figure S2 Confusion matrix of the RandomForestClassifier on the optimized dataset; Figure S3: Confusion matrix of the HistGradientBoostingClassifier on the optimized dataset; Figure S4: Confusion matrix of the BernoulliNB classifier on the optimized dataset; Figure S5: Confusion matrix of the BaggingClassifier on the optimized dataset; Figure S6: Confusion matrix of the CalibratedClassifierCV on the optimized dataset; Figure S7: Confusion matrix of the DecisionTreeClassifier on the optimized dataset; Figure S8: Confusion matrix of the GradientBoostingClassifier on the optimized dataset; Figure S9: Confusion matrix of the LogisticRegressionCV classifier on the optimized dataset; Figure S10: Confusion matrix of the LinearDiscriminantAnalysis on the optimized dataset; Table S1: Summary of extended performance metrics (accuracy, balanced accuracy, precision, recall, F1-score, ROC-AUC, Brier score, and expected calibration error) for the top 10 classifiers on the optimized dataset using 5-fold stratified cross-validation; Table S2: Per-class precision, recall, and F1-score for the top 10 classifiers on the optimized dataset using 5-fold stratified cross-validation, reported separately for mild (class 0), moderate (class 1), and severe (class 2) COVID-19 cases.

Author Contributions

Conceptualization and writing—original draft preparation, M.S., L.S. and M.Š.; writing—review and editing, Y.V., R.S., L.S. and D.M.; supervision, M.Š., Y.V.; project administration, M.Š. D.L. and Y.V.; visualization, A.S., D.M.; funding acquisition, M.S. and L.S. Modelling, Y.V., D.L. and D.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study is available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SNPsingle-nucleotide polymorphisms
MLmachine learning
ROCreceiver operating characteristic
WHOWorld Health Organization
CDCCenters for Disease Control and Prevention
BMIbody mass index
ILinterleukin
MImyocardial infarction
PADperipheral artery disease
BPblood pressure
GFRglomerular filtration rate

References

  1. Chan, J.F.W.; Yuan, S.; Chu, H.; Sridhar, S.; Yuen, K.Y. COVID-19 drug discovery and treatment options. Nat. Revs Microbiol. 2024, 22, 391–407. [Google Scholar] [CrossRef]
  2. Zhao, H.; To, K.K.; Lam, H.; Zhou, X.; Chan, J.F.W.; Peng, Z.; Lee, A.C.Y.; Cai, J.; Chan, W.-M.; Ip, J.D.; et al. Cross-linking peptide and repurposed drugs inhibit both entry pathways of SARS-CoV-2. Nat. Commun. 2021, 12, 1517. [Google Scholar] [CrossRef]
  3. Wu, Y.; Xu, X.; Chen, Z.; Duan, J.; Hashimoto, K.; Yang, L.; Liu, C.; Yang, C. Nervous system involvement after infection with COVID-19 and other coronaviruses. Brain Behav. Immun. 2020, 87, 18–22. [Google Scholar] [CrossRef] [PubMed]
  4. Sokolenko, M.O.; Sydorchuk, L.P.; Sokolenko, L.S.; Sokolenko, A.A. General immunologic reactivity of patients with COVID-19 and its relation to gene polymorphism, severity of clinical course of the disease and combination with comorbidities. Med. Perspekt. 2024, 29, 108. [Google Scholar] [CrossRef]
  5. Chowdhury, M.E.H.; Rahman, T.; Khandakar, A.; Al-Madeed, S.; Zughaier, S.M.; Doi, S.A.R.; Hassen, H.; Islam, M.T. An Early Warning Tool for Predicting Mortality Risk of COVID-19 Patients Using Machine Learning. Cogn. Comput. 2021, 16, 1778–1793. [Google Scholar] [CrossRef]
  6. Nemati, M.; Ansary, J.; Nemati, N. Machine-Learning Approaches in COVID-19 Survival Analysis and Discharge-Time Likelihood Prediction Using Clinical Data. Patterns 2020, 1, 100074. [Google Scholar] [CrossRef] [PubMed]
  7. Sayed, S.A.; Elkorany, A.M.; Sayed, M.S. Applying Different Machine Learning Techniques for Prediction of COVID-19 Severity. IEEE Access 2021, 9, 135697–135707. [Google Scholar] [CrossRef]
  8. Gallo, M.B.; Aghagoli, G.; Lavine, K.; Yang, L.; Siff, E.J.; Chiang, S.S.; Salazar-Mather, T.P.; Dumenco, L.; Savaria, M.C.; Aung, S.N.; et al. Predictors of COVID-19 severity: A literature review. Rev. Med. Virol. 2021, 31, 1–10. [Google Scholar] [CrossRef]
  9. Gazzaruso, C.; Paolozzi, E.; Valenti, C.; Brocchetta, M.; Naldani, D.; Grignani, C.; Salvucci, F.; Marino, F.; Coppola, A.; Gallotti, P. Association between antithrombin and mortality in patients with COVID-19. A possible link with obesity. Nutr. Metab. Cardiovasc. Dis. 2020, 30, 1914–1919. [Google Scholar] [CrossRef]
  10. Malki, Z.; Atlam, E.S.; Hassanien, A.E.; Dagnew, G.; Elhosseini, M.A.; Gad, I. Association between weather data and COVID-19 pandemic predicting mortality rate: Machine learning approaches. Chaos Soliton Fractals 2020, 138, 110137. [Google Scholar] [CrossRef]
  11. Lalmuanawma, S.; Hussain, J.; Chhakchhuak, L. Applications of machine learning and artificial intelligence for COVID-19 (SARS-CoV-2) pandemic: A review. Chaos Soliton Fractals 2020, 139, 110059. [Google Scholar] [CrossRef]
  12. Cobianchi, L.; Piccolo, D.; Dal Mas, F.; Agnoletti, V.; Ansaloni, L.; Balch, J.; Biffl, W.; Butturini, G.; Catena, F.; Coccolini, F.; et al. Surgeons’ perspectives on artificial intelligence to support clinical decision-making in trauma and emergency contexts: Results from an international survey. World J. Emerg. Surg. 2023, 18, 1. [Google Scholar] [CrossRef] [PubMed]
  13. De Simone, B.; Abu-Zidan, F.M.; Saeidi, S.; Deeken, G.; Biffl, W.L.; Moore, E.E.; Sartelli, M.; Coccolini, F.; Ansaloni, L.; Di Saverio, S.; et al. Knowledge, attitudes and practices of using Indocyanine Green (ICG) fluorescence in emergency surgery: An international web-based survey in the ARtificial Intelligence in Emergency and trauma Surgery (ARIES)—WSES project. Updates Surg. 2024, 76, 1969–1981. [Google Scholar] [CrossRef]
  14. Silva, K.; Lee, W.K.; Forbes, A.; Demmer, R.T.; Barton, C.; Enticott, J. Use and performance of machine learning models for type 2 diabetes prediction in community settings: A systematic review and meta-analysis. Int. J. Med. Inform. 2020, 143, 104268. [Google Scholar] [CrossRef]
  15. Krittanawong, C.; Virk, H.U.H.; Bangalore, S.; Wang, Z.; Johnson, K.W.; Pinotti, R.; Zhang, H.; Kaplin, S.; Narasimhan, B.; Kitai, T.; et al. Machine learning prediction in cardiovascular diseases: A meta-analysis. Sci. Rep. 2020, 10, 16057. [Google Scholar] [CrossRef]
  16. Castaldo, R.; Cavaliere, C.; Soricelli, A.; Salvatore, M.; Pecchia, L.; Franzese, M. Radiomic and Genomic Machine Learning Method Performance for Prostate Cancer Diagnosis: Systematic Literature Review. J. Med. Internet Res. 2021, 23, e22394. [Google Scholar] [CrossRef]
  17. Lu, W.; Fu, D.; Kong, X.; Huang, Z.; Hwang, M.; Zhu, Y.; Chen, L.; Jiang, K.; Li, X.; Wu, Y.; et al. FOLFOX treatment response prediction in metastatic or recurrent colorectal cancer patients via machine learning algorithms. Cancer Med. 2020, 9, 1419–1429. [Google Scholar] [CrossRef]
  18. Fleuren, L.M.; Klausch, T.L.T.; Zwager, C.L.; Schoonmade, L.J.; Guo, T.; Roggeveen, L.F.; Swart, E.L.; Girbes, A.R.J.; Thoral, P.; Ercole, A.; et al. Machine learning for the prediction of sepsis: A systematic review and meta-analysis of diagnostic test accuracy. Intensive Care Med. 2020, 46, 383–400. [Google Scholar] [CrossRef] [PubMed]
  19. Lee, Y.; Ragguett, R.M.; Mansur, R.B.; Boutilier, J.J.; Rosenblat, J.D.; Trevizol, A.; Brietzke, E.; Lin, K.; Pan, Z.; Subramaniapillai, M.; et al. Applications of machine learning algorithms to predict therapeutic outcomes in depression: A meta-analysis and systematic review. J. Affect. Disord. 2018, 241, 519–532. [Google Scholar] [CrossRef]
  20. Li, W.T.; Ma, J.; Shende, N.; Castaneda, G.; Chakladar, J.; Tsai, J.C.; Apostol, L.; Honda, C.O.; Xu, J.; Wong, L.M.; et al. Using machine learning of clinical data to diagnose COVID-19: A systematic review and meta-analysis. BMC Med. Inform. Decis. Mak. 2020, 20, 247. [Google Scholar] [CrossRef] [PubMed]
  21. Wu, G.; Yang, P.; Xie, Y.; Woodruff, H.C.; Rao, X.; Guiot, J.; Frix, A.N.; Louis, R.; Moutschen, M.; Li, J.; et al. Development of a clinical decision support system for severity risk prediction and triage of COVID-19 patients at hospital admission: An international multicentre study. Eur. Respir. J. 2020, 56, 2001104. [Google Scholar] [CrossRef]
  22. Wynants, L.; Van Calster, B.; Collins, G.S.; Riley, R.D.; Heinze, G.; Schuit, E.; Bonten, M.M.J.; Dahly, D.L.; Damen, J.A.A.; Debray, T.P.A.; et al. Prediction models for diagnosis and prognosis of covid-19: Systematic review and critical appraisal. BMJ 2020, 369, m1328. [Google Scholar] [CrossRef] [PubMed]
  23. Albahri, A.S.; Hamid, R.A.; Alwan, J.K.; Al-Qays, Z.T.; Zaidan, A.A.; Zaidan, B.B.; Albahri, A.O.S.; AlAmoodi, A.H.; Khlaf, J.M.; Almahdi, E.M.; et al. Role of biological Data Mining and Machine Learning Techniques in Detecting and Diagnosing the Novel Coronavirus (COVID-19): A Systematic Review. J. Med. Syst. 2020, 44, 122. [Google Scholar] [CrossRef] [PubMed]
  24. De Simone, B.; Abu-Zidan, F.M.; Kasongo, L.; Moore, E.E.; Podda, M.; Sartelli, M.; Isik, A.; Bala, M.; Coimbra, R.; Balogh, Z.J.; et al. COVID-19 infection is a significant risk factor for death in patients presenting with acute cholecystitis: A secondary analysis of the ChoCO-W cohort study. World J. Emerg. Surg. 2025, 20, 16. [Google Scholar] [CrossRef] [PubMed]
  25. Barough, S.S.; Safavi-Naini, S.A.A.; Siavoshi, F.; Tamimi, A.; Ilkhani, S.; Akbari, S.; Ezzati, S.; Hatamabadi, H.; Pourhoseingholi, M.A. Generalizable machine learning approach for COVID-19 mortality risk prediction using on-admission clinical and laboratory features. Sci. Rep. 2023, 13, 2399. [Google Scholar] [CrossRef]
  26. Patel, D.; Kher, V.; Desai, B.; Lei, X.; Cen, S.; Nanda, N.; Gholamrezanezhad, A.; Duddalwar, V.; Varghese, B.; A Oberai, A. Machine learning based predictors for COVID-19 disease severity. Sci. Rep. 2021, 11, 4673. [Google Scholar] [CrossRef]
  27. Reichert, M.; Sartelli, M.; Weigand, M.A.; Hecker, M.; Oppelt, P.U.; Noll, J.; Askevold, I.H.; Liese, J.; Padberg, W.; Coccolini, F.; et al. Two years later: Is the SARS-CoV-2 pandemic still having an impact on emergency surgery? An international cross-sectional survey among WSES members. World J. Emerg. Surg. 2022, 17, 34. [Google Scholar] [CrossRef]
  28. Alotaibi, A.; Shiblee, M.; Alshahrani, A. Prediction of severity of COVID-19-infected patients using machine learning techniques. Computers 2021, 10, 31. [Google Scholar] [CrossRef]
  29. Sokolenko, M.; Sydorchuk, L.; Sokolenko, A.; Sydorchuk, R.; Kamyshna, I.; Sydorchuk, A.; Sokolenko, L.; Sokolenko, O.; Oksenych, V.; Kamyshnyi, O. Antiviral Intervention of COVID-19: Linkage of Disease Severity with Genetic Markers FGB (rs1800790), NOS3 (rs2070744) and TMPRSS2 (rs12329760). Viruses 2025, 17, 792. [Google Scholar] [CrossRef]
  30. Protocol “Provision of medical assistance for the treatment of coronavirus disease (COVID-19)”. Approved by the Order of the Ministry of Health of Ukraine of 2 April 2020 No. 762 (As Amended by the Order of the Ministry of Health of Ukraine of 17 May 2023 No. 913. (In Ukrainian). Available online: https://www.dec.gov.ua/wp-content/uploads/2023/05/protokol-covid2023.pdf (accessed on 12 September 2025).
  31. National Medical Care Standard “Coronavirus disease (COVID-19)”. Approved by Order No. 722 of the Ministry of Health of Ukraine Dated 28 March 2020. (In Ukrainian). Available online: https://www.dec.gov.ua/wp-content/uploads/2021/10/2020_722_standart_covid_19.pdf (accessed on 12 September 2025).
  32. CDC 24/7: Saving Lives, Protecting People. Prevention Actions to Use at All COVID-19 Community Levels. Center for Disease Control and Prevention. 2023. Available online: https://www.cdc.gov/covid/prevention/index.html (accessed on 12 September 2025).
  33. Hu, C.; Liu, Z.; Jiang, Y.; Shi, O.; Zhang, X.; Xu, K.; Suo, C.; Wang, Q.; Song, Y.; Yu, K.; et al. Early prediction of mortality risk among patients with severe COVID-19, using machine learning. Int. J. Epidemiol. 2021, 49, 1918–1929. [Google Scholar] [CrossRef]
  34. Lv, H.; Liu, Y.; Yin, H.; Xi, J.; Wei, P. ML Applications in prediction models for COVID-19: Bibliometric analysis. Information 2024, 15, 575. [Google Scholar] [CrossRef]
  35. Han, H.; Ma, Q.; Li, C.; Liu, R.; Zhao, L.; Wang, W.; Zhang, P.; Liu, X.; Gao, G.; Liu, F.; et al. Profiling serum cytokines in COVID-19 patients reveals IL-6 and IL-10 are disease severity predictors. Emerg. Microbes Infect. 2020, 9, 1123–1130. [Google Scholar] [CrossRef]
  36. De Simone, B.; Abu-Zidan, F.M.; Chouillard, E.; Di Saverio, S.; Sartelli, M.; Podda, M.; Gomes, C.A.; Moore, E.E.; Moug, S.J.; Ansaloni, L.; et al. The ChoCO-W prospective observational global study: Does COVID-19 increase gangrenous cholecystitis? World J. Emerg. Surg. 2022, 17, 61. [Google Scholar] [CrossRef] [PubMed]
  37. Tulu, T.W.; Wan, T.K.; Chan, C.L.; Wu, C.H.; Woo, P.Y.M.; Tseng, C.Z.S.; Vodencarevic, A.; Menni, C.; Chan, K.H.K. Machine learning-based prediction of COVID-19 mortality using immunological-metabolic biomarkers. BMC Digit. Health 2023, 1, 6. [Google Scholar]
  38. Sekaran, K.; Gnanasambandan, R.; Thirunavukarasu, R.; Iyyadurai, R.; Karthik, G.; Doss, C.G.P. A systematic review of artificial intelligence-based COVID-19 modeling on multimodal genetic information. Prog. Biophys. Mol. Biol. 2023, 179, 1–9. [Google Scholar] [CrossRef]
  39. Statsenko, Y.; Al Zahmi, F.; Habuza, T.; Gorkom, K.N.; Zaki, N. Prediction of COVID-19 severity using laboratory findings on admission: Informative values, thresholds, ML model performance. BMJ Open 2021, 11, e044500. [Google Scholar] [CrossRef]
  40. Liang, P.; Li, Y.; Meng, L.; Li, Y.; Mai, H.; Li, T.; Ma, J.; Ma, J.; Wang, J.; Zhuan, B.; et al. Prognostic significance of serum interleukin-6 in severe/critical COVID-19 patients treated with tocilizumab: A detailed observational study analysis. Sci. Rep. 2024, 14, 29634. [Google Scholar] [CrossRef]
  41. Li, Y.; Shang, K.; Bian, W.; He, L.; Fan, Y.; Ren, T.; Zhang, J. Prediction of disease progression in patients with COVID-19 by artificial intelligence assisted lesion quantification. Sci. Rep. 2020, 10, 22083. [Google Scholar] [CrossRef] [PubMed]
  42. Nguyen, H.T.T.; Le-Quy, V.; Ho, S.V.; Thomsen, J.H.D.; Pontoppidan Stoico, M.; Tong, H.V.; Nguyen, N.L.; Krarup, H.B.; Nguyen, S.H.; Tran, V.Q.; et al. Outcome prediction model and prognostic biomarkers for COVID-19 patients in Vietnam. ERJ Open Res. 2023, 9, 00481-2022. [Google Scholar] [CrossRef] [PubMed]
  43. Pourhomayoun, M.; Shakibi, M. Predicting mortality risk in patients with COVID-19 using machine learning to help medical decision-making. Smart Health 2021, 20, 100178. [Google Scholar] [CrossRef]
  44. Laatifi, M.; Douzi, S.; Bouklouz, A.; Ezzine, H.; Jaafari, J.; Zaid, Y.; El Ouahidi, B.; Naciri, M. Machine learning approaches in COVID-19 severity risk prediction in Morocco. J. Big Data 2022, 9, 5. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Study methodology visualized using Orange flow diagrams.
Figure 1. Study methodology visualized using Orange flow diagrams.
Computation 13 00238 g001
Figure 2. Lower-triangular correlation matrix of predictor variables after exclusion of highly correlated features. Correlations were computed using Pearson’s coefficient for pairs of continuous variables, Cramér’s V for pairs of categorical variables, and the correlation ratio (η) for mixed continuous-categorical pairs, as described in the Methods section. A threshold of 0.80 was applied to identify strongly collinear pairs (shown in darker shades); no pairs exceeded this threshold in the optimized dataset. The color scale represents the absolute correlation strength, ranging from 0 (light) to 1 (dark).
Figure 2. Lower-triangular correlation matrix of predictor variables after exclusion of highly correlated features. Correlations were computed using Pearson’s coefficient for pairs of continuous variables, Cramér’s V for pairs of categorical variables, and the correlation ratio (η) for mixed continuous-categorical pairs, as described in the Methods section. A threshold of 0.80 was applied to identify strongly collinear pairs (shown in darker shades); no pairs exceeded this threshold in the optimized dataset. The color scale represents the absolute correlation strength, ranging from 0 (light) to 1 (dark).
Computation 13 00238 g002
Figure 3. Heatmap of normalized feature importance across the top 10 classifiers for multiclass prediction of COVID-19 severity, computed using permutation importance with 5-fold stratified cross-validation.
Figure 3. Heatmap of normalized feature importance across the top 10 classifiers for multiclass prediction of COVID-19 severity, computed using permutation importance with 5-fold stratified cross-validation.
Computation 13 00238 g003
Figure 4. Decision tree for patient category determination.
Figure 4. Decision tree for patient category determination.
Computation 13 00238 g004
Table 1. Observations and dataset, class/severity distribution.
Table 1. Observations and dataset, class/severity distribution.
Patients/ObservationCOVID-19 Severity Course
Mild, n (%)Moderate, n (%)Severe, n (%)
Primary dataset (enrolled patients), n = 257 (%)60 (23.35)55 (21.40)142 (55.25)
Final dataset, n = 226 (%)30 (13.27)54 (23.89)142 (62.83)
Class/severity distribution012
Table 2. Correlation coefficients between the target variable (Severity) and predictor variables.
Table 2. Correlation coefficients between the target variable (Severity) and predictor variables.
VariableCorrelation
IL-60.838
Depression0.738
LDL-H0.663
AST0.662
Platelets−0.628
Lymphocytes0.627
ALT0.611
Triglycerides0.602
GFR−0.471
Fibrinogen0.329
BMI0.324
D-dimer0.319
TMPRSS20.252
Smoking−0.205
Age, yo0.176
Vaccination−0.168
ET-10.141
Gene FGB0.087
Gene TMPRSS20.056
Gene NOS30.037
Sex0.018
PI, %−0.002
Table 3. Top 10 classifiers ranked by predictive accuracy.
Table 3. Top 10 classifiers ranked by predictive accuracy.
Full DatasetOptimized Dataset
ModelAccuracy (Mean ± SD)AUC-ROCAccuracy (Mean ± SD)AUC-ROC
ExtraTreesClassifier0.965 (±0.027)1.0000.974 (±0.022)1.000
RandomForestClassifier0.945 (±0.050)1.0000.960 (±0.035)1.000
HistGradientBoostingClassifier0.948 (±0.051)1.0000.960 (±0.038)1.000
BernoulliNB0.945 (±0.044)1.0000.956 (±0.037)1.000
BaggingClassifier0.944 (±0.043)1.0000.951 (±0.036)1.000
CalibratedClassifierCV0.936 (±0.037)0.9980.943 (±0.030)1.000
DecisionTreeClassifier0.932 (±0.051)0.9960.938 (±0.043)1.000
GradientBoostingClassifier0.920 (±0.056)0.9960.934 (±0.046)1.000
LogisticRegressionCV0.923 (±0.029)0.9970.934 (±0.020)1.000
LinearDiscriminantAnalysis0.917 (±0.037)0.9970.929 (±0.029)1.000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sydorchuk, L.; Sokolenko, M.; Škoda, M.; Lajcin, D.; Vyklyuk, Y.; Sydorchuk, R.; Sokolenko, A.; Martjanov, D. Management of Severe COVID-19 Diagnosis Using Machine Learning. Computation 2025, 13, 238. https://doi.org/10.3390/computation13100238

AMA Style

Sydorchuk L, Sokolenko M, Škoda M, Lajcin D, Vyklyuk Y, Sydorchuk R, Sokolenko A, Martjanov D. Management of Severe COVID-19 Diagnosis Using Machine Learning. Computation. 2025; 13(10):238. https://doi.org/10.3390/computation13100238

Chicago/Turabian Style

Sydorchuk, Larysa, Maksym Sokolenko, Miroslav Škoda, Daniel Lajcin, Yaroslav Vyklyuk, Ruslan Sydorchuk, Alina Sokolenko, and Dmytro Martjanov. 2025. "Management of Severe COVID-19 Diagnosis Using Machine Learning" Computation 13, no. 10: 238. https://doi.org/10.3390/computation13100238

APA Style

Sydorchuk, L., Sokolenko, M., Škoda, M., Lajcin, D., Vyklyuk, Y., Sydorchuk, R., Sokolenko, A., & Martjanov, D. (2025). Management of Severe COVID-19 Diagnosis Using Machine Learning. Computation, 13(10), 238. https://doi.org/10.3390/computation13100238

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop