Next Article in Journal
Challenges in the Pediatric Celiac Disease Diagnosis: An Up-to-Date Review
Previous Article in Journal
Kinetics of High-Sensitive Cardiac Troponin I in Patients with ST-Segment Elevation Myocardial Infarction and Non-ST Segment Elevation Myocardial Infarction
Previous Article in Special Issue
A Virtual Reality-Based Multimodal Approach to Diagnosing Panic Disorder and Agoraphobia Using Physiological Measures: A Machine Learning Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Artificial Intelligence to Develop Clinical Decision Support Systems—The Evolving Road of Personalized Oncologic Therapy

by
Elena Chitoran
1,2,
Vlad Rotaru
1,2,*,
Aisa Gelal
1,2,*,
Sinziana-Octavia Ionescu
1,2,
Giuseppe Gullo
3,
Daniela-Cristina Stefan
1 and
Laurentiu Simion
1,2
1
Medicine School, “Carol Davila” University of Medicine and Pharmacy, 050474 Bucharest, Romania
2
General Surgery and Surgical Oncology Department I, Bucharest Institute of Oncology “Prof. Dr. Alexandru Trestioreanu”, 022328 Bucharest, Romania
3
Department of Obstetrics and Gynecology, Villa Sofia Cervello Hospital, University of Palermo, 90146 Palermo, Italy
*
Authors to whom correspondence should be addressed.
Diagnostics 2025, 15(18), 2391; https://doi.org/10.3390/diagnostics15182391
Submission received: 28 July 2025 / Revised: 13 September 2025 / Accepted: 18 September 2025 / Published: 19 September 2025
(This article belongs to the Special Issue A New Era in Diagnosis: From Biomarkers to Artificial Intelligence)

Abstract

Background/Objectives: The use of artificial intelligence (AI) in oncology has the potential to improve decision making, particularly in managing the risk associated with targeted therapies. This study aimed to develop and validate a machine learning-based clinical decision support system (CDSS) capable of predicting complications associated with Bevacizumab or its biosimilars and to translate the resulting predictive model into a clinically applicable tool. Methods: A prospective observational study was conducted on 395 records from patients treated with Bevacizumab or biosimilars for solid tumors. Pretherapeutic variables, such as demographic data, medical history, tumor characteristics and laboratory findings, were retrieved from medical records. Several machine learning models (logistic regression, Random Forest, XGBoost) were trained using 70/30 and 80/20 data splits. Their predictive performances were compared using accuracy, AUC-ROC, sensitivity, specificity, F1-scores and error rate. The best-performing model was used to derive a logistic-based risk score, which was further implemented as an interactive HTML form. Results: The optimized Random Forest model trained on the 80/20 split demonstrated the best balance between accuracy (70.63%), sensitivity (66.67%), specificity (73.85%), and AUC-ROC (0.75). The derived logistic risk score showed good performance (AUC-ROC = 0.720) and calibration. It identified variables, such as age ≥ 65, anemia, elevated urea, leukocytosis, tumor differentiation, and stage, as significant predictors of complications. The final tool provides clinicians with an easy-to-use, offline form that estimates individual risk levels and stratifies patients into low-, intermediate-, or high-risk categories. Conclusions: This study offers a proof of concept for developing AI-supported predictive tools in oncology using real-world data. The resulting logistic risk score and interactive form can assist clinicians in tailoring therapeutic decisions for patients receiving targeted therapies, enhancing the personalization of care without replacing clinical judgment.

1. Introduction

In the last few years, oncological treatment has rapidly evolved and targeted therapies have become readily available for a large number of patients on a global scale. As such, oncology has generated an ever-increasing influx of real-time data. The development of artificial intelligence tools provides an enhanced possibility of data synthesis and opens the way for innovative perspectives in creating advanced clinical tools. These tools (or clinical decision support systems—CDSSs) are not meant to replace medical expertise but to serve as complementary tools and help improve the accuracy and speed of the decisional process [1].
Artificial intelligence (AI) tools can be used to rapidly process high volumes of, often, disjointed data (clinical, paraclinical, imaging, molecular, genetic) and extract predictive patterns difficult to observe through traditional methods, which can be used for developing clinical scores. Interactive forms based on AI algorithms can be used to evaluate the patient’s risk of developing complications of targeted therapies in real time, such as angiogenesis inhibitors. CDSSs can be used during the consultation to guide the discussion with the patient about therapeutic alternatives, possible risks and expected benefits, contributing to a shared medical decision. They can integrate multiple variables—from age and comorbidities to biomarkers and functional scores—to generate a personalized risk assessment [2]. Integrating these tools into current oncology practice offers the opportunity to optimize the risk–benefit ratio of therapies, especially in complex or marginal cases. For example, in elderly patients with multiple comorbidities, AI-based CDSSs (AI-CDSSs) can suggest an increased likelihood of severe toxicity with a particular therapy, thus guiding the medical team towards better-tolerated alternatives. At the same time, these systems can alert the clinician to the need for stricter paraclinical monitoring or dose adjustment but do not replace clinical judgment or therapeutic guidelines [3,4].
Another great advantage of AI-CDSSs is their ability for continuous and dynamic updates, by adding new data as they become available through clinical observation in real oncological practice. The clinical scores can be continuously refined through feed-back and this aspect is especially essential in oncology. As real-life data accumulate, AI models can be automatically recalibrated. This allows for continuous improvements to scores and forms, aligning them with evolving therapies and real-world outcomes in the population [5,6]. It is, however, essential that these tools are rigorously validated in real clinical settings, through multidisciplinary collaboration between oncologists, biostatisticians, AI experts and medical informaticians.
But the integration of AI technology in clinical practice needs to be made with responsibility. Algorithmic bias must be considered—AI models may underestimate the risk of certain groups or misinterpret clinical variables, requiring rigorous validation and representative inclusion of data [7]. No matter how performant the AI-CDSSs are [8], they cannot replace detailed clinical exams, direct discussions with the patients and a correct evaluation of each patient’s medical context. Also, international guidelines, which are based on solid clinical evidence, need to remain the basis of any therapeutic decision [9].
The CDSS interface must be intuitive, integrable into electronic health systems, and comply with patient data protection standards. Their acceptability among professionals also depends on the transparency of the algorithms and their ability to provide clear explanations for the recommendations made. A recent study showed an increased level of physician trust in the AI tool when the results of the AI tool concord with the results of randomized controlled trails [10], thus raising the problems of transparency and model explainability. Moreover, the success of AI-CDSSs in oncology is dependent on the quality of available data [11].
Recent advances in artificial intelligence (AI) and machine learning (ML) have led to highly performant clinical decision support systems (CDSSs). However, most of these models are developed in collaboration with IT experts, often resulting in complex black-box systems that are difficult for clinicians to understand and apply in daily practice. As a result, a gap remains between the potential of AI technologies and their adoption in oncology. Our work addresses this gap by presenting a clinician-led proof of concept. The goal was not to design the most accurate or complex predictive model but to demonstrate that contemporary AI tools have become sufficiently user-friendly to be directly applied by physicians without advanced computational skills. In this way, we aim to reduce clinicians’ apprehension toward AI, to show feasibility, and to bridge the professional distance between medicine and computer science.
The aim of this study is to offer a proof of concept on how AI tools can be used for developing clinical interactive and predictive tools, which may in fact represent a major step to truly personalized oncological care, while simultaneously proving that these kinds of tools are now so readily available and intuitive that clinicians with no formal computational training can integrate them into both research and clinical practice with promising results. At the same time, we emphasize that interdisciplinary collaboration with AI specialists is crucial for model enhancement, performance, interpretability and clinical translation.
The main contributions of this study can be summarized as follows:
  • We developed a clinician-led, proof-of-concept CDSS for predicting complications of Bevacizumab using routinely available pretherapeutic data.
  • We tested several machine learning algorithms (logistic regression, Random Forest, XGBoost) and selected the best-performing approach.
  • We derived a simplified logistic regression-based risk score and translated it into an accessible offline HTML calculator for clinical use.
  • We demonstrated feasibility by applying user-friendly AI tools to a real-world oncology dataset without advanced IT expertise.
  • We positioned this work not as a final model but as a step toward reducing clinicians’ apprehension toward AI and fostering future multidisciplinary collaborations.

2. Materials and Methods

2.1. Patient Selection and Data Collection

This study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Ethics Committee of Bucharest Institute of Oncology “Prof. Dr. Alexandru Trestioreanu” (Protocol Number. 24855, Date of approval: 24 November 2022). The design of this study was open, observational, prospective, conducted under real conditions of current oncological practice on patients treated with Bevacizumab or biosimilars, anticancer drugs associated with quite frequent and sometimes serious side effects [12,13]. We included all patients treated for solid malignant tumors at the Bucharest Oncological Institute “Prof. Dr. Al. Trestioreanu (IOB) between 1 January 2023 and 30 June 2024, consecutively. This research was intended to evaluate the usual clinical practice and the real experience of using this drug, which is why the data were extracted from the patients’ medical records and from the hospital’s electronic database. We did not specify exclusion criteria, and all medical decisions regarding treatment, doses, regimens or follow-up were made by the attending physicians without any limitation or influence. We extended the patient follow-up period until 31 December 2024, when the patients were censored prior to the analysis. Thus, we ensured a minimum follow-up period of 6 months.
All patients admitted in IOB agree, by signing a consent form, to the anonymous usage of their medical records and investigation results for scientific, educational and for the preparation and publication of scientific papers, with no additional consent required unless the patient has withdrawn consent.
Patients who received Bevacizumab or biosimilars were identified from pharmacy consumption tables. If a patient received more than one line of therapy containing Bevacizumab, each line of treatment was considered a separate record and analyzed as such. Similarly, if a patient received Bevacizumab for multiple neoplastic diseases, each indication was considered a separate record.
Case-specific data were obtained from patient charts, electronic records, and by consulting laboratory, imaging, histopathological, and genetic results. Data collected included demographics, comorbidities and relevant medical history, neoplastic disease history (tumor origin, histology, immunohistochemistry, genetic characteristics, metastatic sites, diagnostic dates, treatment initiation, progression, death), treatment aspects (line of treatment, combination therapies, dosage, duration of treatment, adverse events, treatment response).
Complications were defined as any documented adverse event considered at least possibly related to Bevacizumab or its biosimilars, regardless of severity, as recorded in patient charts and according to clinical judgment. We initially aimed to model only severe complications, but due to their low incidence, the dataset was statistically underpowered for such stratification. We, thus, included all complications as a binary outcome, which reflects a real-world setting where even moderate events may require dose adjustments or impact quality of life.

2.2. Variable Selection

We used several predictive models developed through machine learning to estimate the risk of complications associated with Bevacizumab, based on pretherapeutic variables, including the following: socio-demographic data of the patients (such as age, sex, background, health insurance status), the presence of personal history (smoking, alcohol, nutritional status), the presence of pathological history (ischemic heart disease, hypertension, heart failure, arrhythmias, valvular diseases, diabetes mellitus, previous treatment with oral anticoagulants/oral antidiabetics/insulin, chronic renal or pulmonary disease, stroke/TIA, thromboembolic events, chronic hepatitis or cirrhosis, other neoplasia than the one being treated, psychiatric history, recurrent urinary tract infections, surgical history), data on the neoplastic disease (TNM classification, location of metastases, stage of the disease, histopathological form, degree of differentiation, neoplastic infiltration of the margins of the resection, presence of lymphovascular or perineural invasion and presence of peritumoral lymphocytic infiltrate on resection pieces/biopsy samples), indication and method of administration of Bevacizumab (unresectable locally advanced neoplastic disease/metastatic disease/recurrent disease, therapeutic line 1/2/3+/maintenance), history of oncological treatments (chemo-radiotherapy, hormone-, immunotherapy, oncological surgical interventions), pretherapeutic biological parameters (hemoglobin, leukocytes, platelets, blood glucose, urea and creatinine, serum Ca2+, transaminases, GGT, presence of coagulation changes or tumor markers) and quality of life of patients prior to initiation of therapy. Unfortunately, although we tried to collect data on tumor receptors, the presence of oncogenic mutations or deficiencies in repair mechanisms, these data could not be used to generate predictive models even when we tried to merge them, due to the massive percentage of missing data (over 80% of patient records do not contain such data, which indicates a lack of accessibility for most patients to advanced genetic testing methods, which may influence their ability to benefit from truly targeted therapies).

2.3. Database Pre-Processing and Leakage Prevention

In an attempt to achieve this objective, we processed the initial database by eliminating all variables that described the type of complications that occurred and their consequences, maintaining only the pretherapeutic characteristics of the patients and diseases that in theory could serve as predictive factors for the occurrence of complications. Any post-treatment or outcome-linked fields were removed a priori to avoid information leakage. Missing values were imputed within the training data only and the fitted imputers were retrained inside each cross-validation fold before being applied to the corresponding validation fold. No outcome-aware univariate filtering or feature selection was performed outside the cross-validation loop.
We also added to the database a recoded binary variable with possible values 0 for the absence of any Bevacizumab-associated complication and 1 for the occurrence of at least one therapy-associated complication regardless of its severity. We did not attempt to create predictive models for severe complications, because the number of cases where they occurred is relatively small, which may lead to a decrease in the predictive power of the models.
A rigorous cleaning of the database was performed in several successive steps, in order to optimize it. Initially, we identified variables that contained a significant number of missing values. Thus, variables encoding certain biological constants such as Na+, K+ and Cl (frequently determined together but absent in over 20% of the records prior to the initiation of therapy) were completely eliminated from the analysis to avoid introducing errors in the imputation process (automatic completion) of the data. Subsequently, for continuous numerical variables with a small number of missing values (<5%), we opted for an imputation method based on the median value of the entire sample—applied to the variables encoding GGT and serum Ca2+. The last step we performed to clean the dataset was an automated analysis of the mean, minimum and maximum values of the continuous numeric variables that highlighted possible outliers. This identified 2 data entry errors (omitting a comma) that were manually corrected after rechecking the values in the patient records.

2.4. Training Predictive Models

We used the optimized database to train several predictive machine learning models:
  • Simple logistic regression—model based on linear relationships between predictors and log-odds, relatively sensitive to outliers and noisy variables, but which offers the advantage of obtaining coefficients that are easily interpretable in a clinical context.
  • Penalized logistic regression with Elastic Net—method that combines the advantages of L1 (sparse, variable selection) and L2 (stability) regularization.
  • Random Forest—robust nonlinear model, based on decision trees, with great capacity to capture complex relationships.
  • XGBoost—boosting-type model (improves errors gradually) that offers the highest accuracy, superior performance on complicated data and good management of imbalances but at the cost of lower interpretability.
Short, clinician-oriented “algorithm primers” summarizing working principles, strengths, and limitations for Logistic/Elastic-Net, Random Forest, and XGBoost are available in previous works [14].
To mitigate moderate class imbalance, we applied class weights (inverse frequency) for Logistic/Elastic-Net and balanced class weights for RF and XGBoost during training. Unless otherwise stated, metrics were computed at a default 0.50 probability threshold; alternative thresholds can trade sensitivity for specificity depending on clinical priorities.
Statistical analysis and development of predictive models were performed using a cloud-based interactive web platform, Jupyter notebook (Google Colab [15]). The Python scripts used in the modeling process were developed with the help of web-based AI code generation assistants available on the platform and online (Gemini 2.0 model [15] and ChatGPT plus 4o [16]), used exclusively under active human supervision and guided by the requirements formulated by the authors. These tools were also used for figure generation. All analytical and methodological decisions, interpretations and scientific writing were made exclusively by the authors.
In the process of developing the predictive models, datasets were initially divided in a proportion of 70% for training the model and 30% for testing, a common choice that provides a good balance between learning and evaluation. However, to improve the models’ learning ability (especially in the context of class imbalance), we decided to explore the 80/20 split version (which allows for training the models on a larger number of cases and, at least in theory, offers the possibility of achieving superior performance to the 70/30 split).
In an attempt to obtain the best performance of each model, we decided to optimize the models as follows:
  • Logistic regression: Inverse weights were applied for the classes (to correct class imbalance) and L1 (Lasso), L2 (Ridge) and Elastic-Net-type penalties were tested (to reduce the overfitting effect and select the relevant variables); it is a method that combines the advantages of L1 (sparse regularization, variable selection) and L2 (stability).
  • Random Forest: Hyperparameters were adjusted (such as the number of trees, maximum tree depth, minimum number of samples per leaf), and weights were applied to reduce minority class classification errors.
  • XGBoost: Boost parameters (learning_rate, max_depth, n_estimators, subsample) and class weights were adjusted to improve performance in detecting cases with complications.
Hyperparameters for each algorithm were tuned using a grid search approach on the training set. The final parameter settings are summarized in Supplementary Table S1.

2.5. Comparing the Performance of Predictive Models

All models were tested comparatively (in both unoptimized and optimized variants) by applying the same evaluation procedure on an unrevised test set (hold-out technique) comprising either 20 or 30% of the initial data depending on the model split. The following indicators were used to evaluate the performance of each model:
  • Overall accuracy.
  • AUC ROC (area under the receiver operating characteristic curve) score—to measure the overall ability to discriminate between classes.
  • Sensitivity (recall) for class 1 (ability to correctly identify patients with complications)—essential in a medical context, as the inability to identify cases that have the potential to suffer complications can have serious clinical consequences.
  • Specificity (ability to correctly identify cases without complications)—clinically significant, as it confirms the possibility of safe drug administration.
  • F1-score for class 1—indicates the balance between precision and sensitivity.
  • Error rate—indicates the total proportion of incorrect classifications (false positive and false negative).
Following the comparative analysis, we identified the model with the best performance (optimized Random Forest 80/20), which was subsequently used to generate clinical tools with practical applicability—a risk score and an interactive HTML form that provides access to an automatic risk calculator.

2.6. Internal Validation

To provide a more robust estimate of model generalizability and reduce the risk of overfitting, we subsequently applied a 5-fold stratified cross-validation procedure across the entire dataset. Performance metrics (AUC, sensitivity, specificity, accuracy) were calculated as the mean across folds. To assess stability, we also performed repeated 5 × 5 CV (25 runs). In addition, we generated out-of-fold (OOF) predictions across all runs and computed a bootstrap 95% CI for AUC (B = 2000). While a nested cross-validation framework would have been more rigorous, this was not feasible in the current proof of concept due to dataset size limitations.

3. Results

3.1. General Description of the Cohorts and Distribution of the Target Variable

The analytic dataset comprised 395 treatment episodes from a single center, of which 177 (44.8%) developed at least one bevacizumab-related complication and 218 (55.2%) did not. Each row represents one treatment episode; when patients contributed multiple episodes, these were handled as independent observations after confirming no data leakage across the train–validation partitions. The primary outcome was the occurrence of any bevacizumab-related complication during the observation window.
The overall sample contained 130 male and 265 female patients. The patients mostly came from urban areas (341 patients—86.33%), suggesting an inequality of access to oncological medical services for the rural population. Further, 271 patients were retired at the time of treatment, 90 were employees and 34 were without insurance, benefiting from access to complex treatments through the National Oncology Program (the fact that 8.61% of the total patients belong to this group suggests that the program is functional in the IOB). Figure 1 presents an age histogram in the studied cohort. From the age histogram, it is observed that the distribution of this variable is relatively normal with a slight negative asymmetry (skewness = −0.18). A relatively normal distribution indicates that our group is representative for the statistical analysis, the age of the patients not being a bias factor. The age of the patients ranged from 22 to 88 years, with a mean age of 61.6 ± 10.88, which confirms that the majority of patients treated with Bevacizumab for solid tumors were in their sixth decade of life, which coincides with the usual age of diagnosis in oncology. The presence of young patients (under 40 years), although a minority, proves that there are advanced cases in young populations as well—these patients have certain psychosocial characteristics that must be taken into account when establishing the therapeutic plan. The fact that the two groups do not appear to be statistically different (p = 0.475) is important because the correlation between advanced age and patients’ comorbidities is known. The general fragility of the body induced by advancing age may, at least in theory, increase the frequency of complications associated with Bevacizumab therapy.
Table 1, Table 2 and Table 3 present the distribution of pretherapeutic variables, based on which we analyzed the global sample. We observe that most pretherapeutic variables have a relatively balanced distribution between the two groups. The only variables that seem to be more frequent in the group with complications are cardiac and thromboembolic history, chronic treatment with anticoagulants, pre-existing renal damage, low serum hemoglobin values and certain tumor characteristics (colorectal or cervical location of the primary tumor, poor/absent cellular differentiation and more aggressive histologies).
Table 4 summarizes the number, type and severity of complications occurring in our patients under Bevacizumab/biosimilar therapy.

3.2. Comparative Interpretation of Predictive Models’ Performances

Model optimization had heterogeneous effects on performance: while the Random Forest model benefited significantly from hyperparameter tuning (optimization resulted in an increase in AUC from 0.66 to 0.75, while maintaining a good balance between sensitivity and specificity, thus resulting in the best-performing predictive model), no notable improvements in performance metrics were observed for predictive models based on logistic regression or XGBoost. Moreover, in the case of the XGBoost 80/20 model, a reduction in performance was observed after hyperparameter tuning (although the AUC remained at approximately the same value, the sensitivity of the model decreased to 0.400).
In the course of trying to improve the models’ learning ability (especially in the context of class imbalance), we decided to explore the 80/20 split version (which allows for training the models on a larger number of cases and, at least in theory, offers the possibility of obtaining superior performance to the 70/30 split initially used). Comparing identical models trained on two different data splits provides the following conclusions: models trained on the 80/20 split generally perform better than their counterparts trained on the 70/30 split, both in terms of AUC-ROC and sensitivity, providing a better balance between learning and evaluation.
The comparative performances of all trained predictive models are presented in Figure 2 in the form of a heatmap for easier visualization.

3.3. Performances of the Optimal Predictive Model

After training various predictive machine learning models, we found that the Random Forest model with an 80/20 split and optimized by adjusting hyperparameters and applying class weights to reduce classification errors of the minority class achieved the best prediction performance for the probability of complications associated with therapy.
Paired DeLong tests on the common validation set showed no statistically significant differences in AUC among RF, Logistic/Elastic-Net, and XGBoost (all p > 0.05). McNemar tests for accuracy at the 0.50 threshold were also non-significant across pairwise contrasts, supporting the interpretation that observed differences are modest; RF was, therefore, selected for downstream translation due to its balanced sensitivity/specificity and amenability to SHAP-based explanations
The overall performance of the RF 80/20 optimized model was as follows:
  • Overall accuracy: 70.63%.
  • Error rate: 29.37%.
  • Sensitivity (recall)—detection of patients with complications: 66.67%.
  • Specificity (detection of patients without complications): 73.85%.
  • AUC ROC: 0.75.
  • F1-score for class 1 (patients with complications): 0.67.
  • Precision: 0.673.
The optimized Random Forest 80/20 model, trained on the group of 395 patients (177 presenting complications and 218 who did not develop complications), demonstrated moderate predictive performance but very well balanced between classes. Moreover, the overall accuracy of 70.63% tells us that the model manages to correctly classify 7 out of 10 patients. The recall value for the minority class of 66.67% means that the model manages to detect two out of three patients who will develop complications associated with therapy, while the specificity of 73.85% shows a satisfactory ability to exclude patients who will not develop complications (correctly detecting three out of four patients in this category). The clinical importance of these results, especially in the context of oncology practice, can be considerable, avoiding unnecessary degradation of the patient’s quality of life and interruption of systemic therapy. Such a model can be successfully used as a tool for initial triage of patients, but the error rate of 30% requires caution in its use as the sole decision-making criterion. The accuracy of 67.3% of the model (a value comparable to the sensitivity) suggests that our model is well calibrated between detecting cases and avoiding, as much as possible, false alarms. And the F1-score value of 67%, being the harmonic mean between sensitivity and precision, supports the conclusion that the optimized Random Forest 80/20 model has a satisfactory, balanced performance, without being unbalanced in a certain direction (increased precision at the expense of poor recall or vice versa). Such a model finds its clinical applicability in situations where both risks (of missing certain cases or of generating false alarms) are important. Finally, the AUC ROC value (0.75) indicates that this model has a reasonable discriminative power (well above the random threshold of 0.5) in a varied range of decision thresholds.
In terms of possible errors, the model records both false-negative results (patients who will develop complications but whom the model incorrectly labeled as risk free) and false-positive results (patients who will not develop complications but whom the model classifies as at risk). In the context of current oncological practice, false-negative results can have serious consequences through the lack of adequate monitoring or the absence of treatment adjustment. In contrast, false-positive results, although not inherently dangerous for the patient, can cause unnecessary anxiety, additional costs associated with the medical act generated by over-investigation and excessively cautious monitoring. In Figure 3, we present a confusion matrix of the optimized Random Forest 80/20 model, which gives us an overview of these potential errors.
Based on the above, we can conclude that the optimized Random Forest 80/20 model can be used as an automated screening tool that allows the treating physician to identify patients they intend to treat with Bevacizumab who are prone to develop complications associated with therapy, contributing to dose adjustment and individualization of monitoring protocols. This model offers the most robust predictive performance, combined with a good balance between detecting cases at risk (sensitivity) and avoiding false alarms (specificity). However, the moderate performances suggest that the model can be used strictly to complement clinical assessment and not as a substitute for it.

3.4. Model Validation

To assess generalizability, we complemented the initial 80/20 hold-out with internal cross-validation. In 5-fold stratified cross-validation, the optimized Random Forest model achieved an AUC of 0.62 ± 0.10, accuracy 0.60 ± 0.08, sensitivity 0.43 ± 0.11, specificity 0.73 ± 0.09, precision 0.57 ± 0.10, and F1 0.49 ± 0.10 (Supplementary Table S2). Repeating the procedure (5 × 5 cross-validation; 25 runs) yielded nearly identical estimates: AUC 0.62 ± 0.07, accuracy 0.59 ± 0.06, sensitivity 0.43 ± 0.08, specificity 0.72 ± 0.06, precision 0.56 ± 0.08, and F1 0.48 ± 0.08 (Supplementary Table S3). Out-of-fold predictions aggregated across all runs yielded a global AUC of 0.612 with a bootstrap 95% CI of 0.555–0.665 (B = 2000), overlapping the cross-validated means and supporting the stability of our estimates (Supplementary Table S4).
Class proportions were preserved across folds (positives ≈45%), indicating that the observed trade-off—higher specificity than sensitivity—reflects the model rather than sampling artifacts. This profile suggests a ‘rule-in’ orientation: positive flags are clinically meaningful for intensifying monitoring in higher-risk patients, while negative predictions should not be used to de-escalate care.
These findings are in line with this study’s proof-of-concept aim—demonstrating clinician-led feasibility on real-world data—rather than maximizing performance. Future multicenter external validation, feature enrichment, and threshold calibration are expected to improve sensitivity and overall utility.

3.5. Model Interpretability (SHapley Additive exPlanations—SHAP)

To examine how predictions were formed, we generated a SHAP summary plot for the optimized Random Forest model (Figure 4). This step increases clinical transparency and aligns with current recommendations for explainable AI in medicine [17,18,19].
Two variables concentrate most of the explanatory signal: age (Varsta) and diagnosis (Diagnostic). In the plot, each dot is one patient; the horizontal position represents the direction and magnitude of that feature’s contribution to the predicted probability of complications (positive values push the prediction toward “complication”), while color encodes the feature value (blue = low, red = high).
Overall, higher age values systematically shift predictions toward higher risk, whereas younger age shifts them downward. For diagnosis, certain diagnostic categories tend to push risk upward while others push it downward, reflecting the expected heterogeneity across tumor sites/contexts. Importantly, the absolute SHAP magnitudes are small (roughly within ±0.05), indicating no single dominant driver and aligning with the modest discrimination observed in cross-validation (AUC ≈ 0.62). In other words, risk is multi-factorial, with several routine variables contributing incrementally rather than any one predictor determining the output.
From a clinical standpoint, these explanations are plausible and transparent: older or diagnostically less favorable profiles nudge predictions upward, supporting the tool’s intended use for risk awareness and triaging of monitoring intensity rather than dictating therapy. The plot, thus, illustrates that our clinician-led model behaves in a way that is consistent with clinical intuition while remaining interpretable for non-IT users.

3.6. Translating the Predictive Model into Clinical Tools Applicable in Oncological Practice

Following the comparative analysis, we identified the model with the best performance (optimized Random Forest 80/20), which was subsequently used to generate clinical tools with practical applicability—a risk score and an interactive HTML form that provides access to an automatic risk calculator. This step-by-step process consisted of the following:
  • Selection of major predictive variables: The first step in the construction of the clinical score consisted of extracting the most relevant predictive variables by analyzing the importance coefficients of the variables generated by the selected model. We were, thus, able to identify variables that the predictive model considered to have greater predictive power. Of these, those that were easily accessible in current practice, available pretherapeutically and clinically significant were retained. The list of biological parameters considered predictive included the following: hemoglobin level, platelet and leukocyte count, blood glucose, creatinine, urea and transaminase values. Clinical parameters, such as age, location of neoplastic disease, stage of disease and degree of tumor differentiation, were also introduced into the list.
  • Construction of the initial clinical score: For each selected numerical variable, binary thresholds were defined (e.g., Hg < 11 g/dL, urea ≥ 40 mg/dL), based on the distribution in the dataset and clinical expertise. Then, each category was assigned a score proportional to its relative importance in the optimized Random Forest 80/20 model. The scores were aggregated into a cumulative score, which was subsequently divided into three risk classes: low (0–2 points), intermediate (3–5 points) and high (≥6 points).
  • Testing of the initial clinical score: The score was applied retrospectively to the entire cohort to evaluate its performance in identifying patients at risk. We analyzed the distribution of patients according to score, the actual complication rates in each risk group and the statistical performances (AUC ROC, accuracy, sensitivity and specificity). The performances of the initial clinical score were as follows: overall accuracy—59%, AUC ROC—0.555, sensitivity (recall for class 1)—47.2%, specificity—60.8%, F1-score class 1—0.502. These results suggest a low capacity to triage patients exposed to Bevacizumab according to the risk of developing specific complications. Moreover, since the initial score has a discrete nature, with fixed thresholds, the premises of some limitations in the fine representation of individual risk are created. These aspects motivated us to try to optimize the score.
  • Optimization of clinical score performances: The refinement of the clinical score was performed by applying a logistic regression model, using the derived raw score as predictors. A calibrated probability for the occurrence of complications was, thus, generated, which allowed us to eliminate arbitrary thresholds and generate a continuous risk estimate and generate a new logistic-derived clinical score. The advantages of the logistic-derived score are as follows: personalized probabilities of risk occurrence, better separation of high-risk from low-risk cases, and the possibility of validation through calibration (observed vs. predicted ratio). A simplified logistic regression-based risk score was derived from the multivariate logistic model. Regression coefficients (β) were scaled and rounded to obtain integer weights, which were then summed into a total score. Cut-off points for risk stratification (low, intermediate, high) were selected based on Youden’s J index from ROC analysis.
  • Performance evaluation of the optimized score: The performances of the new score were compared with the initial score and with the Random Forest model directly through ROC curves, analysis of the score distribution and performance indicators—accuracy, sensitivity, specificity, F1-score. These evaluations showed superior performances to the initial score, comparable to those of the Random Forest model, but with the additional advantage of clinical portability and increased interpretability (Figure 5 and Figure 6 and Table 5). When analyzing the comparative distribution of scores, we find that for the initial score, no clear separation is observed (the scores of patients with and without complications often overlap), indicating a limited discrimination capacity, with a close mean between groups and a large variation. In contrast, the logistic-derived score shows a clearer separation between groups. Patients who developed complications tend to have higher probability scores, while patients without complications are concentrated in the lower part of the distribution.
Validation of the probabilities predicted by the logistic-derived score: In order to best assess whether the probabilities given by the logistic-derived score reflect the real risk of complications, but also for a better comparison with the initial score, we generated the comparative calibration plot shown in Figure 6.
The calibration plot shows the agreement between the probabilities of complications predicted by each score and their actual observed frequency. The dotted diagonal line represents the ideal scenario, in which there is perfect agreement (the model accurately predicts the risk in all cases). In Figure 6, we observe that the logistic derivative score curve approaches the ideal line, which indicates both increased discrimination power (the model separates patients with and without complications quite well) and the fact that the score correctly estimates the probability of risk. Minor deviations from the ideal line in certain areas can be attributed to the natural imbalance existing in the dataset, but this does not significantly affect the overall performance. Therefore, we concluded that the optimized score has a robust calibrated prediction capacity, which makes it suitable for clinical applications.

3.7. Building an Interactive Form Applicable in Oncologic Practice

Based on the logistic-derived score, we built a digital HTML form with a responsive design (which allows for automatic adaptation to the electronic device from which it is accessed—mobile, tablet, laptop), accessible and offline, in which the doctor can tick the presence of risk factors, and the system automatically returns the cumulative score, risk class and estimated percentage risk. The form is easy to use and has the potential to be integrated into oncological practice, offering treating physicians a tool for pretherapeutic stratification of the risk of developing therapeutic complications. Table 6 and Table 7 present the final version of the logistic-derived score (variables, categories and scores awarded) and the clinical risk thresholds.
Supplementary Material File S1 consists of an interactive HTML form of the risk calculator based on the clinical risk score, proving its functionality.

4. Discussion

In the continuous effort to ameliorate post-therapeutic results [20,21,22], the oncology landscape is rapidly evolving, with a huge influx of real-time clinical, paraclinical, imaging, genetic and molecular data that exceed the capacity of the standard decision-making processes of the human physician. Advanced AI tools can synthesize this complexity and support personalized treatment decisions, without replacing clinical experience, normal interaction between patient and physician or evidence-based international guidelines. Various forms of AI tools have been employed in developing CDSSs that can be integrated into oncological practice with ease and playing multiple roles:
  • Machine learning algorithms (Random Forest, Gradient Boosting such as XGBoost, Support Vector Machines, logistic regression)—used for developing predictive models based on historical data. They work by identifying patterns in a dataset and applying them to new cases, providing quantifiable and reproducible results. Possible applications include prediction of risk of toxicity/adverse effects, prognostic stratification of patients or classifying therapeutic response to various agents. One such example is xDECIDE, which is a web-based platform that combines AI power with human clinical expertise and offers personalized oncological therapeutic recommendations. Studies have proven that this platform has real impact on decision making; changes in the initial treatment plan proposed by the clinician were recorded after using the platform [23,24].
  • Deep Learning is an extension of machine learning, which employes complex neural networks with multiple layers to learn highly abstract representations of data. These models, such as CNN (Convolutional Neural Networks) or LSTM (Long Short-Term Memory), are capable of analyzing medical images (such as CT, MRI, histopathology) or time-series data (such as evolution of biological parameters). Deep Learning is used for the automatic recognition of anomalies and interpretation of sequential data. Deep Learning can be used for automatic recognition of lesions on CT scans and MRI or for early detection of toxicity signs based on serial analyses. An example of real applications of Deep Learning-based CDSS is DeepMind Health (Google), which employs Complementarity-Driven Deferral to Clinical Workflow (CoDoC) for screening of breast lesions. Compared to double reading with arbitration, the predictive model reduced false-positive rates by 25% and reduced clinician workload by 66% [25]. Clinical trials on the deployment of Deep Learning CDSSs in healthcare are underway with encouraging results [26].
  • Natural Language Processing (NLP)—allows AI to “understand” and process human written speak, thus being able to extract valuable information from unstructured medical records (charts, clinical or trial notes, guidelines) and transforming the data into structured data usable for CDSS development. NLP tools, such as Named Entity Recognition (NER), text summarization or Embedding models (e.g., Word2Vec, BERT), can be used for automatic extraction of data from medical records, identifying eligible patients for clinical trials and generation of guideline-concordant therapeutic recommendations [27,28,29,30,31,32].
  • Large Language Models (LLMs)—recent developments like ChatGPT, Flamingo, PaLI or Gemini led to the emergence of conversational AIs capable of coherent text generation and real-time synthetization of data from multiple sources. LLMs can be incorporated into interactive interfaces where the clinician can interact with the system, ask questions, demand explanations or viable alternatives for the proposed clinical/therapeutic decision. LLMs can also be used for automatic generation of documentation, thus reducing clinician workload. AMIE and MedGemini (medical LLMs created by Google DeepMind in 2025) have managed to generate complex oncologic therapy plans in more than 90% of the simulated cases and surpass human experts in text summarization and document generation [33,34,35,36].
  • Reinforcement Learning (RL)—a lesser-known but very promising type of AI, which learns through interactions and rewards. It can be used in scenarios in which the therapeutic decision needs to be dynamically adaptative (e.g., adjusting doses based on individual patients’ response and reaction) [37,38,39]. This type of AI resembles the way a physician refines his clinical intuition after gaining experience based on the results of previous decisions.
  • In order for AI to be truly useful in clinical practice, it needs to be transparent, and the variable which formed the bases for a decision needs to be easily explained to the human doctor/patient. Higher transparency will ensure better adoption of AI technologies into clinical practice. Explainable AI (XAI), through technologies such as LIME or SHAP, has precisely this role [40,41].
  • At the heart of all these technologies is Data Engineering—the process of collecting, cleaning, and shaping medical data. Without this step, no AI model will work properly. Libraries like Pandas, NumPy or TensorFlow Data are used to convert raw data into formats that can be used by algorithms.
  • Finally, in the context of multicenter collaboration, Federated Learning offers an ethical and safe solution to train AI models on distributed databases (data from multiple medical centers) without sharing patient information between institutions (e.g., Flower, NVIDIA Clara, TensorFlow Federated). In this way, hospitals can contribute to shared models while respecting data privacy under GDPR or HIPAA regulations.
These types of AI are not competing but complementary to each other and to human clinical expertise. Each adds an essential piece to the complex puzzle of assisted medical decision making.

4.1. Related Work and State-of-the-Art (SOTA) Approaches

Recent frameworks, such as the Yonsei Cancer Data Library, demonstrate how continuous multimodal data integration empowers evidence-based guidance across over 170,000 cancer cases [42]. Combined, multiple AI tools enable the transition from standard guidelines to personalized real-time assistance, supporting the oncologist in their mission to provide safe, effective and tailored care for each patient [2,4,5,9,43,44]. In Table 8, we present some of the more prominent usage of AI-CDSS in oncology.
While these SOTA systems illustrate the remarkable potential of AI in medicine, our work positions itself differently, as a clinician-led proof of concept, showing that accessible AI tools can already be integrated into everyday oncology practice, even outside large research consortia or IT-driven environments. The distinctive feature of this study is that it was designed and conducted entirely by clinicians, without formal training in computational sciences or artificial intelligence. This underscores the important shift in the field: AI-based technologies have become sufficiently accessible and intuitive to be employed by non-specialists using real-world clinical data (especially once clinicians surpass their initial reticence and embrace the new technologies). However, we recognize that collaboration with AI experts is crucial to further model enhancement, optimization, interpretability, validation and clinical implementation. Integration of AI-CDSS into oncological practice offers a transformative opportunity, but to ensure clinical relevance and performance over time, a dynamic, feed-back-driven continuous development approach must be considered and implemented. In Figure 7, we propose a structured workflow, which allows for continuous refinement of AI-CDSS-based clinical scores for predicting complications associated with oncologic agents.
In order for it to be possible to integrate new data and improve the model over time, first, we need to establish an adaptive clinical data repository (data lake) that captures both baseline and outcome variables from real-world clinical practice. The database needs to be designed for expansion with the inclusion of new key clinical parameters as they emerge from evolving research or clinical observation. As recommended by frameworks for AI-CDSS, incoming real-world clinical data fed into adaptive databases are the bases for incremental learning and retraining of predictive models [54,55]. By design, database architecture should allow for modular data expansion, thus allowing the incorporation of new emerging biomarkers or clinical variables without disrupting the existing workflows [4,23,42]. An upgradable AI-CDSS driven by clinician feed-back, real-life data accrual and periodic model retraining offers a scalable and ethical approach to maintaining the relevance of the tool. This dynamic framework ensures that the CDSS does not become outdated but rather evolves parallel to medical evidence, therapeutic innovations and patient-specific complexity.

4.2. Study Strengths, Limitations and Future Directions of Research

This study has several notable strengths. First, the model is built exclusively on real-world, prospectively collected clinical data and uses only pretherapeutic variables, thus enabling early risk stratification before the actual initiation of Bevacizumab/biosimilars therapy. A key innovation is that the entire model development was carried out by clinicians with no formal computational training, underscoring the growing accessibility and usability of AI tools in daily medical practice. Moreover, the model was translated into intuitive, offline, HTML interactive tools, with immediate deployment capability, designed for direct clinical use. And, while the predictors identified by the model (e.g., age, anemia, renal function, tumor differentiation) are well known to oncologists, the added value of this tool lies in its ability to synthesize multiple variables into a single, individualized risk estimate that can aid in triaging and shared decision making, especially in borderline or frail patients. Special attention is given to model transparency, simplicity and clinical relevance, bridging the gap between theoretic modeling and clinical practice. As such, this study does not position itself as a technical AI project but rather as a clinician-led proof of concept, demonstrating how AI tools can now be realistically implemented in oncology care. At the same time, we emphasize the value of interdisciplinary collaboration with AI experts to further refine model performance, interpretability and clinical integration. This combination of accessibility, clinical grounding and translational focus contributes to the growing body of evidence that AI can support (and not replace) medical expertise in personalizing cancer care.
The most important limitations of our study are as follows:
  • The most significant limitation of our study is the fact that our risk score is not developed using an adaptive database, and, as a result, our predictive models cannot be retrained once new clinical data become available. This was done to highlight that even in the absence of a trained AI expert, AI tools have become so readily available and intuitive that medical professionals can employ them in their current research and practice. However, collaboration with AI experts is highly recommended for optimal results.
  • Relatively small and unbalanced sample size: The relatively modest size of the dataset and the unequal distribution between patients in the group with and without complications (unbalanced classes) may limit the generalizability of the predictive models that we trained. While 395 treatment episodes were included, we acknowledge that the sample size may still limit the statistical power of the model, particularly in edge-case scenarios. This reinforces the need for further multicenter validation.
  • The data come from a single medical unit, which introduces an institutional selection bias, which reduces the external validity of the results obtained and increases the risk of overfitting.
  • Absence of potentially relevant variables: Some clinical and biological information that could have predictive value for the occurrence of complications (such as concomitant treatments, detailed ECOG score, immunohistochemical factors, the presence of oncogenic mutations and mutation repair deficiencies) was available for a very small number of patients, which made them statistically unusable.
  • The final risk score has not been validated on an external dataset, nor has it been prospectively tested in clinical practice to prove its reliability.
These limitations also lead to future directions for further research:
  • Implementing a data-lake adaptive system will allow for continuous refinement of clinical tools developed and maintaining clinical relevance.
  • Increasing the sample size by introducing data from other medical centers (multicenter study) could strengthen the robustness and generalizability of the predictive models.
  • Prospective validation of the score on an external sample is necessary to test its performance and utility in bedside decision making. The current study represents a single-center proof of concept. Future work should aim to externally validate the model on independent cohorts, ideally through multicentric collaborations.
  • Intense collaboration with IT experts would mainstream the process of developing better AI tools.
  • Future versions of the predictive model could include other variables (such as immunohistochemical and genetic markers or other imaging, clinical and therapeutic data) to increase the accuracy of predictions
  • The development of offline digital applications such as the one already created paves the way for personalized decision support tools, easy to use for clinicians.

5. Conclusions

The present study has proven the possibility of constructing and internally validating a clinical score that is easy to introduce into current oncological practice, using predictive models based on machine learning algorithms, capable of estimating the risk of patients developing a complication associated with Bevacizumab or biosimilar therapy based on the data usually available pretherapeutically. Through a rigorous, staged approach, we obtained a clinical tool with practical value that is very easy to implement. Moreover, the transformation of the score into a digital HTML form with a responsive design (which allows automatic adaptation to the electronic device from which it is accessed—mobile, tablet, laptop), accessible and offline, facilitates clinical implementation and makes it directly usable by physicians within the therapeutic decision-making process. The clinical risk score facilitates good risk stratification and allows for personalized management of the oncology patient treated with Bevacizumab or biosimilars. Our tool is not designed to dictate therapeutic decisions but rather to support risk awareness and individualization of monitoring strategies. Even a somewhat modest predictive performance (70% accuracy with balanced sensitivity/specificity) can help avoid underestimating risks in complex patients and facilitate shared decision making between doctor and patient.
The present study is a clinician-led proof of concept aimed at demonstrating that contemporary AI tools can be safely explored by physicians with transparent outputs, not an IT state-of-the-art benchmark. This study is single center and lacks external validation; discrimination is modest, with a specificity-dominant profile (rule-in orientation). Although not meant as a substitute for clinical evaluation and paraclinical monitoring of oncologic patients, these tools may in fact provide a therapeutic decision aid for physicians and adjust the risk/benefit ratio of targeted therapies, thus potentially being the next step towards truly personalized medicine.
In sum, the present work is not intended to replace clinicians or IT experts but to reduce apprehension and bridge disciplines by showing that user-friendly AI can already support risk awareness and monitoring triage in routine oncology. Superior models will require—and we explicitly invite—multidisciplinary collaboration and external validation.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics15182391/s1, Supplementary Table S1: Final hyperparameter settings for model optimization; Supplementary Table S2: Cross validation performance; Supplementary Table S3: Repeated CV Performance; Supplementary Table S4: OOF Bootstrap AUC; Supplementary File S1: Interactive HTML form.

Author Contributions

Conceptualization, E.C.; methodology, E.C. and D.-C.S.; formal analysis, E.C., V.R., G.G., S.-O.I. and A.G.; software, E.C., A.G., S.-O.I. and G.G.; writing—original draft preparation, E.C.; writing—review and editing, V.R., S.-O.I., L.S. and E.C.; supervision, L.S. and D.-C.S. All authors have read and agreed to the published version of the manuscript.

Funding

Publication of this paper was supported by the University of Medicine and Pharmacy “Carol Davila” through the institutional program “Publish not Perish”.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Ethics Committee of Bucharest Institute of Oncology “Prof. Dr. Alexandru Trestioreanu” (Protocol Number. 24855, Date of approval: 24 November 2022).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets presented in this article are not readily available because the data supporting the findings of this study are not publicly available due to institutional restrictions. Requests to access the datasets should be directed to corresponding author V.R.

Acknowledgments

The training of the predictive models based on machine learning and the development of the clinical score and interactive form were possible with the help of artificial intelligence tools (ChatGPT plus—v.4o and Gemini) and freely available cloud-based interactive web platform Jupyter notebook (Google Colab). The authors reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
AI-CDSSArtificial Intelligence-based Clinical Decision Support System
AMIEArticulate Medical Intelligence Explorer
AUCArea Under the Curve
BERTBidirectional Encoder Representations from Transformers
CDSSClinical Decision Support System
CNNConvolutional Neural Network
CTComputed Tomography
DVTDeep Vein Thrombosis
ECOGEastern Cooperative Oncology Group (performance status score)
EGFREpidermal Growth Factor Receptor
EHRElectronic Health Record
F1F1-Score (harmonic mean of precision and recall)
GDPRGeneral Data Protection Regulation
GGTGamma-Glutamyl Transferase
HIPAAHealth Insurance Portability and Accountability Act
HTMLHyperText Markup Language
LIMELocal Interpretable Model-agnostic Explanations
LLMLarge Language Model
LSTMLong Short-Term Memory
MINDACTMicroarray In Node-negative and 1 to 3 positive lymph node Disease may Avoid ChemoTherapy
MLMachine Learning
MRIMagnetic Resonance Imaging
NERNamed Entity Recognition
NLPNatural Language Processing
NSCLCNon-Small Cell Lung Cancer

References

  1. Nafees, A.; Khan, M.; Chow, R.; Fazelzad, R.; Hope, A.; Liu, G.; Letourneau, D.; Raman, S. Evaluation of clinical decision support systems in oncology: An updated systematic review. Crit. Rev. Oncol. Hematol. 2023, 192, 104143. [Google Scholar] [CrossRef] [PubMed]
  2. Oehring, R.; Ramasetti, N.; Ng, S.; Roller, R.; Thomas, P.; Winter, A.; Maurer, M.; Moosburner, S.; Raschzok, N.; Kamali, C.; et al. Use and accuracy of decision support systems using artificial intelligence for tumor diseases: A systematic review and meta-analysis. Front. Oncol. 2023, 13, 1224347. [Google Scholar] [CrossRef] [PubMed]
  3. Elhaddad, M.; Hamam, S. AI-Driven Clinical Decision Support Systems: An Ongoing Pursuit of Potential. Cureus 2024, 16, e57728. [Google Scholar] [CrossRef] [PubMed]
  4. Karuppan Perumal, M.K.; Rajan Renuka, R.; Kumar Subbiah, S.; Manickam Natarajan, P. Artificial intelligence-driven clinical decision support systems for early detection and precision therapy in oral cancer: A mini review. Front. Oral Health 2025, 6, 1592428. [Google Scholar] [CrossRef]
  5. Verlingue, L.; Boyer, C.; Olgiati, L.; Brutti Mairesse, C.; Morel, D.; Blay, J.Y. Artificial intelligence in oncology: Ensuring safe and effective integration of language models in clinical practice. Lancet Reg. Health Eur. 2024, 46, 101064. [Google Scholar] [CrossRef]
  6. Kann, B.H.; Hosny, A.; Aerts, H.J.W.L. Artificial Intelligence for Clinical Oncology. Cancer Cell 2021, 39, 916. [Google Scholar] [CrossRef]
  7. Cross, J.L.; Choma, M.A.; Onofrey, J.A. Bias in medical AI: Implications for clinical decision-making. PLoS Digit. Health 2024, 3, e0000651. [Google Scholar] [CrossRef]
  8. Abdeldjouad, F.Z.; Brahami, M.; Sabri, M. Evaluating the Effectiveness of Artificial Intelligence in Predicting Adverse Drug Reactions among Cancer Patients: A Systematic Review and Meta-Analysis. In Proceedings of the 2023 IEEE International Conference Challenges and Innovations on TIC, I2CIT 2023, Hammamet, Tunisia, 21–23 December 2023. [Google Scholar] [CrossRef]
  9. Wang, L.; Chen, X.; Zhang, L.; Li, L.; Huang, Y.B.; Sun, Y.; Yuan, X. Artificial intelligence in clinical decision support systems for oncology. Int. J. Med. Sci. 2023, 20, 79–86. [Google Scholar] [CrossRef]
  10. Hussain, Z.; Lam, B.D.; Acosta-Perez, F.A.; Riaz, I.B.; Jacobs, M.; Yee, A.J.; Sontag, D. Evaluating Physician-AI Interaction for Cancer Management: Paving the Path towards Precision Oncology. arXiv 2024, arXiv:2404.15187. [Google Scholar] [CrossRef]
  11. Grüger, J.; Geyer, T.; Brix, T.; Storck, M.; Leson, S.; Bley, L.; Braun, S.A. AI-Driven Decision Support in Oncology: Evaluating Data Readiness for Skin Cancer Treatment. arXiv 2025, arXiv:2503.09164. [Google Scholar] [CrossRef]
  12. Chitoran, E.; Rotaru, V.; Stefan, D.C.; Gullo, G.; Simion, L. Blocking Tumoral Angiogenesis VEGF/VEGFR Pathway: Bevacizumab—20 Years of Therapeutic Success and Controversy. Cancers 2025, 17, 1126. [Google Scholar] [CrossRef]
  13. Chitoran, E.; Rotaru, V.; Ionescu, S.-O.; Gelal, A.; Capsa, C.-M.; Bohiltea, R.-E.; Mitroiu, M.-N.; Serban, D.; Gullo, G.; Stefan, D.-C.; et al. Bevacizumab-Based Therapies in Malignant Tumors—Real-World Data on Effectiveness, Safety, and Cost. Cancers 2024, 16, 2590. [Google Scholar] [CrossRef] [PubMed]
  14. Shateri, A.; Nourani, N.; Dorrigiv, M.; Nasiri, H. An Explainable Nature-Inspired Framework for Monkeypox Diagnosis: Xception Features Combined with NGBoost and African Vultures Optimization Algorithm. arXiv 2025, arXiv:2504.17540. [Google Scholar] [CrossRef]
  15. Welcome To Colab-Colab, n.d. Available online: https://colab.research.google.com/notebooks/intro.ipynb (accessed on 23 July 2025).
  16. ChatGPT, n.d. Available online: https://chatgpt.com/?model=gpt-4o (accessed on 23 July 2025).
  17. Farzipour, A.; Elmi, R.; Nasiri, H. Detection of Monkeypox Cases Based on Symptoms Using XGBoost and Shapley Additive Explanations Methods. Diagnostics 2023, 13, 2391. [Google Scholar] [CrossRef] [PubMed]
  18. De Paola, L.; Treglia, M.; Napoletano, G.; Treves, B.; Ghamlouch, A.; Rinaldi, R. Legal and Forensic Implications in Robotic Surgery. Clin. Ter. 2025, 176, 233–240. [Google Scholar] [CrossRef]
  19. Marinelli, S.; De Paola, L.; Stark, M.; Vergallo, G.M. Artificial Intelligence in the Service of Medicine: Current Solutions and Future Perspectives, Opportunities, and Challenges. Clin. Ter. 2025, 176, 77–82. [Google Scholar] [CrossRef]
  20. Cirimbei, C.; Rotaru, V.; Chitoran, E.; Cirimbei, S. Laparoscopic Approach in Abdominal Oncologic Pathology. In Proceedings of the 35th Balkan Medical Week, Athens, Greece, 25–27 September 2018; pp. 260–265. Available online: https://www.webofscience.com/wos/woscc/full-record/WOS:000471903700043 (accessed on 20 July 2023).
  21. Rotaru, V.; Chitoran, E.; Cirimbei, C.; Cirimbei, S.; Simion, L. Preservation of Sensory Nerves During Axillary Lymphadenectomy. In Proceedings of the 35th Balkan Medical Week, Athens, Greece, 25–27 September 2018; pp. 271–276. Available online: https://www.webofscience.com/wos/woscc/full-record/WOS:000471903700045 (accessed on 20 July 2023).
  22. Chitoran, E.; Rotaru, V.; Mitroiu, M.-N.; Durdu, C.-E.; Bohiltea, R.-E.; Ionescu, S.-O.; Gelal, A.; Cirimbei, C.; Alecu, M.; Simion, L. Navigating Fertility Preservation Options in Gynecological Cancers: A Comprehensive Review. Cancers 2024, 16, 2214. [Google Scholar] [CrossRef]
  23. Shapiro, M.A.; Stuhlmiller, T.J.; Wasserman, A.; Kramer, G.A.; Federowicz, B.; Hoos, W.; Kaufman, Z.; Chuyka, D.; Mahoney, W.; Newton, M.E.; et al. AI-Augmented Clinical Decision Support in a Patient-Centric Precision Oncology Registry. AI Precis. Oncol. 2024, 1, 58–68. [Google Scholar] [CrossRef]
  24. Xu, F.; Sepúlveda, M.-J.; Jiang, Z.; Wang, H.; Li, J.; Liu, Z.; Yin, Y.; Roebuck, M.C.; Shortliffe, E.H.; Yan, M.; et al. Effect of an Artificial Intelligence Clinical Decision Support System on Treatment Decisions for Complex Breast Cancer. JCO Clin. Cancer Inform. 2020, 4, CCI.20.00018. [Google Scholar] [CrossRef]
  25. Dvijotham, K.; Winkens, J.; Barsbey, M.; Ghaisas, S.; Stanforth, R.; Pawlowski, N.; Strachan, P.; Ahmed, Z.; Azizi, S.; Bachrach, Y.; et al. Enhancing the reliability and accuracy of AI-enabled diagnosis via complementarity-driven deferral to clinicians. Nat. Med. 2023, 29, 1814–1820. [Google Scholar] [CrossRef]
  26. Abràmoff, M.D.; Lavin, P.T.; Birch, M.; Shah, N.; Folk, J.C. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. npj Digit. Med. 2018, 1, 39. [Google Scholar] [CrossRef] [PubMed]
  27. Rojas-Carabali, W.; Agrawal, R.; Gutierrez-Sinisterra, L.; Baxter, S.L.; Cifuentes-González, C.; Wei, Y.C.; Abisheganaden, J.; Kannapiran, P.; Wong, S.; Lee, B.; et al. Natural Language Processing in medicine and ophthalmology: A review for the 21st-century clinician. Asia-Pac. J. Ophthalmol. 2024, 13, 100084. [Google Scholar] [CrossRef] [PubMed]
  28. Natural Language Processing in Healthcare: 8 Key Use Cases n.d. Available online: https://kms-healthcare.com/blog/natural-language-processing-in-healthcare/ (accessed on 27 July 2025).
  29. Mendonça, E.A.; Haas, J.; Shagina, L.; Larson, E.; Friedman, C. Extracting information on pneumonia in infants using natural language processing of radiology reports. J. Biomed. Inform. 2005, 38, 314–321. [Google Scholar] [CrossRef] [PubMed]
  30. Aramaki, E.; Wakamiya, S.; Yada, S.; Nakamura, Y. Natural Language Processing: From Bedside to Everywhere. Yearb. Med. Inform. 2022, 31, 243. [Google Scholar] [CrossRef]
  31. Locke, S.; Bashall, A.; Al-Adely, S.; Moore, J.; Wilson, A.; Kitchen, G.B. Natural language processing in medicine: A review. Trends Anaesth. Crit. Care 2021, 38, 4–9. [Google Scholar] [CrossRef]
  32. Natural Language Processing in Healthcare Medical Records n.d. Available online: https://www.foreseemed.com/natural-language-processing-in-healthcare (accessed on 27 July 2025).
  33. Palepu, A.; Dhillon, V.; Niravath, P.; Weng, W.H.; Prasad, P.; Saab, K.; Tu, T. Exploring Large Language Models for Specialist-level Oncology Care. arXiv 2024, arXiv:2411.03395. [Google Scholar] [CrossRef]
  34. Saab, K.; Tu, T.; Weng, W.H.; Tanno, R.; Stutz, D.; Wulczyn, E.; Natarajan, V. Capabilities of Gemini Models in Medicine. arXiv 2024, arXiv:2404.18416. [Google Scholar] [CrossRef]
  35. Yang, L.; Xu, S.; Sellergren, A.; Kohlberger, T.; Zhou, Y.; Ktena, I.; Golden, D. Advancing Multimodal Medical Capabilities of Gemini. arXiv 2024, arXiv:2405.03162. [Google Scholar] [CrossRef]
  36. McDuff, D.; Schaekermann, M.; Tu, T.; Palepu, A.; Wang, A.; Garrison, J.; Singhal, K.; Sharma, Y.; Azizi, S.; Kulkarni, K.; et al. Towards Accurate Differential Diagnosis with Large Language Models. Nature 2023, 642, 451–457. [Google Scholar] [CrossRef]
  37. Yu, C.; Liu, J.; Nemati, S.; Yin, G. Reinforcement Learning in Healthcare: A Survey. ACM Comput. Surv. 2022, 55, 1–36. [Google Scholar] [CrossRef]
  38. Eckardt, J.-N.; Wendt, K.; Bornhäuser, M.; Middeke, J.M. Reinforcement Learning for Precision Oncology. Cancers 2021, 13, 4624. [Google Scholar] [CrossRef] [PubMed]
  39. Jayaraman, P.; Desman, J.; Sabounchi, M.; Nadkarni, G.N.; Sakhuja, A. A Primer on Reinforcement Learning in Medicine for Clinicians. npj Digit. Med. 2024, 7, 337. [Google Scholar] [CrossRef] [PubMed]
  40. Almadhor, A.; Ojo, S.; Nathaniel, T.I.; Alsubai, S.; Alharthi, A.; Hejaili AAl et, a.l. An interpretable XAI deep EEG model for schizophrenia diagnosis using feature selection and attention mechanisms. Front. Oncol. 2025, 15, 1630291. [Google Scholar] [CrossRef]
  41. Abas Mohamed, Y.; Ee Khoo, B.; Shahrimie Mohd Asaari, M.; Ezane Aziz, M.; Rahiman Ghazali, F. Decoding the black box: Explainable AI (XAI) for cancer diagnosis, prognosis, and treatment planning-A state-of-the art systematic review. Int. J. Med. Inform. 2025, 193, 105689. [Google Scholar] [CrossRef]
  42. Chang, J.S.; Kim, H.; Baek, E.S.; Choi, J.E.; Lim, J.S.; Kim, J.S.; Shin, S.J. Continuous multimodal data supply chain and expandable clinical decision support for oncology. npj Digit. Med. 2025, 8, 1–13. [Google Scholar] [CrossRef]
  43. Mahadevaiah, G.; Prasad, R.V.; Bermejo, I.; Jaffray, D.; Dekker, A.; Wee, L. Artificial intelligence-based clinical decision support in modern medical physics: Selection, acceptance, commissioning, and quality assurance. Med. Phys. 2020, 47, e228–e235. [Google Scholar] [CrossRef]
  44. Walsh, S.; de Jong, E.E.C.; van Timmeren, J.E.; Ibrahim, A.; Compter, I.; Peerlings, J.; Sanduleanu, S.; Refaee, T.; Keek, S.; Larue, R.T.H.M.; et al. Decision Support Systems in Oncology. JCO Clin. Cancer Inform. 2019, 3, 1–9. [Google Scholar] [CrossRef]
  45. Haddad, T.; Helgeson, J.M.; E Pomerleau, K.; Preininger, A.M.; Roebuck, M.C.; Dankwa-Mullan, I.; Jackson, G.P.; Goetz, M.P. Accuracy of an Artificial Intelligence System for Cancer Clinical Trial Eligibility Screening: Retrospective Pilot Study. JMIR Med. Inform. 2021, 9, e27767. [Google Scholar] [CrossRef]
  46. Audeh, W.; Blumencranz, L.; Kling, H.; Trivedi, H.; Srkalovic, G. Prospective Validation of a Genomic Assay in Breast Cancer: The 70-gene MammaPrint Assay and the MINDACT Trial. Acta Med. Acad. 2019, 48, 18–34. [Google Scholar] [CrossRef]
  47. Filho, O.M.; Cardoso, F.; Poncet, C.; Desmedt, C.; Linn, S.; Wesseling, J.; Hilbers, F.; Delaloge, S.; Pierga, J.-Y.; Brain, E.; et al. Survival outcomes for patients with invasive lobular cancer by MammaPrint: Results from the MINDACT phase III trial. Eur. J. Cancer 2025, 217, 115222. [Google Scholar] [CrossRef]
  48. Cardoso, F.; Piccart-Gebhart, M.; Van’t Veer, L.; Rutgers, E. The MINDACT trial: The first prospective clinical validation of a genomic tool. Mol. Oncol. 2007, 1, 246–251. [Google Scholar] [CrossRef]
  49. Schefler, A.C.; Skalet, A.; Oliver, S.C.N.; Mason, J.; Daniels, A.B.; Alsina, K.M.; Plasseraud, K.M.; A Monzon, F.; Firestone, B. Prospective evaluation of risk-appropriate management of uveal melanoma patients informed by gene expression profiling. Melanoma Manag. 2020, 7, MMT37. [Google Scholar] [CrossRef]
  50. Harbour, J.W.; Chen, R. The DecisionDx-UM Gene Expression Profile Test Provides Risk Stratification and Individualized Patient Care in Uveal Melanoma. PLoS Curr. 2013, 9, 5. [Google Scholar] [CrossRef] [PubMed]
  51. Sun, W.; Hu, G.; Long, G.; Wang, J.; Liu, D.; Hu, G. Predictive value of a serum-based proteomic test in non-small-cell lung cancer patients treated with epidermal growth factor receptor tyrosine kinase inhibitors: A meta-analysis. Curr. Med. Res. Opin. 2014, 30, 2033–2039. [Google Scholar] [CrossRef] [PubMed]
  52. Molina-Pinelo, S.; Pastor, M.D.; Paz-Ares, L. VeriStrat: A prognostic and/or predictive biomarker for advanced lung cancer patients? Expert Rev. Respir. Med. 2014, 8, 1–4. [Google Scholar] [CrossRef] [PubMed]
  53. Ferber, D.; El Nahhas, O.S.; Wölflein, G.; Wiest, I.C.; Clusmann, J.; Leßman, M.E.; Foersch, S.; Lammert, J.; Tschochohei, M.; Jäger, D.; et al. Autonomous Artificial Intelligence Agents for Clinical Decision Making in Oncology. arXiv 2024, arXiv:2404.04667. [Google Scholar] [CrossRef]
  54. Lee, C.S.; Lee, A.Y. Clinical applications of continual learning machine learning. Lancet Digit. Health 2020, 2, e279–e281. [Google Scholar] [CrossRef]
  55. Labkoff, S.; Oladimeji, B.; Kannry, J.; Solomonides, A.; Leftwich, R.; Koski, E.; Joseph, A.L.; Lopez-Gonzalez, M.; A Fleisher, L.; Nolen, K.; et al. Toward a responsible future: Recommendations for AI-enabled clinical decision support. J. Am. Med. Inform. Assoc. 2024, 31, 2730–2739. [Google Scholar] [CrossRef]
Figure 1. Histogram of the age of patients in the studied cohort.
Figure 1. Histogram of the age of patients in the studied cohort.
Diagnostics 15 02391 g001
Figure 2. Comparative performances of trained predictive models.
Figure 2. Comparative performances of trained predictive models.
Diagnostics 15 02391 g002
Figure 3. Confusion matrix of the optimized Random Forest model (80/20).
Figure 3. Confusion matrix of the optimized Random Forest model (80/20).
Diagnostics 15 02391 g003
Figure 4. SHAP summary plot for the optimized Random Forest model. Each dot represents one patient; the x-axis shows the feature’s contribution to the predicted probability of complications (positive values increase risk). Color encodes the feature value (blue = low, red = high). Age and Diagnosis are the most influential features: higher age and specific diagnostic categories push predictions toward higher risk, whereas lower age and other categories push them downward. The modest effect sizes (|SHAP| ≲ 0.05) mirror the cross-validated performance (AUC ≈ 0.62), indicating that risk is distributed across several routine variables rather than dominated by a single predictor.
Figure 4. SHAP summary plot for the optimized Random Forest model. Each dot represents one patient; the x-axis shows the feature’s contribution to the predicted probability of complications (positive values increase risk). Color encodes the feature value (blue = low, red = high). Age and Diagnosis are the most influential features: higher age and specific diagnostic categories push predictions toward higher risk, whereas lower age and other categories push them downward. The modest effect sizes (|SHAP| ≲ 0.05) mirror the cross-validated performance (AUC ≈ 0.62), indicating that risk is distributed across several routine variables rather than dominated by a single predictor.
Diagnostics 15 02391 g004
Figure 5. Distribution of clinical risk scores according to the occurrence of complications—comparison between the initial version (a) versus the logistic derived score (b). The Y-axis shows the raw model score (log-odds); positive values increase the estimated probability and negative values decrease it.
Figure 5. Distribution of clinical risk scores according to the occurrence of complications—comparison between the initial version (a) versus the logistic derived score (b). The Y-axis shows the raw model score (log-odds); positive values increase the estimated probability and negative values decrease it.
Diagnostics 15 02391 g005aDiagnostics 15 02391 g005b
Figure 6. Comparing calibration plot for the logistic derivative score and the original score.
Figure 6. Comparing calibration plot for the logistic derivative score and the original score.
Diagnostics 15 02391 g006
Figure 7. Proposed structured workflow for continuous refinement of an AI-CDSS-based risk score used in oncology clinical practice.
Figure 7. Proposed structured workflow for continuous refinement of an AI-CDSS-based risk score used in oncology clinical practice.
Diagnostics 15 02391 g007
Table 1. Distribution of pretherapeutic socio-demographic and medical variables.
Table 1. Distribution of pretherapeutic socio-demographic and medical variables.
VariableWhole SampleStudy Group (with Complications)Witness Group (Without Complications)p-Value *
Sex 0.259
  Male130 (32.91%)6664
  Female265 (67.09%)152113
Age 0.475
  Extremes 22–8861.6 ± 10.8861 ± 11.062.4 ± 10.7
Residence environment 0.167
  Urban area341 (86.33%)183158
  Rural Area54 (13.67%)3519
Blood type 0.987
  A164 (41.52%)9272
  O139 (35.19%)7663
  B58 (14.68%)3226
  AB34 (8.61)1816
Rhesus factor 0.667
  Rh-positive321 (81.27%)175146
  Rh-negative74 (18.73%)4331
Comorbidities
  Hypertension147 (37.22%)73 (33.5%)74 (41.8%)0.1103
  Ischemic heart disease29 (7.34%)10 (4.6%)19 (10.7%)0.0327
  Cardiac insufficiency33 (8.35%)12 (5.5%)21 (11.9%)0.0367
  Valvopathies62 (15.70%)26 (11.9%)36 (20.3%)0.0318
  Arrythmias32 (8.10%)15 (6.9%)17 (9.6%)0.4230
  Stroke/TIA3 (0.76%)1 (0.5%)2 (1.1%)0.8560
  PTE/DVT20 (5.06%)6 (2.8%)14 (7.9%)0.0362
  Diabetes44 (11.14%)18 (8.3%)26 (14.7%)0.0629
  Chronic lung disease20 (5.06%)9 (4.1%)11 (6.2%)0.4779
  Chronic renal disease10 (2.53%)1 (0.5%)9 (5.1%)0.0096
  Chronic hepatitis or cirrhosis13 (3.29%)6 (2.8%)7 (4.0%)0.7020
  Previous cancer34 (8.61%)21 (9.6%)13 (7.3%)0.5313
  Psychiatric disorders15 (3.80%)6 (2.8%)9 (5.1%)0.3465
  Recurrent urinary infections15 (3.80%)11 (5.0%)4 (2.3%)0.2396
  Previous surgery288 (72.91%)158 (72.5%)130 (73.4%)0.9190
Chronic/previous treatments
  Oral Anticoagulants29 (7.34%)9 (4.1%)20 (11.3%)0.0116
  Oral Antidiabetics20 (5.06%)7 (3.2%)13 (7.3%)0.1025
  Insulin4 (1.01%)1 (0.5%)3 (1.7%)0.4746
  Bevacizumab55 (13.92%)28 (12.8%)27 (15.3%)0.5878
  Neoadjuvant Immunotherapy 94 (23.80%)44 (20.2%)50 (28.2%)0.0796
  Neoadjuvant Chemotherapy 234 (59.24%)134 (61.5%)100 (56.5%)0.3698
  Neoadjuvant Radiotherapy 92 (23.29%)47 (21.6%)45 (25.4%)0.4331
  Hormonotherapy5 (1.27%)2 (0.9%)3 (1.7%)0.8143
Nutritional status 0.2531
  Normal weight257 (65.06%)148 (67.9%)109 (61.6%)
  Overweight/Obesity73 (18.48%)40 (18.3%)33 (18.6%)
  Underweight/Cachexia65 (16.46%)30 (13.8%)35 (19.8%)
Abbreviations used: TIA—transient ischemic attack, PTE—pulmonary thromboembolism, DVT—deep venous thrombosis, *—p-value is calculated using the Chi-square test, t-test or Mann–Whitney U (Wilcoxon rank-sum) test depending on the type of variable.
Table 2. Distribution of pretherapeutic variables specific to the neoplastic disease.
Table 2. Distribution of pretherapeutic variables specific to the neoplastic disease.
VariableWhole SampleStudy Group (with Complications)Witness Group (Without Complications)p-Value *
Type of cancer 0.0002
  Colo-rectal209 (52.91%)104 (47.7%)105 (59.3%)
  Ovarian/tubal/primary peritoneal102 (25.82%)70 (32.1%)32 (18.1%)
  Cervical40 (10.13%)16 (7.3%)24 (13.6%)
  Breast28 (7.09%)22 (10.1%)6 (3.4%)
  Pulmonary7 (1.77%)4 (1.8%)3 (1.7%)
  Others9 (2.28%)2 (0.9%)7 (4.0%)
Histopathology 0.0013
  Adenocarcinoma206 (52.15%)104 (47.7%)102 (57.6%)
  Serous carcinoma93 (23.54%)63 (28.9%)30 (16.9%)
  Squamous carcinoma33 (8.35%)13 (6.0%)20 (11.3%)
  Invasive ductal/lobular carcinoma25 (6.33%)20 (9.2%)5 (2.8%)
  Mucinous adenocarcinoma23 (5.82%)9 (4.1%)14 (7.9%)
  Others15 (3.80%)9 (4.1%)6 (3.4%)
Differentiation degree 0.0065
  Poorly differentiated (G3)144 (36.46%)92 (42.2%)53 (29.9%)
  Moderate differentiated (G2)227 (57.47%)110 (50.5%)117 (66.1%)
  Well differentiated (G1)24 (6.07%)16 (7.3%)8 (4.52%)
Perineural invasion75 (18.99%)40 (18.3%)35 (19.8%)0.7049
Lymphovascular invasion117 (29.62%)64 (29.4%)53 (29.9%)0.9877
Peritumoral lymphocyte infiltrate79 (20.00%)36 (16.5%)43 (24.3%)0.1473
Stage 0.3306
  IV316 (80.00%)168 (77.1%)148 (83.6%)
  III74 (18.73%)46 (21.1%)28 (15.8%)
  II4 (1.01%)3 (1.4%)1 (0.6%)
  I1 (0.25)1 (0.5%)0 (0.0%)
Metastasis
  Peritoneale142 (35.95%)81 (37.2%)61 (34.5%)0.6533
  Bone25 (6.33%)14 (6.4%)11 (6.2%)1.0000
  Hepatic169 (42.78%)89 (40.8%)82 (46.3%)0.238
  Pulmonary95 (24.05%)53 (24.3%)42 (23.7%)0.9869
  Lymphatic46 (11.65%)28 (12.8%)18 (10.2%)0.5052
  Cerebral8 (2.03%)4 (1.8%)4 (2.3%)1.0000
  Others15 (3.80%)8 (3.7%)7 (4.0%)1.0000
Type of disease 0.1061
  Metastatic314 (79.49%)170 (78.0%)144 (81.4%)
  Locally advanced/Unresectable52 (13.16%)35 (16.1%)17 (9.6%)
  Recurrent29 (7.34%)13 (6.0%)16 (9.0%)
Bevacizumab/biosimilar administration 0.3935
  First line180 (45.57%)95 (43.6%)85 (48.0%)
  Second line146 (36.96)84 (38.5%)62 (35.0%)
  Subsequent therapeutic lines54 (13.67%)28 (12.8%)26 (14.7%)
  Maintenance15 (3.80%)11 (5.0%)4 (2.3%)
* p-value is calculated using the Chi-square test for categorical variables.
Table 3. Distribution of pretherapeutic paraclinical variables.
Table 3. Distribution of pretherapeutic paraclinical variables.
VariableStudy Group (with Complications)Witness Group (Without Complications)p-Value *
Hemoglobin11.84 ± 1.7412.38 ± 1.560.0009
Leucocytes7.44 ± 2.707.07 ± 2.660.1378
Thrombocytes286.110 ± 108.420272.790 ± 111.2400.2500
Glycemia101.83 ± 26.25103.79 ± 27.010.3049
Creatinine0.77 ± 0.210.82 ± 0.250.0845
Urea32.16 ± 13.2934.55 ± 15.480.1354
TGO23.67 ± 13.3026.38 ± 23.240.8259
TGP22.29 ± 13.2926.04 ± 38.480.5718
GGT67.38 ± 94.0775.01 ± 106.160.0773
Ca2+9.35 ± 0.599.30 ± 0.570.1655
Coagulation modifications29 (13.3%)39 (22.0%)0.0314
Tumoral markers alterations74 (33.9%)72 (40.7%)0.2027
* p-value is calculated using the Chi-square test for binary categorical variables and the Mann–Whitney U (Wilcoxon rank-sum) test for numerical variables.
Table 4. Number, type and severity of Bevacizumab/biosimilars related complications in study group.
Table 4. Number, type and severity of Bevacizumab/biosimilars related complications in study group.
Type of ComplicationsLow-Medium Severity (Total Number of Cases = 245)High Severity (Total Number of Cases = 60 *)Total
Septic/Infectious complications351449
  Sepsis 257
  Abscess7310
  Urinary infection11213
  Wound complications404
  Pneumonia123
  Others10212
Cardiovascular complications51758
  Thromboembolic events12113
  Hemorrhagic events (other than digestive)8210
  Arrythmia101
  Arterial hypertension30434
Gastro-intestinal complications341448
  Digestive hemorrhage426
  Ileus202
  Fistulas/perforative complications3912
  Abdominal pain617
  Gastritis101
  Stomatitis303
  Nausea/vomiting10010
  Diarrhea527
Genito-urinary complications291342
  Proteinuria29433
  Renal insufficiency077
  Genito-urinary fistula022
Hematologic complications72981
  Anemia34034
  Leucopenia18624
  Thrombocytopenia18220
  Neutropenia200
  Pancytopenia011
Others24327
  Headache/Migraine516
  Neuropathies 606
  Rhinitis101
  Dyspnea325
  Dehydration303
  Anorexia101
  Cutaneous complications505
* the total number of patients with severe therapy-related complications was 56—the difference between 56 and the total number of severe complications is explained by the fact that some patients developed more than one severe complication, while others, although having adverse reactions, none of those were classified as severe by their oncologist.
Table 5. Performance metrics of the logistic derivative score versus the original score.
Table 5. Performance metrics of the logistic derivative score versus the original score.
Performance MetricsInitial Clinical ScoreLogistic Derived Score
AUC ROC0.5550.720
Accuracy0.5900.680
Sensitivity (Recall)0.4720.642
Specificity0.6080.698
F1-Score0.5020.662
Table 6. Logistic-derived score variables and points awarded for each category.
Table 6. Logistic-derived score variables and points awarded for each category.
VariableCategory and Points Awarded for Each Category
Age<65 years—0 points
≥65 years—1 point
Serum Urea<40 mg/dL—0 points
≥40 mg/dL—1 point
Leucocytes<10.000/mmc—0 points
≥10.000/mmc—1 point
Hemoglobin≥10 g/dL—0 points
<10 g/dL—1 point
Transaminases (TGO/TGP)<40 U/L—0 points
≥40 U/L—1 point
Creatinine<1.5 mg/dL—0 points
≥1.5 mg/dL—1 point
StageStage I–II—0 points
Stage III–IV—1 point
Differentiation degreeG1–G2—0 points
G3—1 point
Lymphovascular invasionAbsent—0 points
Present—1 point
Type of cancerBreast, Ovarian, Cervical—0 points
Colo-rectal, Pulmonary, Others—1 point
Table 7. Clinical risk thresholds of the logistic-derived score.
Table 7. Clinical risk thresholds of the logistic-derived score.
PointsRisk GroupProbability of Complications
0–3 pointsLow risk<25%
4–6 pointsIntermediate risk25–60%
7–10 pointsHigh risk>60%
Table 8. AI-CDSS-based tools developed for oncology—related work and state-of-the-art (SOTA) approaches.
Table 8. AI-CDSS-based tools developed for oncology—related work and state-of-the-art (SOTA) approaches.
CDSS/TestOncology DomainAI TypeKey Results
Yonsei Cancer Data Library [42]Multiple cancer typesMultimodal Deep Learning (Electronic Health Record + Genomic data)Personalized care for >170K patients; streamlined workflow
AI for
Trial Eligibility [45]
Breast cancerNLP + Rule-based
Screening
Higher sensitivity and specificity vs manual screening
MammaPrint
[46,47,48]
Breast cancerGenomic Risk Classifier (ML-based)Avoids chemotherapy in low-risk patients (MINDACT validation)
DecisionDx-UM [49,50]Uveal melanomaGene Expression Classifier (ML-based)Accurate metastasis risk classification; accepted standard of care
VeriStrat
[51,52]
NSCLCProteomic Classifier (ML-based)Predicts response to EGFR inhibitors like erlotinib
LLM Agents
[33,53]
Oncology simulationsLLM + Retrieval-Augmented Generation>91% accuracy in simulated oncologic decisions with explainability
Abbreviation: CDSS—clinical decision support system; AI—artificial intelligence; NLP—Natural Language Processing; ML—Machine Learning; LLM—Large Language Model; NSCLC—Non-Small Cell Lung Cancer; EGFR—Epidermal Growth Factor Receptor; MINDACT—Microarray In Node-negative and 1 to 3 positive lymph node Disease may Avoid ChemoTherapy (clinical trial).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chitoran, E.; Rotaru, V.; Gelal, A.; Ionescu, S.-O.; Gullo, G.; Stefan, D.-C.; Simion, L. Using Artificial Intelligence to Develop Clinical Decision Support Systems—The Evolving Road of Personalized Oncologic Therapy. Diagnostics 2025, 15, 2391. https://doi.org/10.3390/diagnostics15182391

AMA Style

Chitoran E, Rotaru V, Gelal A, Ionescu S-O, Gullo G, Stefan D-C, Simion L. Using Artificial Intelligence to Develop Clinical Decision Support Systems—The Evolving Road of Personalized Oncologic Therapy. Diagnostics. 2025; 15(18):2391. https://doi.org/10.3390/diagnostics15182391

Chicago/Turabian Style

Chitoran, Elena, Vlad Rotaru, Aisa Gelal, Sinziana-Octavia Ionescu, Giuseppe Gullo, Daniela-Cristina Stefan, and Laurentiu Simion. 2025. "Using Artificial Intelligence to Develop Clinical Decision Support Systems—The Evolving Road of Personalized Oncologic Therapy" Diagnostics 15, no. 18: 2391. https://doi.org/10.3390/diagnostics15182391

APA Style

Chitoran, E., Rotaru, V., Gelal, A., Ionescu, S.-O., Gullo, G., Stefan, D.-C., & Simion, L. (2025). Using Artificial Intelligence to Develop Clinical Decision Support Systems—The Evolving Road of Personalized Oncologic Therapy. Diagnostics, 15(18), 2391. https://doi.org/10.3390/diagnostics15182391

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop