Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,307)

Search Parameters:
Keywords = object based random forest

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 8025 KB  
Article
Quantitative Analysis of Smooth Pursuit and Saccadic Eye Movements in Multiple Sclerosis
by Pavol Skacik, Lucia Kotulova, Ema Kantorova, Egon Kurca and Stefan Sivak
Neurol. Int. 2026, 18(2), 22; https://doi.org/10.3390/neurolint18020022 - 26 Jan 2026
Abstract
Introduction: Multiple sclerosis (MS) is a chronic inflammatory and neurodegenerative disease of the central nervous system, frequently associated with visual and oculomotor disturbances. Quantitative analysis of eye movements represents a non-invasive method for assessing central nervous system dysfunction beyond conventional imaging; however, [...] Read more.
Introduction: Multiple sclerosis (MS) is a chronic inflammatory and neurodegenerative disease of the central nervous system, frequently associated with visual and oculomotor disturbances. Quantitative analysis of eye movements represents a non-invasive method for assessing central nervous system dysfunction beyond conventional imaging; however, the diagnostic and predictive value of oculomotor metrics remains insufficiently defined. Objectives: The aims of this study were to compare smooth pursuit gain and reflexive saccade parameters (latency, velocity, and precision) between individuals with MS and healthy controls, and to evaluate their ability to discriminate disease status. Methods: This cross-sectional study included 46 clinically stable patients with MS (EDSS ≤ 6.5) and 46 age- and sex-matched healthy controls. Oculomotor function was assessed using videonystagmography under standardized conditions. Group differences across horizontal and vertical gaze directions were analyzed using linear mixed-effects models. Random forest models were applied to assess the discriminative performance of oculomotor parameters, with permutation-based feature importance and receiver operating characteristic (ROC) curve analysis. Results: Patients with MS showed significantly reduced smooth pursuit gain across most horizontal and vertical directions compared with controls. Saccadic latency was significantly prolonged in all tested movement directions. Saccadic velocity exhibited selective directional impairment consistent with subtle medial longitudinal fasciculus involvement, whereas saccadic precision did not differ significantly between groups. A random forest model combining pursuit and saccadic parameters demonstrated only moderate discriminative performance between MS patients and controls (AUC = 0.694), with saccadic latency contributing most strongly to classification. Conclusions: Quantitative eye-movement assessment revealed widespread oculomotor abnormalities in MS, particularly reduced smooth pursuit gain and prolonged saccadic latency. Although the overall discriminative accuracy of oculomotor parameters was limited, these findings support their potential role as complementary markers of central nervous system dysfunction. Further longitudinal and multimodal studies are required to clarify their clinical relevance and prognostic value. Full article
(This article belongs to the Special Issue Advances in Multiple Sclerosis, Third Edition)
Show Figures

Graphical abstract

23 pages, 3441 KB  
Article
Integrating Large Language Models with Deep Learning for Breast Cancer Treatment Decision Support
by Heeseung Park, Serin Ok, Taewoo Kang and Meeyoung Park
Diagnostics 2026, 16(3), 394; https://doi.org/10.3390/diagnostics16030394 - 26 Jan 2026
Abstract
Background/Objectives: Breast cancer is one of the most common malignancies, but its heterogeneous molecular subtypes make treatment decision-making complex and patient-specific. Both the pathology reports and the electronic medical record (EMR) play a critical role for an appropriate treatment decision. This study [...] Read more.
Background/Objectives: Breast cancer is one of the most common malignancies, but its heterogeneous molecular subtypes make treatment decision-making complex and patient-specific. Both the pathology reports and the electronic medical record (EMR) play a critical role for an appropriate treatment decision. This study aimed to develop an integrated clinical decision support system (CDSS) that combines a large language model (LLM)-based pathology analysis with deep learning-based treatment prediction to support standardized and reliable decision-making. Methods: Real-world data (RWD) obtained from a cohort of 5015 patients diagnosed with breast cancer were analyzed. Meta-Llama-3-8B-Instruct automatically extracted the TNM stage and tumor size from the pathology reports, which were then integrated with EMR variables. A multi-label classification of 16 treatment combinations was performed using six models, including Decision Tree, Random Forest, GBM, XGBoost, DNN, and Transformer. Performance was evaluated using accuracy, macro/micro-averaged precision, recall, F1 score, and AUC. Results: Using combined LLM-extracted pathology and EMR features, GBM and XGBoost achieved the highest and most stable predictive performance across all feature subset configurations (macro-F1 ≈ 0.88–0.89; AUC = 0.867–0.868). Both models demonstrated strong discrimination ability and consistent recall and precision, highlighting their robustness for multi-label classification in real-world settings. Decision Tree and Random Forest showed moderate but reliable performance (macro-F1 = 0.84–0.86; AUC = 0.849–0.821), indicating their applicability despite lower predictive capability. By contrast, the DNN and Transformer models produced comparatively lower scores (macro-F1 = 0.74–0.82; AUC = 0.780–0.757), especially when using the full feature set, suggesting limited suitability for structured clinical data without strong contextual dependencies. These findings indicate that gradient-boosting ensemble approaches are better optimized for tabular medical data and generate more clinically reliable treatment recommendations. Conclusions: The proposed artificial intelligence-based CDSS improves accuracy and consistency in breast cancer treatment decision support by integrating automated pathology interpretation with deep learning, demonstrating its potential utility in real-world cancer care. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

22 pages, 4317 KB  
Article
Non-Contact Temperature Monitoring in Dairy Cattle via Thermal Infrared Imaging and Environmental Parameters
by Kaixuan Zhao, Shaojuan Ge, Yinan Chen, Qianwen Li, Mengyun Guo, Yue Nian and Wenkai Ren
Agriculture 2026, 16(3), 306; https://doi.org/10.3390/agriculture16030306 - 26 Jan 2026
Abstract
Core body temperature is a critical physiological indicator for assessing and diagnosing animal health status. In bovines, continuously monitoring this metric enables accurate evaluation of their physiological condition; however, traditional rectal measurements are labor-intensive and cause stress in animals. To achieve intelligent, contactless [...] Read more.
Core body temperature is a critical physiological indicator for assessing and diagnosing animal health status. In bovines, continuously monitoring this metric enables accurate evaluation of their physiological condition; however, traditional rectal measurements are labor-intensive and cause stress in animals. To achieve intelligent, contactless temperature monitoring in cattle, we proposed a non-invasive method based on thermal imaging combined with environmental data fusion. First, thermal infrared images of the cows’ faces were collected, and the You Only Look Once (YOLO) object detection model was used to locate the head region. Then, the YOLO segmentation network was enhanced with the Online Convolutional Re-parameterization (OREPA) and High-level Screening-feature Fusion Pyramid Network (HS-FPN) modules to perform instance segmentation of the eye socket area. Finally, environmental variables—ambient temperature, humidity, wind speed, and light intensity—were integrated to compensate for eye socket temperature, and a random forest algorithm was used to construct a predictive model of rectal temperature. The experiments were conducted using a thermal infrared image dataset comprising 33,450 frontal-view images of dairy cows with a resolution of 384 × 288 pixels, along with 1471 paired samples combining thermal and environmental data for model development. The proposed method achieved a segmentation accuracy (mean average precision, mAP50–95) of 86.59% for the eye socket region, ensuring reliable temperature extraction. The rectal temperature prediction model demonstrated a strong correlation with the reference rectal temperature (R2 = 0.852), confirming its robustness and predictive reliability for practical applications. These results demonstrate that the proposed method is practical for non-contact temperature monitoring of cattle in large-scale farms, particularly those operating under confined or semi-confined housing conditions. Full article
(This article belongs to the Section Farm Animal Production)
Show Figures

Figure 1

14 pages, 1371 KB  
Article
AI-Based Electromyographic Analysis of Single-Leg Landing for Injury Risk Prediction in Taekwondo Athletes
by Jun-Sik Kim, Fatima Faridoon, Jaeyeop Choi, Junghwan Oh, Juhyun Kang and Hae Gyun Lim
Healthcare 2026, 14(3), 292; https://doi.org/10.3390/healthcare14030292 - 23 Jan 2026
Viewed by 90
Abstract
Background/Objectives: Improper landing mechanics in Taekwondo can lead to non-contact injuries such as ankle sprains and knee ligament tears, highlighting the necessity for objective methods to evaluate landing stability and injury risk. Electromyography (EMG) enables the examination of muscle activation patterns; however, conventional [...] Read more.
Background/Objectives: Improper landing mechanics in Taekwondo can lead to non-contact injuries such as ankle sprains and knee ligament tears, highlighting the necessity for objective methods to evaluate landing stability and injury risk. Electromyography (EMG) enables the examination of muscle activation patterns; however, conventional analyses based on simple averages have limited predictive value. Methods: This study analyzed EMG signals recorded during single-leg landings (45 cm height) in 30 elite male Taekwondo athletes. Participants were divided into regular exercise groups (REG, n = 15) and non-exercise groups (NEG, n = 15). Signals were segmented into two phases. Eight features were extracted per muscle per phase. Classification models (Random Forest, XGBoost, Logistic Regression, Voting Classifier) were used to classify between groups, while regression models (Ridge, Random Forest, XGBoost) predicted continuous muscle activation changes as injury risk indicators. Results: The Random Forest Classifier achieved an accuracy of 0.8365 and an F1-score of 0.8547. For regression, Ridge Regression indicated high performance (R2 = 0.9974, MAE = 0.2620, RMSE = 0.4284, 5-fold CV MAE: 0.2459 ± 0.0270), demonstrating strong linear correlations between EMG features and outcomes. Conclusions: The AI-enabled EMG analysis can be used as an objective measure of the study of the individual landing stability and risk of injury in Taekwondo athletes, but its clinical application has to be validated in the future by biomechanical injury indicators and prospective cohort studies. Full article
(This article belongs to the Section Artificial Intelligence in Healthcare)
20 pages, 1962 KB  
Article
Machine Learning-Based Prediction and Feature Attribution Analysis of Contrast-Associated Acute Kidney Injury in Patients with Acute Myocardial Infarction
by Neriman Sıla Koç, Can Ozan Ulusoy, Berrak Itır Aylı, Yusuf Bozkurt Şahin, Veysel Ozan Tanık, Arzu Akgül and Ekrem Kara
Medicina 2026, 62(1), 228; https://doi.org/10.3390/medicina62010228 - 22 Jan 2026
Viewed by 28
Abstract
Background and Objectives: Contrast-associated acute kidney injury (CA-AKI) is a frequent and clinically significant complication in patients with acute myocardial infarction (AMI) undergoing coronary angiography. Early and accurate risk stratification remains challenging with conventional models that rely on linear assumptions and limited [...] Read more.
Background and Objectives: Contrast-associated acute kidney injury (CA-AKI) is a frequent and clinically significant complication in patients with acute myocardial infarction (AMI) undergoing coronary angiography. Early and accurate risk stratification remains challenging with conventional models that rely on linear assumptions and limited variable integration. This study aimed to evaluate and compare the predictive performance of multiple machine learning (ML) algorithms with traditional logistic regression and the Mehran risk score for CA-AKI prediction and to explore key determinants of risk using explainable artificial intelligence methods. Materials and Methods: This retrospective, single-center study included 1741 patients with AMI who underwent coronary angiography. CA-AKI was defined according to KDIGO criteria. Multiple ML models, including gradient boosting machine (GBM), random forest (RF), XGBoost, support vector machine, elastic net, and standard logistic regression were developed using routinely available clinical and laboratory variables. A weighted ensemble model combining the best-performing algorithms was constructed. Model discrimination was assessed using area under the receiver operating characteristic curve (AUC), along with sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV). Model interpretability was evaluated using feature importance and SHapley Additive exPlanations (SHAP). Results: CA-AKI occurred in 356 patients (20.4%). In multivariable logistic regression, lower left ventricular ejection fraction, higher contrast volume, lower sodium, lower hemoglobin, and higher neutrophil-to-lymphocyte ratio (NLR) were independently associated with CA-AKI. Among ML approaches, the weighted ensemble model demonstrated the highest discriminative performance (AUC 0.721), outperforming logistic regression and the Mehran risk score (AUC 0.608). Importantly, the ensemble model achieved a consistently high NPV (0.942), enabling reliable identification of low-risk patients. Explainability analyses revealed that inflammatory markers, particularly NLR, along with sodium, uric acid, baseline renal indices, and contrast burden, were the most influential predictors across models. Conclusions: In patients with AMI undergoing coronary angiography, interpretable ML models, especially ensemble and gradient boosting-based approaches, provide superior risk stratification for CA-AKI compared with conventional methods. The high negative predictive value highlights their clinical utility in safely identifying low-risk patients and supporting individualized, risk-adapted preventive strategies. Full article
(This article belongs to the Section Urology & Nephrology)
Show Figures

Figure 1

32 pages, 2929 KB  
Article
Policy Plateau and Structural Regime Shift: Hybrid Forecasting of the EU Decarbonisation Gap Toward 2030 Targets
by Oksana Liashenko, Kostiantyn Pavlov, Olena Pavlova, Olga Demianiuk, Robert Chmura, Bożena Sowa and Tetiana Vlasenko
Sustainability 2026, 18(2), 1114; https://doi.org/10.3390/su18021114 - 21 Jan 2026
Viewed by 92
Abstract
This study investigates the structural evolution and projected trajectory of greenhouse gas (GHG) emissions across the EU27 from 1990 to 2030, with a particular focus on their implications for the effectiveness of European climate policy. Drawing on official sectoral data and employing a [...] Read more.
This study investigates the structural evolution and projected trajectory of greenhouse gas (GHG) emissions across the EU27 from 1990 to 2030, with a particular focus on their implications for the effectiveness of European climate policy. Drawing on official sectoral data and employing a multi-method framework combining time series modelling (ARIMA), machine learning (Random Forest), regime-switching analysis, and segmented linear regression, we assess past dynamics, detect structural shifts, and forecast future trends. Empirical findings, based on Markov-switching models and segmented regression analysis, indicate a statistically significant regime change around 2014, marking a transition to a new emissions pattern characterised by a deceleration in reduction rates. While the energy sector experienced the most significant decline, agriculture and industry have gained relative prominence, underscoring their growing strategic importance as targets for policy interventions. Hybrid ARIMA–ML forecasts indicate that, under current trajectories, the EU is unlikely to meet its 2030 Fit for 55 targets without adaptive and sector-specific interventions, with a projected shortfall of 12–15 percentage points relative to 1990 levels, excluding LULUCF. The results underscore critical weaknesses in the EU’s climate policy architecture and reveal a clear need for transformative recalibration. Without accelerated action and strengthened governance mechanisms, the post-2014 regime risks entrenching a plateau in emissions reductions, jeopardising long-term climate objectives. Full article
(This article belongs to the Section Air, Climate Change and Sustainability)
Show Figures

Figure 1

21 pages, 2691 KB  
Article
Interturn Short-Circuit Fault Diagnosis in a Permanent Magnet Synchronous Generator Using Wavelets and Binary Classifiers
by Jose Antonio Alvarez-Salas, Francisco Javier Villalobos-Pina, Mario Arturo Gonzalez-Garcia and Ricardo Alvarez-Salas
Processes 2026, 14(2), 377; https://doi.org/10.3390/pr14020377 - 21 Jan 2026
Viewed by 63
Abstract
Condition monitoring and diagnosis in a permanent magnet synchronous generator (PMSG) are crucial for ensuring its service continuity and reliability. Recent advancements have introduced innovative, non-invasive techniques for detecting mechanical and electrical faults in this machine. This paper proposes a novel application of [...] Read more.
Condition monitoring and diagnosis in a permanent magnet synchronous generator (PMSG) are crucial for ensuring its service continuity and reliability. Recent advancements have introduced innovative, non-invasive techniques for detecting mechanical and electrical faults in this machine. This paper proposes a novel application of the discrete wavelet transform and binary classifiers for diagnosing interturn short-circuit faults in a PMSG with high accuracy and low computational burden. The objective of fault diagnosis is to detect the presence of an interturn short-circuit fault (fault vs. no-fault) under different fault severities and operating speeds. Multiple binary models were trained separately for each fault scenario. The three-phase currents from the PMSG are processed using the discrete wavelet transform to extract features, which are then fed into a binary classifier based on a Random Forest algorithm. Optimization techniques are used to improve the performance of the binary classifiers. Experimental results obtained under various stator fault conditions in the PMSG are presented. Metrics such as accuracy and confusion matrices are used to evaluate the performance of binary classifiers. Full article
(This article belongs to the Special Issue Fault Diagnosis of Equipment in the Process Industry)
Show Figures

Figure 1

17 pages, 783 KB  
Article
Hospital-Wide Sepsis Detection: A Machine Learning Model Based on Prospectively Expert-Validated Cohort
by Marcio Borges-Sa, Andres Giglio, Maria Aranda, Antonia Socias, Alberto del Castillo, Cristina Pruenza, Gonzalo Hernández, Sofía Cerdá, Lorenzo Socias, Victor Estrada, Roberto de la Rica, Elisa Martin and Ignacio Martin-Loeches
J. Clin. Med. 2026, 15(2), 855; https://doi.org/10.3390/jcm15020855 - 21 Jan 2026
Viewed by 82
Abstract
Background/Objectives: Sepsis detection remains challenging due to clinical heterogeneity and limitations of traditional scoring systems. This study developed and validated a hospital-wide machine learning model for sepsis detection using retrospectively developed data from prospectively expert-validated cases, aiming to improve diagnostic accuracy beyond conventional [...] Read more.
Background/Objectives: Sepsis detection remains challenging due to clinical heterogeneity and limitations of traditional scoring systems. This study developed and validated a hospital-wide machine learning model for sepsis detection using retrospectively developed data from prospectively expert-validated cases, aiming to improve diagnostic accuracy beyond conventional approaches. Methods: This retrospective cohort study analysed 218,715 hospital episodes (2014–2018) at a tertiary care centre. Sepsis cases (n = 11,864, 5.42%) were prospectively validated in real-time by a Multidisciplinary Sepsis Unit using modified Sepsis-2 criteria with organ dysfunction. The model integrated structured data (26.95%) and unstructured clinical notes (73.04%) extracted via natural language processing from 2829 variables, selecting 230 relevant predictors. Thirty models including random forests, support vector machines, neural networks, and gradient boosting were developed and evaluated. The dataset was randomly split (5/7 training, 2/7 testing) with preserved patient-level independence. Results: The BiAlert Sepsis model (random forest + Sepsis-2 ensemble) achieved an AUC-ROC of 0.95, sensitivity of 0.93, and specificity of 0.84, significantly outperforming traditional approaches. Compared to the best rule-based method (Sepsis-2 + qSOFA, AUC-ROC 0.90), BiAlert reduced false positives by 39.6% (13.10% vs. 21.70%, p < 0.01). Novel predictors included eosinopenia and hypoalbuminemia, while traditional variables (MAP, GCS, platelets) showed minimal univariate association. The model received European Medicines Agency approval as a medical device in June 2024. Conclusions: This hospital-wide machine learning model, trained on prospectively expert-validated cases and integrating extensive NLP-derived features, demonstrates superior sepsis detection performance compared to conventional scoring systems. External validation and prospective clinical impact studies are needed before widespread implementation. Full article
Show Figures

Figure 1

13 pages, 6367 KB  
Article
Gene Expression-Based Colorectal Cancer Prediction Using Machine Learning and SHAP Analysis
by Yulai Yin, Zhen Yang, Xueqing Li, Shuo Gong and Chen Xu
Genes 2026, 17(1), 114; https://doi.org/10.3390/genes17010114 - 20 Jan 2026
Viewed by 216
Abstract
Objective: To develop and validate a genetic diagnostic model for colorectal cancer (CRC). Methods: First, differential expression genes (DEGs) between colorectal cancer and normal groups were screened using the TCGA database. Subsequently, a two-sample Mendelian randomization analysis was performed using the eQTL genomic [...] Read more.
Objective: To develop and validate a genetic diagnostic model for colorectal cancer (CRC). Methods: First, differential expression genes (DEGs) between colorectal cancer and normal groups were screened using the TCGA database. Subsequently, a two-sample Mendelian randomization analysis was performed using the eQTL genomic data from the IEU OpenGWAS database and colorectal cancer outcomes from the R12 Finnish database to identify associated genes. The intersecting genes from both methods were selected for the development and validation of the CRC genetic diagnostic model using nine machine learning algorithms: Lasso Regression, XGBoost, Gradient Boosting Machine (GBM), Generalized Linear Model (GLM), Neural Network (NN), Support Vector Machine (SVM), k-Nearest Neighbors (KNN), Random Forest (RF), and Decision Tree (DT). Results: A total of 3716 DEGs were identified from the TCGA database, while 121 genes were associated with CRC based on the eQTL Mendelian randomization analysis. The intersection of these two methods yielded 27 genes. Among the nine machine learning methods, XGBoost achieved the highest AUC value of 0.990. The top five genes predicted by the XGBoost method—RIF1, GDPD5, DBNDD1, RCCD1, and CLDN5—along with the five most significantly differentially expressed genes (ASCL2, IFITM3, IFITM1, SMPDL3A, and SUCLG2) in the GSE87211 dataset, were selected for the construction of the final colorectal cancer (CRC) genetic diagnostic model. The ROC curve analysis revealed an AUC (95% CI) of 0.9875 (0.9737–0.9875) for the training set, and 0.9601 (0.9145–0.9601) for the validation set, indicating strong predictive performance of the model. SHAP model interpretation further identified IFITM1 and DBNDD1 as the most influential genes in the XGBoost model, with both making positive contributions to the model’s predictions. Conclusions: The gene expression profile in colorectal cancer is characterized by enhanced cell proliferation, elevated metabolic activity, and immune evasion. A genetic diagnostic model constructed based on ten genes (RIF1, GDPD5, DBNDD1, RCCD1, CLDN5, ASCL2, IFITM3, IFITM1, SMPDL3A, and SUCLG2) demonstrates strong predictive performance. This model holds significant potential for the early diagnosis and intervention of colorectal cancer, contributing to the implementation of third-tier prevention strategies. Full article
(This article belongs to the Section Bioinformatics)
Show Figures

Figure 1

14 pages, 1176 KB  
Systematic Review
The Efficacy of Electronic Health Record-Based Artificial Intelligence Models for Early Detection of Pancreatic Cancer: A Systematic Review and Meta-Analysis
by George G. Makiev, Igor V. Samoylenko, Valeria V. Nazarova, Zahra R. Magomedova, Alexey A. Tryakin and Tigran G. Gevorkyan
Cancers 2026, 18(2), 315; https://doi.org/10.3390/cancers18020315 - 20 Jan 2026
Viewed by 126
Abstract
Background: The persistently low 5-year survival rate for pancreatic cancer (PC) underscores the critical need for early detection. However, population-wide screening remains impractical. Artificial Intelligence (AI) models using electronic health record (EHR) data offer a promising avenue for pre-symptomatic risk stratification. Objective: To [...] Read more.
Background: The persistently low 5-year survival rate for pancreatic cancer (PC) underscores the critical need for early detection. However, population-wide screening remains impractical. Artificial Intelligence (AI) models using electronic health record (EHR) data offer a promising avenue for pre-symptomatic risk stratification. Objective: To systematically review and meta-analyze the performance of AI models for PC prediction based exclusively on structured EHR data. Methods: We systematically searched PubMed, MedRxiv, BioRxiv, and Google Scholar (2010–2025). Inclusion criteria encompassed studies using EHR-derived data (excluding imaging/genomics), applying AI for PC prediction, reporting AUC, and including a non-cancer cohort. Two reviewers independently extracted data. Random-effects meta-analysis was performed for AUC, sensitivity (Se), and specificity (Sp) using R software version 4.5.1. Heterogeneity was assessed using I2 statistics and publication bias was evaluated. Results: Of 946 screened records, 19 studies met the inclusion criteria. The pooled AUC across all models was 0.785 (95% CI: 0.759–0.810), indicating good overall discriminatory ability. Neural Network (NN) models demonstrated a statistically significantly higher pooled AUC (0.826) compared to Logistic Regression (LogReg, 0.799), Random Forests (RF, 0.762), and XGBoost (XGB, 0.779) (all p < 0.001). In analyses with sufficient data, models like Light Gradient Boosting (LGB) showed superior Se and Sp (99% and 98.7%, respectively) compared to NNs and LogReg, though based on limited studies. Meta-analysis of Se and Sp revealed extreme heterogeneity (I2 ≥ 99.9%), and the positive predictive values (PPVs) reported across studies were consistently low (often < 1%), reflecting the challenge of screening a low-prevalence disease. Conclusions: AI models using EHR data show significant promise for early PC detection, with NNs achieving the highest pooled AUC. However, high heterogeneity and typically low PPV highlight the need for standardized methodologies and a targeted risk-stratification approach rather than general population screening. Future prospective validation and integration into clinical decision-support systems are essential. Full article
Show Figures

Figure 1

17 pages, 2787 KB  
Article
Machine Learning Model for Sepsis Prediction in Prolonged and Chronic Critical Illness: Development and Validation Using Retrospective Real-World ICU Data
by Mikhail Ya. Yadgarov, Olga Yu. Rebrova, Levan B. Berikashvili, Petr A. Polyakov, Kristina K. Kadantseva, Alexey A. Yakovlev, Andrey V. Grechko and Valery V. Likhvantsev
J. Clin. Med. 2026, 15(2), 777; https://doi.org/10.3390/jcm15020777 - 18 Jan 2026
Viewed by 138
Abstract
Background: No machine learning (ML) models for sepsis prediction have been specifically developed for patients with prolonged or chronic critical illness (PCI/CCI). Objective: This study aimed to develop and validate an ML-based sepsis prediction model for this cohort. Methods: We analyzed ICU admissions [...] Read more.
Background: No machine learning (ML) models for sepsis prediction have been specifically developed for patients with prolonged or chronic critical illness (PCI/CCI). Objective: This study aimed to develop and validate an ML-based sepsis prediction model for this cohort. Methods: We analyzed ICU admissions from the Russian Intensive Care Dataset (RICD, 575 patients with PCI/CCI) and two public ICU datasets from the PhysioNet (>40,000 patients with acute critical illness). Models were trained within a right-aligned prediction framework using a case–crossover–control sampling approach and a 6 h prediction window. Two strategies were evaluated: (1) a PCI/CCI-focused model trained on RICD with external testing on PhysioNet data and (2) a universal model trained on combined RICD and PhysioNet cohorts. Models were developed with tree-based algorithms (XGBoost, LightGBM, Random Forest, AdaBoost), with internal and external validation. Primary outcome was model discrimination (AUROC). Subgroup analyses were performed for sepsis phenotypes. Results: The PCI/CCI-focused XGBoost model achieved an AUROC of 0.82 in the RICD cohort but failed to generalize to external ICU populations (AUROC 0.47). A universal model trained on mixed data demonstrated reduced discrimination in PCI/CCI patients (AUROC mean difference 0.02, p = 0.0012). Respiratory rate, heart rate, body temperature, and age were among the most important features. Predictive performance was higher in hypoinflammatory sepsis phenotype (AUROC 0.84 vs. 0.81 for hyperinflammatory, p < 0.001). Despite worthless positive predictive value (up to 21%) for PCI/CCI-focused model, negative predictive value exceeded 97%. Conclusions: A right-aligned ML model tailored to PCI/CCI demonstrated strong internal performance for sepsis exclusion but limited cross-population generalizability, underscoring the need for population-specific prediction models and prospective validation before clinical application. Full article
(This article belongs to the Section General Surgery)
Show Figures

Figure 1

12 pages, 2513 KB  
Article
Missing Data in OHCA Registries: How Multiple Imputation Methods Affect Research Conclusions—Paper II
by Stella Jinran Zhan, Seyed Ehsan Saffari, Marcus Eng Hock Ong and Fahad Javaid Siddiqui
J. Clin. Med. 2026, 15(2), 732; https://doi.org/10.3390/jcm15020732 - 16 Jan 2026
Viewed by 124
Abstract
Background/Objectives: Missing data in clinical observational studies, such as out-of-hospital cardiac arrest (OHCA) registries, can compromise statistical validity. Single imputation methods are simple alternatives to complete-case analysis (CCA) but do not account for imputation uncertainty. Multiple imputation (MI) is the standard for handling [...] Read more.
Background/Objectives: Missing data in clinical observational studies, such as out-of-hospital cardiac arrest (OHCA) registries, can compromise statistical validity. Single imputation methods are simple alternatives to complete-case analysis (CCA) but do not account for imputation uncertainty. Multiple imputation (MI) is the standard for handling missing-at-random (MAR) data, yet its implementation remains challenging. This study evaluated the performance of MI in association analysis compared with CCA and single imputation methods. Methods: Using a simulation framework with real-world Singapore OHCA registry data (N = 13,274 complete cases), we artificially introduced 20%, 30%, and 40% missingness under MAR. MI was implemented using predictive mean matching (PMM), random forest (RF), and classification and regression trees (CART) algorithms, with 5–20 imputations. Performance was assessed based on bias and precision in a logistic regression model evaluating the association between alert issuance and bystander CPR. Results: CART outperformed PMM, providing more accurate β coefficients and stable CIs across missingness levels. Although K-Nearest Neighbours (KNN) produced similar point estimates, it underestimated imputation uncertainty. PMM showed larger bias, wider and less stable CIs, and in some settings performed similarly to CCA. MI methods produced wider CIs than single imputation, appropriately capturing imputation uncertainty. Increasing the number of imputations had minimal impact on point estimates but modestly narrowed CIs. Conclusions: MI performance depends strongly on the chosen algorithm. CART and RF methods offered the most robust and consistent results for OHCA data, whereas PMM may not be optimal and should be selected with caution. MI using tree-based methods (CART/RF) remains the preferred strategy for generating reliable conclusions in OHCA research. Full article
Show Figures

Figure 1

27 pages, 2279 KB  
Article
Sustainability-Driven Design Optimization of Aircraft Parts Using Mathematical Modeling
by Aikaterini Anagnostopoulou, Dimitris Sotiropoulos, Ioannis Sioutis and Konstantinos Tserpes
Aerospace 2026, 13(1), 95; https://doi.org/10.3390/aerospace13010095 - 15 Jan 2026
Viewed by 182
Abstract
The design of aircraft components is a complex process that must simultaneously account for environmental impact, manufacturability, cost and structural performance to meet modern regulatory requirements and sustainability objectives. When these factors are integrated from the early design stages, the approach transcends traditional [...] Read more.
The design of aircraft components is a complex process that must simultaneously account for environmental impact, manufacturability, cost and structural performance to meet modern regulatory requirements and sustainability objectives. When these factors are integrated from the early design stages, the approach transcends traditional eco-design and becomes a genuinely sustainability-oriented design methodology. This study proposes a sustainability-driven design framework for aircraft components and demonstrates its application to a fuselage panel consisting of a curved skin, four frames, seven stringers, and twenty-four clips. The design variables investigated include the material selection, joining methods, and subcomponent thicknesses. The design space is constructed through a combinatorial generation process coupled with compatibility and feasibility constraints. Sustainability criteria are evaluated using a combination of parametric Life Cycle Assessment (LCA) and Life Cycle Costing (LCC) regression models, parametric Finite Element Analysis (FEA), and Random Forest surrogate modeling trained on a stratified set of simulation results. Two methodological pathways are introduced: 1. Cluster-based optimization, involving customized clustering followed by multi-criteria decision-making (MCDM) within each cluster. 2. Global optimization, performed across the full decision matrix using Pareto front analysis and MCDM techniques. A stability analysis of five objective-weighting methods and four normalization techniques is conducted to identify the most robust methodological configuration. The results—based on a full cradle-to-grave assessment that includes the use phase over a 30-year A319 aircraft operational lifetime—show that the thermoplastic CFRP panel joined by welding emerges as the most sustainable design alternative. Full article
(This article belongs to the Special Issue Composite Materials and Aircraft Structural Design)
Show Figures

Figure 1

22 pages, 3392 KB  
Systematic Review
Factors Affecting the Treatment Heterogeneity of PPARγ and Pan-PPAR Agonists in Type 2 Diabetes Mellitus: A Systematic Review and Machine Learning-Based Meta-Regression Analysis
by Xinlei Zhang, Yingning Liu, Ming Chu, Linong Ji and Xiantong Zou
Pharmaceuticals 2026, 19(1), 139; https://doi.org/10.3390/ph19010139 - 13 Jan 2026
Viewed by 197
Abstract
Background/Objectives: Significant heterogeneity in the treatment response to peroxisome proliferator-activated receptor γ (PPARγ) agonists exists, and predictive factors for their efficacy remain unclear. We aimed to assess the relationships between routinely available clinical features and the efficacy of PPARγ agonists and pan-PPAR [...] Read more.
Background/Objectives: Significant heterogeneity in the treatment response to peroxisome proliferator-activated receptor γ (PPARγ) agonists exists, and predictive factors for their efficacy remain unclear. We aimed to assess the relationships between routinely available clinical features and the efficacy of PPARγ agonists and pan-PPAR agonists by meta-regression analysis. Methods: We searched PubMed, Embase, Cochrane Library, ClinicalTrials.gov and the WHO International Clinical Trials Registry Platform (ICTRP) and included randomised controlled trials involving type 2 diabetes patients with 12-week or longer treatment durations with PPARγ agonists or pan-PPAR agonists published before 11 November 2023 (PROSPERO registration number: CRD42024578987). We conducted mixed-effect meta-regression analyses between baseline variables and treatment response. Moreover, we developed a machine learning-based meta-forest model and ranked the relative importance of each variable. Results: In 147 studies involving 29,250 participants, PPARγ and pan-PPAR agonists significantly reduced HbA1c (mean difference(MD) = −0.8876 [95% confidence interval (CI): −0.8999, −0.8754]; p < 0.0001, I2 = 96.0%) and FPG = (MD = −1.7900 [95% CI: −1.9137, −1.6663]; p < 0.0001, I2 = 92.0%). Multivariable association analysis suggested that a greater proportion of female participants (β = 0.0066 [95% CI: 0.0012, 0.0121]; p = 0.017), younger age (β = −0.0314 [95% CI: −0.05, −0.0129]; p = 0.0009) and lower HDL-C levels (β = −0.9304 [95% CI: −1.5176, −0.3431]; p = 0.0019) were significantly associated with a greater decrease in HbA1c. A greater proportion of female participants (β = 0.0112 [95% CI: 0.0019, 0.0205]; p = 0.0178) and lower baseline HDL-C levels (β = −1.8722 [95% CI: −2.812, −0.9323]; p < 0.0001) were significantly associated with a greater decrease in FPG. These variables also ranked among the top five most important predictors of drug response in the meta-random forest models. Conclusions: Our study demonstrated that female sex, younger age, and lower HDL-C levels were associated with greater glycaemic lowering effect from PPARγ and pan-PPAR agonists. Full article
(This article belongs to the Section Pharmacology)
Show Figures

Figure 1

10 pages, 1829 KB  
Proceeding Paper
Machine Learning Based Agricultural Price Forecasting for Major Food Crops in India Using Environmental and Economic Factors
by P. Ankit Krishna, Gurugubelli V. S. Narayana, Siva Krishna Kotha and Debabrata Pattnayak
Biol. Life Sci. Forum 2025, 54(1), 7; https://doi.org/10.3390/blsf2025054007 - 12 Jan 2026
Viewed by 217
Abstract
The contemporary agricultural market is profoundly volatile, where agricultural prices are based on a complex supply chain, climatic irregularity or unscheduled market demand. Prices of crops need to be predicted in a reliable and timely manner for farmers, policy-makers and other stakeholders to [...] Read more.
The contemporary agricultural market is profoundly volatile, where agricultural prices are based on a complex supply chain, climatic irregularity or unscheduled market demand. Prices of crops need to be predicted in a reliable and timely manner for farmers, policy-makers and other stakeholders to take evidence-based decisions ultimately for the benefit towards sustainable agriculture and economic sustainability. Objective: The objective of this study is to develop and evaluate a comprehensive machine learning model for predicting agricultural prices incorporating logistic, economic and environmental considerations. It is the desire to make agriculture more profitable by building simple and accurate forecasting models. Methods: An assorted dataset was collected, which covers major factors to constitute the dataset of temperature, rainfall, fertiliser use, pest and disease attack level, cost of transportation, market demand-supply ratio and regional competitiveness. The data was subjected to pre-processing and feature extraction for quality control/quality assurance. Several machine learning models (Linear Regression, Support Vector Machines, AdaBoost, Random Forest, and XGBoost) were trained and evaluated using performance metrics such as R2 score, Root Mean Squared Error (RMSE), and Mean Absolute Error (MAE). Results: Out of the model approaches that were analysed, predictive performance was superior for XGBoost (with an R2 Score of 0.94, RMSE of 12.8 and MAE of 8.6). To generate accurate predictions, the ability to account for complex non-linear relationships between market and environmental information was necessary. Conclusions: The forecast model of the XGBoost-based prediction system is reliable, of low complexity and widely applicable to large-scale real-time forecasting of agricultural monitoring. The model substantially reduces the uncertainty of price forecasting, and does so by including multivariate environmental and economic aspects that permit more profitable management practices in a schedule for future sustainable agriculture. Full article
(This article belongs to the Proceedings of The 3rd International Online Conference on Agriculture)
Show Figures

Figure 1

Back to TopTop