Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (8,818)

Search Parameters:
Keywords = metric learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
42 pages, 16476 KB  
Article
PIMSEL: A Physically Guided Multi-Modal Semi-Supervised Learning Framework for Earthquake-Induced Landslide Reactivation Risk Assessment
by Bingxin Shi, Hongmei Guo, Zongheng He, Shi Chen, Jia Guo, Yunxi Dong, Bingyang Shi, Jingren Zhou, Yusen He and Huajin Li
Remote Sens. 2026, 18(9), 1320; https://doi.org/10.3390/rs18091320 (registering DOI) - 25 Apr 2026
Abstract
Earthquake-induced landslide reactivation poses a sustained hazard for years following major seismic events, yet operational prediction remains constrained by heterogeneous multi-modal data, sparse supervision, and the absence of uncertainty-aware frameworks. This paper presents PIMSEL, a physically guided multi-modal semi-supervised framework for post-seismic landslide [...] Read more.
Earthquake-induced landslide reactivation poses a sustained hazard for years following major seismic events, yet operational prediction remains constrained by heterogeneous multi-modal data, sparse supervision, and the absence of uncertainty-aware frameworks. This paper presents PIMSEL, a physically guided multi-modal semi-supervised framework for post-seismic landslide reactivation risk assessment. PIMSEL integrates satellite-derived morphological features, precipitation time series, and seismic hazard attributes through four components: entropy-regularized optimal transport for cross-modal semantic alignment without paired supervision; causally constrained hierarchical fusion enforcing domain-consistent modal weighting; scenario-based prototype mutation for semi-supervised learning from sparse expert annotations; and prototype-anchored variational graph clustering that simultaneously stratifies landslides into HIGH, MEDIUM, and LOW risk tiers and produces decomposed aleatoric and epistemic uncertainty estimates for operational triage. The HIGH risk tier operationally corresponds to predicted reactivation, validated against 598 documented reactivation events across 7482 co-seismic landslides from three Sichuan Province earthquake sequences: the 2013 Lushan (Mw 7.0), 2017 Jiuzhaigou (Mw 7.0), and 2022 Luding (Mw 6.8) events. PIMSEL achieves 82.5% reactivation recall and 66.4% precision, outperforming twelve baselines across clustering quality, classification, and uncertainty calibration metrics. Ablation studies confirm that optimal transport alignment contributes the largest individual performance gain. Current limitations include quarterly assessment frequency and dependence on optical imagery under cloud cover, which future integration of real-time meteorological triggers and SAR data should address. Full article
24 pages, 24917 KB  
Article
BCDA-Net: A Bottleneck-Free Channel Dual-Path Aggregation Network for Infrared Image Destriping
by Lingzhi Chen, Feng Dong, Lingfeng Huang and Yutian Fu
Remote Sens. 2026, 18(9), 1321; https://doi.org/10.3390/rs18091321 (registering DOI) - 25 Apr 2026
Abstract
The inherent non-uniformity of Infrared Focal Plane Arrays (IRFPA) inevitably results in stripe noise, which severely degrades image quality and hinders downstream applications. Existing deep learning methods often struggle to strike a balance between effective denoising and the preservation of fine thermal textures. [...] Read more.
The inherent non-uniformity of Infrared Focal Plane Arrays (IRFPA) inevitably results in stripe noise, which severely degrades image quality and hinders downstream applications. Existing deep learning methods often struggle to strike a balance between effective denoising and the preservation of fine thermal textures. To address this issue, we propose a Bottleneck-free Channel Dual-path Aggregation Network (BCDA-Net) based on a “Perception-Reconstruction” design principle. In the perception stage, the network jointly employs the Dual-Path Channel Down-sampling (DCD) module and the Context-Guided Stripe Attention Block (CGSAB). The DCD module utilizes a channel split strategy to simultaneously extract semantic features and preserve high-frequency textures, while the CGSAB performs global context modeling on these features to precisely perceive and locate global stripe noise patterns. In the reconstruction stage, we integrate the Cascaded Dense Feature Aggregation (CDFA) module with a Bottleneck-Free Aggregation Strategy (BFAS). The CDFA utilizes the perceived information to densely aggregate features and progressively reconstruct clean image details, whereas the BFAS structurally blocks the propagation of low-resolution noise during decoding, effectively mitigating aliasing artifacts induced by deep feature upsampling. Together, these components form a complete closed loop from accurate noise perception to high-fidelity reconstruction. Extensive experiments on public and real-world datasets demonstrate that BCDA-Net maximally preserves image details while removing non-uniform stripe noise. Both objective metrics and subjective visual quality outperform existing state-of-the-art methods. Full article
(This article belongs to the Section Remote Sensing Image Processing)
22 pages, 3386 KB  
Article
UAV Visual Localization via Multimodal Fusion and Multi-Scale Attention Enhancement
by Yiheng Wang, Yushuai Zhang, Zhenyu Wang, Jianxin Guo, Feng Wang, Rui Zhu and Dejing Lin
Sustainability 2026, 18(9), 4277; https://doi.org/10.3390/su18094277 (registering DOI) - 25 Apr 2026
Abstract
For power-grid applications such as transmission corridor inspection, substation asset inspection, and post-disaster emergency repair, reliable UAV self-localization under GNSS-degraded or GNSS-denied conditions is critical to ensuring operational safety and accurate defect geotagging. Due to substantial discrepancies in viewpoint, scale, and geometric structure [...] Read more.
For power-grid applications such as transmission corridor inspection, substation asset inspection, and post-disaster emergency repair, reliable UAV self-localization under GNSS-degraded or GNSS-denied conditions is critical to ensuring operational safety and accurate defect geotagging. Due to substantial discrepancies in viewpoint, scale, and geometric structure between oblique UAV images and nadir satellite images, conventional RGB-based cross-view retrieval methods often suffer from unstable alignment and insufficient geometric modeling, particularly in scenarios with repetitive textures and partial overlap. To address these challenges, we propose a cross-view visual geo-localization model that integrates RGBD multimodal inputs with multi-scale attention enhancement. Specifically, MiDaS is used to estimate relative depth from UAV imagery, which is concatenated with RGB to form a four-channel input, while satellite images are padded with an additional zero channel to maintain dimensional consistency. A shared-weight ViTAdapter is adopted to learn joint semantic–geometric representations, and a lightweight Efficient Multi-scale Attention (EMA) module is adopted on spatial feature maps to strengthen multi-scale spatial consistency. In addition, an IoU-weighted InfoNCE loss is employed to accommodate partial matching during training, thereby improving the robustness of feature alignment. Experiments on the GTA-UAV dataset under the cross-area protocol show stable performance across both retrieval and localization metrics. Specifically, Recall@1, Recall@5, and Recall@10 reach 18.12%, 38.83%, and 49.47%, respectively; AP is 28.01 and SDM@3 is 0.53; meanwhile, the top-1 geodesic distance error Dis@1 is 1052.73 m. These results indicate that explicit geometric priors combined with multi-scale spatial enhancement can effectively improve cross-view feature alignment, leading to enhanced robustness and accuracy for localization in challenging power inspection scenarios. Full article
Show Figures

Figure 1

52 pages, 2293 KB  
Review
From Model-Driven to AI-Native Physical Layer Design: Deep Learning Architectures and Optimization Paradigms for Wireless Communications
by Evelio Astaiza Hoyos, Héctor Fabio Bermúdez-Orozco and Nasly Cristina Rodriguez-Idrobo
Information 2026, 17(5), 410; https://doi.org/10.3390/info17050410 (registering DOI) - 25 Apr 2026
Abstract
The increasing complexity of next-generation wireless systems challenges the scalability and generalization capabilities of traditional model-driven physical layer (PHY) design, which relies on analytically derived channel models and optimization frameworks. This paper presents a comprehensive survey and critical review of deep learning (DL) [...] Read more.
The increasing complexity of next-generation wireless systems challenges the scalability and generalization capabilities of traditional model-driven physical layer (PHY) design, which relies on analytically derived channel models and optimization frameworks. This paper presents a comprehensive survey and critical review of deep learning (DL) architectures enabling the transition toward AI-native PHY design. A unified optimization perspective is developed in which all PHY tasks—including channel estimation, channel state information (CSI) feedback, massive MIMO processing, signal detection, channel coding, beamforming, resource allocation, and semantic-aware transmission—are formulated under a common empirical risk minimization (ERM) framework. Neural architectures such as autoencoders, convolutional and recurrent networks, transformers, and reinforcement learning models are examined through their underlying optimization formulations, loss functions, training methodologies, and representation learning mechanisms. The review compares model-driven and AI-native approaches in terms of performance metrics, computational complexity, robustness, generalization capability, and practical deployment constraints, including hardware limitations, energy efficiency, and real-time feasibility. The analysis highlights the conditions under which AI-native architectures provide adaptability and performance improvements while identifying trade-offs in complexity, latency, and interpretability. The study concludes by outlining prioritized research directions toward fully adaptive and self-optimizing wireless communication systems. Full article
(This article belongs to the Section Wireless Technologies)
25 pages, 8307 KB  
Article
A Physics–Data Hybrid Framework Using Uncalibrated Consumer CMOS Vision: Pilot Study on Monocular Automatic TUG Assessment Towards Early Parkinson’s Disease Risk Screening
by Yuxiang Qiu, Xiaodong Sun, Fan Yang, Jarred Fastier-Wooller, Shun Muramatsu, Michitaka Yamamoto and Toshihiro Itoh
Micromachines 2026, 17(5), 523; https://doi.org/10.3390/mi17050523 (registering DOI) - 25 Apr 2026
Abstract
The Timed Up and Go (TUG) test is a clinical gold standard for assessing elderly mobility, yet its automated deployment in home-monitoring and resource-limited areas is hindered by high hardware costs and expert calibration requirements. This study introduces a Physics–Data Hybrid framework specifically [...] Read more.
The Timed Up and Go (TUG) test is a clinical gold standard for assessing elderly mobility, yet its automated deployment in home-monitoring and resource-limited areas is hindered by high hardware costs and expert calibration requirements. This study introduces a Physics–Data Hybrid framework specifically designed for uncalibrated consumer-grade CMOS cameras, enabling a “plug-and-play” solution for early Parkinson’s disease (PD) risk screening. The proposed pipeline integrates learning-based pose perception with a self-evolving physics model to recover absolute metric-scale motion without manual checkerboard calibration. A noise-adaptive fusion strategy is implemented to reconcile 2D pixel dynamics with 3D kinematic consistency, overcoming the inherent scale ambiguity of monocular vision. Crucially, this framework enables the extraction of high-dimensional spatiotemporal parameters—such as stride length coefficient of variation and mean gait velocity—which provide a finer diagnostic resolution for capturing subtle motor fluctuations than conventional timing-only systems. Results from our pilot study with a cohort of 10 subjects demonstrate that these extracted metric features serve as decisive markers for risk staging simulated by dual-task-induced cognitive-motor-interference, achieving 98% screening accuracy and an overall classification accuracy of 87.32%. This framework provides a robust, low-cost tool for ubiquitous telehealth, potentially supporting early PD risk assessment in underserved populations. Full article
38 pages, 6938 KB  
Article
DeepSense: An Adaptive Scalable Ensemble Framework for Industrial IoT Anomaly Detection
by Amir Firouzi and Ali A. Ghorbani
Sensors 2026, 26(9), 2662; https://doi.org/10.3390/s26092662 (registering DOI) - 24 Apr 2026
Abstract
The Industrial Internet of Things (IIoT) has become a cornerstone of modern industrial automation, enabling real-time monitoring, intelligent decision-making, and large-scale connectivity across cyber–physical systems. However, the growing scale, heterogeneity, and dynamic behavior of IIoT environments significantly expand the attack surface and challenge [...] Read more.
The Industrial Internet of Things (IIoT) has become a cornerstone of modern industrial automation, enabling real-time monitoring, intelligent decision-making, and large-scale connectivity across cyber–physical systems. However, the growing scale, heterogeneity, and dynamic behavior of IIoT environments significantly expand the attack surface and challenge the effectiveness of conventional security mechanisms. In this paper, we propose DeepSense, a hybrid and adaptive anomaly and intrusion detection framework specifically designed for resource-constrained and heterogeneous IIoT deployments. DeepSense integrates three complementary components: DataSense, a realistic data pipeline and experimental testbed supporting synchronized sensor and network data processing; RuleSense, a lightweight rule-based detection layer that provides fast, deterministic, and interpretable anomaly screening at the edge; and NeuroSense, a learning-driven detection module comprising an adaptive ensemble of 22 machine learning and deep learning models spanning classical, neural, hybrid, and Transformer-based architectures. NeuroSense operates as a second detection stage that validates suspicious events flagged by RuleSense and enables both coarse-grained and fine-grained attack classification. To support rigorous and practical assessment, this work further introduces a comprehensive performance evaluation framework that extends beyond accuracy-centric metrics by jointly considering detection quality, latency, resource efficiency, and detection coverage, alongside an optimization-based process for selecting Pareto-optimal model ensembles under realistic IIoT constraints. Extensive experiments across diverse detection scenarios demonstrate that DeepSense exhibits strong generalization, lower false positive rates, and robust performance under evolving attack behaviors. The proposed framework provides a scalable and efficient IIoT security solution that meets the operational requirements of Industry 4.0 and the resilience-oriented objectives of Industry 5.0. Full article
51 pages, 7385 KB  
Article
Spiking Neural Networks with Continual Learning for Steering Angle Regression: A Sustainable AI Perspective
by Fernando S. Martínez, Sergio Costa and Raúl Parada
Sensors 2026, 26(9), 2656; https://doi.org/10.3390/s26092656 - 24 Apr 2026
Abstract
This work explores the application of Spiking Neural Networks (SNNs) and Continual Learning (CL) methodologies to the problem of steering angle regression, using autonomous driving simulation as the experimental context, with a focus on energy efficiency and alignment with sustainable computing objectives. The [...] Read more.
This work explores the application of Spiking Neural Networks (SNNs) and Continual Learning (CL) methodologies to the problem of steering angle regression, using autonomous driving simulation as the experimental context, with a focus on energy efficiency and alignment with sustainable computing objectives. The primary goal was to design and implement CL techniques in SNNs to assess the model’s ability to maintain accuracy in explored environments while reducing CO2 emissions through the optimized use of a subset of the data. This study emerges in response to the increasing energy demand of deep learning models, which poses a challenge to sustainability. SNNs, inspired by the efficiency of biological neural systems, offer significant advantages in terms of computational and energy consumption, making them a promising alternative. CL techniques, such as Elastic Weight Consolidation and replay memory, are integrated to mitigate catastrophic forgetting in sequential learning tasks. The methodology includes adapting the PilotNet architecture for SNNs, preprocessing datasets generated in the Udacity driving simulator, and evaluating models in incremental learning scenarios. The experiments compare the performance of SNNs with CL against baseline models without CL, using mean squared error (MSE), computational efficiency, and equivalent CO2 emissions as evaluation metrics. The results demonstrate that replay memory enables the retention of prior knowledge with a limited increase in energy consumption. This work concludes that SNNs with CL are a viable alternative for sustainable AI applications. Future research directions include a focus primarily on hardware-specific implementations and real-world testing. Full article
66 pages, 1148 KB  
Review
Explainability and Trust in Deep Learning for Cancer Imaging: Systematic Barriers, Clinical Misalignment, and a Translational Roadmap
by Surekha Borra, Nilanjan Dey, Simon Fong, R. Simon Sherratt and Fuqian Shi
Cancers 2026, 18(9), 1361; https://doi.org/10.3390/cancers18091361 - 24 Apr 2026
Abstract
Deep learning (DL) has transformed cancer imaging by enabling automated tumour detection, classification, and risk prediction. Despite impressive diagnostic performance, limited explainability and poor uncertainty calibration continue to restrict clinical integration. This review is guided by five research questions that examine the challenges, [...] Read more.
Deep learning (DL) has transformed cancer imaging by enabling automated tumour detection, classification, and risk prediction. Despite impressive diagnostic performance, limited explainability and poor uncertainty calibration continue to restrict clinical integration. This review is guided by five research questions that examine the challenges, impact, and translational implications of explainable artificial intelligence (XAI) in oncology imaging. We identify key barriers to trust, including dataset bias, shortcut learning, opacity of convolutional neural networks, and workflow misalignment. Evidence suggests that explainable models can increase clinician confidence, reduce false positives, and improve collaborative decision-making when explanations are faithful, semantically meaningful, and uncertainty aware. We evaluate architectural strategies that embed interpretability such as concept-bottleneck models, prototype-based learning, and attention regularization along with post hoc techniques. Beyond performance metrics, we examine how interpretable AI aligns with clinical reasoning processes and analyse regulatory, ethical, and medico-legal considerations influencing deployment. The findings indicate that explainability alone is insufficient, durable trust requires epistemic alignment, prospective validation, lifecycle governance, and equity-focused evaluation. By reframing explainability as a structural design principle rather than a supplementary feature, this review outlines a pathway toward accountable and clinically dependable AI systems in oncology. Full article
(This article belongs to the Section Cancer Informatics and Big Data)
Show Figures

Figure 1

22 pages, 2381 KB  
Article
An RMST-Integrated Machine Learning Framework for Interpretable Survival Analysis Under Non-Proportional Hazards: Application to the METABRIC Cohort
by Fangya Tan, Yang Zhou, Shuqiao Li, Chun Jiang, Jian-Guo Zhou and Srikar Bellur
Algorithms 2026, 19(5), 329; https://doi.org/10.3390/a19050329 - 24 Apr 2026
Abstract
(1) Background: Advances in machine learning (ML)-based survival modeling enable the analysis of high-dimensional biomedical data. However, many approaches rely on the proportional hazards (PH) assumption, which is frequently violated in oncology and can limit the interpretability of hazard ratio-based results. Using Estrogen [...] Read more.
(1) Background: Advances in machine learning (ML)-based survival modeling enable the analysis of high-dimensional biomedical data. However, many approaches rely on the proportional hazards (PH) assumption, which is frequently violated in oncology and can limit the interpretability of hazard ratio-based results. Using Estrogen Receptor (ER) status in the METABRIC breast cancer cohort as a case study, we propose a framework that integrates machine learning survival models with Restricted Mean Survival Time (RMST) to provide a more robust and clinically interpretable approach for survival analysis under non-proportional hazards. (2) Methods: Overall survival was analyzed in 1104 patients. PH violations were confirmed using Schoenfeld residuals and Kaplan–Meier inspection. We compared four models: stratified Cox Elastic Net (Cox E-Net), Random Survival Forest (RSF), Gradient Boosting Survival Analysis (GBSA), and DeepHit. Performance was assessed using Harrell’s C-index, time-dependent IPCW C-index, and Integrated Brier Score (IBS). RMST at 180 months was utilized to quantify absolute survival differences between ER subgroups. To improve the stability of the estimates, 200 bootstrap resamples were performed, and 95% confidence intervals were derived from the bootstrap distribution. (3) ER status demonstrated significant PH violation (p < 0.005) with crossing survival curves. Discrimination (C-index 0.664–0.725) and calibration (IBS 0.149–0.169) were comparable across models, with RSF achieving the highest overall performance. Despite similar accuracy, survival curve structures differed substantially. Cox E-Net and RSF reproduced the observed crossing pattern, whereas GBSA generated smoother trajectories and DeepHit showed marked compression of subgroup separation. In the independent test cohort, the empirical RMST difference at 180 months was 16.6 months (ER-positive: 130.4; ER-negative: 113.8). Model-based RMST differences ranged from 1 month (DeepHit) to 27 months (Cox E-Net), with RSF and GBSA (12.8 and 13.8 months) most closely approximating the empirical benchmark. (4) Conclusions: We propose a novel, model-agnostic ML + RMST framework that addresses non-proportional hazards while providing quantifiable, time-specific clinical benefit. Moreover, models with similar discrimination and calibration produced markedly different survival curve behavior and absolute RMST estimates, demonstrating that accuracy metrics alone are insufficient for clinical interpretation. By linking prognostic modeling with absolute survival quantification, this framework advances survival evaluation beyond relative risk ranking toward individualized, clinically meaningful decision support. Full article
Show Figures

Figure 1

25 pages, 5193 KB  
Article
Scenario-Adaptive Visibility Level Retrieval via Multi-Source Synergy: Enhancing Physical Traceability and Scene Decoupling Within a Tree-Routed TabPFN Framework
by Chuhan Lu, Shanwen Luo and Zhiyuan Han
Remote Sens. 2026, 18(9), 1307; https://doi.org/10.3390/rs18091307 - 24 Apr 2026
Abstract
Accurate retrieval of visibility grades is critical for transportation safety. Due to the highly complex meteorological backgrounds, traditional global deep learning models frequently struggle with limited physical traceability and feature heterogeneity. To address these challenges by enhance physical traceability and reduces heterogeneity, this [...] Read more.
Accurate retrieval of visibility grades is critical for transportation safety. Due to the highly complex meteorological backgrounds, traditional global deep learning models frequently struggle with limited physical traceability and feature heterogeneity. To address these challenges by enhance physical traceability and reduces heterogeneity, this study proposes a scenario-adaptive visibility retrieval framework based on multi-source synergy, namely TabPFN-ExtraTrees (TabPFN-ET), targeting major transportation routes in Anhui Province, China. Fusing Fengyun-4 (FY-4A/4B) satellite multispectral observations with ground meteorological data, this framework utilizes the divide-and-conquer routing mechanism of ExtraTrees to decouple the complex, heterogeneous feature space into highly homogeneous sub-scenarios. Subsequently, the TabPFN model conducts high-precision inference within each specific subspace. Evaluations on a class-balanced benchmark demonstrate that TabPFN-ET achieves an Overall Accuracy of 0.681, outperforming baseline models such as SAINT across various metrics. Furthermore, this paper conducts a physically consistent analysis of the framework. Feature importance and node profiling corroborate its physical consistency: the FY-4 upper-level water vapor channel (Channel 09) and near-surface humidity act as the macroscopic atmospheric stability and microscopic thermodynamic constraints, respectively, driving the model’s scene decoupling and inference. Cross-regional tests in Jiangsu provide preliminary indications of context-specific transferability. Full article
26 pages, 1015 KB  
Article
AI-Driven Biopsychosocial Screening for Breast Cancer: Enhancing Risk Prediction via Differential Evolutionary Linear Discriminant Analysis for Feature Extraction
by José Luis Llaguno-Roque, Adriana Laura López-Lobato, Juan Carlos Pérez-Arriaga, Héctor Gabriel Acosta-Mesa, Ángel J. Sánchez-García, Gabriel Gutiérrez-Ospina, Antonia Barranca-Enríquez and Tania Romo-González
Math. Comput. Appl. 2026, 31(3), 66; https://doi.org/10.3390/mca31030066 - 24 Apr 2026
Abstract
In Mexico, the high prevalence and mortality rates associated with breast cancer (BC) constitute a critical public health challenge that demands context-specific preventive measures. This study proposes an integrative framework for predicting BC risk based on a biopsychosocial model. We hypothesize that emotional [...] Read more.
In Mexico, the high prevalence and mortality rates associated with breast cancer (BC) constitute a critical public health challenge that demands context-specific preventive measures. This study proposes an integrative framework for predicting BC risk based on a biopsychosocial model. We hypothesize that emotional suppression and repression act as key neuroendocrine disruptors and predisposing factors within the Mexican female population. To test this, we systematically compared the predictive performance of various machine learning classification models using the clinical, psychological, and combined profiles of 110 women. These models were evaluated with and without the application of a robust evolutionary algorithm: Differential Evolutionary Linear Discriminant Analysis for Feature Extraction (DELDAFE). The results demonstrated that integrating clinical and psychological data into a combined latent space significantly improved the performance of the classification algorithms. The Artificial Neural Network achieved the highest metrics (0.9975 Precision; 0.9976 F1-score). However, due to the inherent “black-box” nature of these models (limited clinical interpretability), the Decision Tree emerged as the optimal practical alternative, providing highly competitive (0.8874 Precision; 0.8853 F1-score) and interpretable results. These findings provide empirical evidence that psychological factors, rather than being mere incidental comorbidities, could be associated with the etiology of breast cancer and be used as risk factors in predicting the disease. Ultimately, this AI-driven biopsychosocial screening model offers a scalable, low-cost, and context-adapted risk assessment tool for early BC diagnosis in Mexican women. Full article
(This article belongs to the Special Issue New Trends in Computational Intelligence and Applications 2025)
Show Figures

Figure 1

25 pages, 703 KB  
Review
Eye-Tracking-Based Interventions for School-Age Specific Learning Disorders: A Narrative Review of Functional Assessment and Gaze-Contingent Training
by Pierluigi Diotaiuti, Francesco Di Siena, Salvatore Vitiello, Alessandra Zanon, Pio Alfredo Di Tore and Stefania Mancone
J. Eye Mov. Res. 2026, 19(3), 42; https://doi.org/10.3390/jemr19030042 - 24 Apr 2026
Abstract
Eye tracking (ET) provides process-level indices of how students sample task-relevant information during core academic activities. In school-age learners (6–18 years) with specific learning disorders (SLDs; dyslexia, dysgraphia, and dyscalculia), ET can complement behavioural assessment by quantifying oculomotor patterns linked to decoding, model–production [...] Read more.
Eye tracking (ET) provides process-level indices of how students sample task-relevant information during core academic activities. In school-age learners (6–18 years) with specific learning disorders (SLDs; dyslexia, dysgraphia, and dyscalculia), ET can complement behavioural assessment by quantifying oculomotor patterns linked to decoding, model–production coordination, and stepwise strategy execution. This narrative review synthesises ET findings in SLD across reading, handwriting/copying, and arithmetic and translates them into an applied framework for school-oriented use. We summarise key metrics and Areas of Interest (AOI)-based analyses, highlight technical and data-quality requirements for valid acquisition in educational settings, and outline compact functional assessment protocols integrated with standard academic and neuropsychological measures. Building on these foundations, we propose six hypothesis-driven gaze-contingent paradigms (H1–H6) as preliminary models for future experimental testing rather than as established interventions, and we map each to its current level of empirical support, specifying primary gaze outcomes and curriculum-relevant behavioural endpoints. We emphasise that eye-movement findings in specific learning disorders are heterogeneous and may vary as a function of age, task demands, and comorbidity. Accordingly, credible training effects require retention and transfer probes under standard, non-contingent display conditions, appropriate controls, and explicit developmental interpretation. Eye tracking is positioned as complementary functional evidence and as a platform for experimentally testable, mechanism-based interventions in school-age specific learning disorders. Full article
(This article belongs to the Special Issue Eye Movements in Reading and Related Difficulties)
Show Figures

Figure 1

28 pages, 880 KB  
Article
Prediction Pipeline Selection for Incomplete Clinical Data via Missingness Fingerprints and Instance Augmentation
by Runze Li, Zhuyi Shen, Chengkai Wu, Jingsong Li and Yu Tian
Bioengineering 2026, 13(5), 497; https://doi.org/10.3390/bioengineering13050497 (registering DOI) - 24 Apr 2026
Abstract
Clinical prediction from electronic health records (EHRs) is complicated by pervasive missingness and label scarcity, which make performance sensitive to the match between data conditions and pipeline choice. Choosing the best pipeline for a new incomplete dataset still requires costly trial-and-error. We cast [...] Read more.
Clinical prediction from electronic health records (EHRs) is complicated by pervasive missingness and label scarcity, which make performance sensitive to the match between data conditions and pipeline choice. Choosing the best pipeline for a new incomplete dataset still requires costly trial-and-error. We cast this as an algorithm selection problem and address two bottlenecks—instance scarcity and distance quality—that have so far prevented meta-learning from reaching clinical settings. Graph neural networks offer diverse strategies (patient similarity networks, bipartite imputation graphs, attention-driven feature interaction), yet no single architecture dominates across missingness patterns, and selecting the best pipeline for a new dataset remains a trial-and-error approach. Formal algorithm selection could automate this choice but requires many characterized meta-instances—more than clinical settings typically provide. We propose two solutions: (1) constructive instance augmentation, applying controlled quality perturbations (MCAR and MNAR missingness injection, label trimming) to 20 base EHR datasets to expand the meta-knowledge base to 83 characterized meta-instances, each described by a 10-dimensional missingness fingerprint, without additional model training; and (2) dynamic-supervised metric learning, using differential evolution to optimize fingerprint feature weights so that static distances preserve method-performance similarity captured by dynamic fingerprints, which require model sweeps and are unavailable at deployment. Under base-dataset-level leave-one-dataset-out cross-validation over 21 pipelines, the resulting metric-learned kNN recommender attains the highest win rate (20.5%) among non-oracle strategies on the augmented store, selecting the correct pipeline more often than any fixed default. At deployment, the recommender needs only the 10-dimensional static fingerprint with pre-learned weights; no sweep data is required for new datasets. Cross-domain evaluation on 25 external subsets (colorectal cancer, kidney disease, MIMIC-IV) demonstrates framework modularity: when the fingerprint module is adapted (standard meta-features in place of the missingness-specific set), the recommender achieves regret of 0.025 (55% below random selection). Full article
31 pages, 6114 KB  
Article
A Multi-Stage YOLOv11-Based Deep Learning Framework for Robust Instance Segmentation and Material Quantification of Mixed Plastic Waste
by Andrew N. Shafik, Mohamed H. Khafagy, Alber S. Aziz and Shereen A. Hussein
Computers 2026, 15(5), 271; https://doi.org/10.3390/computers15050271 - 24 Apr 2026
Abstract
Instance segmentation in heterogeneous waste scenes remains challenging due to object variability, deformable shapes, partial occlusion, and large appearance differences across packaging types. This study presents a YOLOv11-based deep learning framework for mixed plastic waste instance segmentation, developed to connect visual perception with [...] Read more.
Instance segmentation in heterogeneous waste scenes remains challenging due to object variability, deformable shapes, partial occlusion, and large appearance differences across packaging types. This study presents a YOLOv11-based deep learning framework for mixed plastic waste instance segmentation, developed to connect visual perception with reliable material quantification. The framework integrates curated instance-level annotations, strict split isolation, multi-stage optimization, training strategy ablation, and seed-robustness analysis to support reproducible model selection. Experimental results on a held-out test set show that the optimized model achieves a mask mAP@50:95 of 0.9337, indicating strong segmentation performance under heterogeneous waste-scene conditions. To extend the analysis beyond standard vision metrics, the framework incorporates a physics-informed mask-to-mass module that converts predicted masks into class-specific mass estimates using geometric calibration and material priors. Applied to a representative stream of 1253 detected objects, the system estimated a total plastic mass of 15.48 ± 1.08 kg, corresponding to a theoretical H2 potential of 0.41 ± 0.04 kg and a greenhouse-gas avoidance of 34.57 ± 4.15 kg CO2e. Overall, the proposed framework extends waste-scene understanding beyond vision-level assessment toward physically grounded, data-driven decision support for smart material recovery systems. Full article
(This article belongs to the Special Issue Machine Learning: Innovation, Implementation, and Impact)
10 pages, 298 KB  
Article
Machine-Learning Models Outperform Clinicians in Predicting Postnatal Growth Failure Among Very Low Birth Weight Infants
by Joohee Lim, Sook Hyun Park, Teahyen Cha, So Jin Yoon, Jung Ho Han, Jeong Eun Shin, In Gyu Song, Soon Min Lee, Ho Seon Eun and Min Soo Park
Diagnostics 2026, 16(9), 1282; https://doi.org/10.3390/diagnostics16091282 - 24 Apr 2026
Abstract
Background/Objectives: Early detection of postnatal growth failure (PGF) is essential for optimizing nutritional management in preterm infants, as PGF is associated with adverse neurodevelopmental outcomes. Early prediction remains difficult because postnatal growth is influenced by multiple clinical factors including gestation age, birth [...] Read more.
Background/Objectives: Early detection of postnatal growth failure (PGF) is essential for optimizing nutritional management in preterm infants, as PGF is associated with adverse neurodevelopmental outcomes. Early prediction remains difficult because postnatal growth is influenced by multiple clinical factors including gestation age, birth weight, nutritional status, and comorbidities. Machine-learning approaches have been proposed to predict complex neonatal outcomes. This study compared the predictive performance of neonatologists with that of a machine-learning model for predicting PGF. Methods: PGF was defined as a decrease in weight z-score greater than 1.28 at discharge compared with birth. A machine-learning model based on extreme gradient boosting (XGBoost) was trained using a dataset of 7954 very low birth weight (VLBW) infants. Nine neonatologists independently assessed 100 clinical cases through a questionnaire-based evaluation, including 50 patients with PGF. Predictive performance was evaluated using seven metrics: area under the receiver operating characteristic curve (AUROC), accuracy, error rate, positive predictive value (PPV), sensitivity, specificity, and F1 score. Results: The neonatologists had a median of 5 years (range: 4–10 years) of clinical experience. The median prediction score among the neonatologists was 52/100 (range, 44–60), whereas the XGBoost model achieved 79/100. The XGBoost model achieved an AUROC of 0.79, accuracy of 0.79, error rate of 0.21, sensitivity of 0.82, and an F1 score of 0.80, demonstrating superior overall performance compared to the neonatologists. In addition, the XGBoost model had a lower error rate than the neonatologists (0.21 vs. 0.49), whereas specificity (0.76 vs. 0.86) and PPV (0.77 vs. 0.53) did not differ significantly. Conclusions: The machine-learning model demonstrated superior or comparable predictive performance to that of neonatologists in detecting PGF. Machine-learning-based prediction models may support early risk stratification and targeted nutritional management in VLBW infants. Full article
(This article belongs to the Special Issue Artificial Intelligence in Clinical Decision Support—2nd Edition)
Show Figures

Figure 1

Back to TopTop