Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (12,901)

Search Parameters:
Keywords = image feature extraction

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 773 KB  
Article
A Radiomics-Based Machine Learning Model for Predicting Pneumonitis During Durvalumab Treatment in Locally Advanced NSCLC
by Takeshi Masuda, Daisuke Kawahara, Wakako Daido, Nobuki Imano, Naoko Matsumoto, Kosuke Hamai, Yasuo Iwamoto, Yusuke Takayama, Sayaka Ueno, Masahiko Sumii, Hiroyasu Shoda, Nobuhisa Ishikawa, Masahiro Yamasaki, Yoshifumi Nishimura, Shigeo Kawase, Naoki Shiota, Yoshikazu Awaya, Soichi Kitaguchi, Yuji Murakami, Yasushi Nagata and Noboru Hattoriadd Show full author list remove Hide full author list
AI 2026, 7(1), 32; https://doi.org/10.3390/ai7010032 (registering DOI) - 16 Jan 2026
Abstract
Introduction: Pneumonitis represents one of the clinically significant adverse events observed in patients with non-small-cell lung cancer (NSCLC) who receive durvalumab as consolidation therapy after chemoradiotherapy (CRT). Although clinical factors such as radiation dose (e.g., V20) and interstitial lung abnormalities (ILAs) have been [...] Read more.
Introduction: Pneumonitis represents one of the clinically significant adverse events observed in patients with non-small-cell lung cancer (NSCLC) who receive durvalumab as consolidation therapy after chemoradiotherapy (CRT). Although clinical factors such as radiation dose (e.g., V20) and interstitial lung abnormalities (ILAs) have been reported as risk predictors, accurate and objective prognostication remains difficult. This study aimed to develop a radiomics-based machine learning model to predict grade ≥ 2 pneumonitis. Methods: This retrospective study included patients with unresectable NSCLC who received CRT followed by durvalumab. Radiomic features, including first-order and texture and shape-based features with wavelet transformation were extracted from whole-lung regions on pre-durvalumab computed tomography (CT) images. Machine learning models, support vector machines, k-nearest neighbor, neural networks, and naïve Bayes classifiers were developed and evaluated using a testing cohort. Model performance was assessed using five-fold cross-validation. Conventional predictors, including V20 and ILAs, were also assessed using logistic regression and receiver operating characteristic analysis. Results: Among 123 patients, 44 (35.8%) developed grade ≥ 2 pneumonitis. The best-performing model, a support vector machine, achieved an AUC of 0.88 and accuracy of 0.81, the conventional model showed lower performance with an AUC of 0.71 and accuracy of 0.64. Conclusions: Radiomics-based machine learning demonstrated superior performance over clinical parameters in predicting pneumonitis. This approach may enable individualized risk stratification and support early intervention in patients with NSCLC. Full article
(This article belongs to the Section Medical & Healthcare AI)
Show Figures

Figure 1

23 pages, 8155 KB  
Article
MRMAFusion: A Multi-Scale Restormer and Multi-Dimensional Attention Network for Infrared and Visible Image Fusion
by Liang Dong, Guiling Sun, Haicheng Zhang and Wenxuan Luo
Appl. Sci. 2026, 16(2), 946; https://doi.org/10.3390/app16020946 - 16 Jan 2026
Abstract
Infrared and visible image fusion improves the visual representation of scenes. Current deep learning-based fusion methods typically rely on either convolution operations for local feature extraction or Transformers for global feature extraction, often neglecting the contribution of multi-scale features to fusion performance. To [...] Read more.
Infrared and visible image fusion improves the visual representation of scenes. Current deep learning-based fusion methods typically rely on either convolution operations for local feature extraction or Transformers for global feature extraction, often neglecting the contribution of multi-scale features to fusion performance. To address this limitation, we propose MRMAFusion, a nested connection model that relies on the multi-scale restoration-Transformer (Restormer) and multi-dimensional attention. We construct an encoder–decoder architecture on UNet++ network with multi-scale local and global feature extraction using convolution blocks and Restormer. Restormer can provide global dependency and more comprehensive attention to texture details of the target region along the vertical dimension, compared to extracting features by convolution operations. Along the horizontal dimension, we enhance MRMAFusion’s multi-scale feature extraction and reconstruction capability by incorporating multi-dimensional attention into the encoder’s convolutional blocks. We perform extensive experiments on the public datasets TNO, NIR and RoadScene and compare with other state-of-the-art methods for both objective and subjective evaluation. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
38 pages, 16828 KB  
Article
Hybrid ConvNeXtV2–ViT Architecture with Ontology-Driven Explainability and Out-of-Distribution Awareness for Transparent Chest X-Ray Diagnosis
by Naif Almughamisi, Gibrael Abosamra, Adnan Albar and Mostafa Saleh
Diagnostics 2026, 16(2), 294; https://doi.org/10.3390/diagnostics16020294 - 16 Jan 2026
Abstract
Background: Chest X-ray (CXR) is widely used for the assessment of thoracic diseases, yet automated multi-label interpretation remains challenging due to subtle visual patterns, overlapping anatomical structures, and frequent co-occurrence of abnormalities. While recent deep learning models have shown strong performance, limitations in [...] Read more.
Background: Chest X-ray (CXR) is widely used for the assessment of thoracic diseases, yet automated multi-label interpretation remains challenging due to subtle visual patterns, overlapping anatomical structures, and frequent co-occurrence of abnormalities. While recent deep learning models have shown strong performance, limitations in interpretability, anatomical awareness, and robustness continue to hinder their clinical adoption. Methods: The proposed framework employs a hybrid ConvNeXtV2–Vision Transformer (ViT) architecture that combines convolutional feature extraction for capturing fine-grained local patterns with transformer-based global reasoning to model long-range contextual dependencies. The model is trained exclusively using image-level annotations. In addition to classification, three complementary post hoc components are integrated to enhance model trust and interpretability. A segmentation-aware Gradient-weighted class activation mapping (Grad-CAM) module leverages CheXmask lung and heart segmentations to highlight anatomically relevant regions and quantify predictive evidence inside and outside the lungs. An ontology-driven neuro-symbolic reasoning layer translates Grad-CAM activations into structured, rule-based explanations aligned with clinical concepts such as “basal effusion” and “enlarged cardiac silhouette”. Furthermore, a lightweight out-of-distribution (OOD) detection module based on confidence scores, energy scores, and Mahalanobis distance scores is employed to identify inputs that deviate from the training distribution. Results: On the VinBigData test set, the model achieved a macro-AUROC of 0.9525 and a Micro AUROC of 0.9777 when trained solely with image-level annotations. External evaluation further demonstrated strong generalisation, yielding macro-AUROC scores of 0.9106 on NIH ChestXray14 and 0.8487 on CheXpert (frontal views). Both Grad-CAM visualisations and ontology-based reasoning remained coherent on unseen data, while the OOD module successfully flagged non-thoracic images. Conclusions: Overall, the proposed approach demonstrates that hybrid convolutional neural network (CNN)–vision transformer (ViT) architectures, combined with anatomy-aware explainability and symbolic reasoning, can support automated chest X-ray diagnosis in a manner that is accurate, transparent, and safety-aware. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
18 pages, 2868 KB  
Article
AdaDenseNet-LUC: Adaptive Attention DenseNet for Laryngeal Ultrasound Image Classification
by Cunyuan Luan and Huabo Liu
BioMedInformatics 2026, 6(1), 5; https://doi.org/10.3390/biomedinformatics6010005 - 16 Jan 2026
Abstract
Evaluating the difficulty of endotracheal intubation during pre-anesthesia assessment has consistently posed a challenge for clinicians. Accurate prediction of intubation difficulty is crucial for subsequent treatment planning. However, existing diagnostic methods often suffer from low accuracy. To tackle this issue, this study presented [...] Read more.
Evaluating the difficulty of endotracheal intubation during pre-anesthesia assessment has consistently posed a challenge for clinicians. Accurate prediction of intubation difficulty is crucial for subsequent treatment planning. However, existing diagnostic methods often suffer from low accuracy. To tackle this issue, this study presented an automated airway classification method utilizing Convolutional Neural Networks (CNNs). We proposed Adaptive Attention DenseNet for Laryngeal Ultrasound Classification (AdaDenseNet-LUC), a network architecture that enhances classification performance by integrating an adaptive attention mechanism into DenseNet (Dense Convolutional Network), enabling the extraction of deep features that aid in difficult airway classification. This model associates laryngeal ultrasound images with actual intubation difficulty, providing healthcare professionals with scientific evidence to help improve the accuracy of clinical decision-making. Experiments were performed on a dataset of 1391 ultrasound images, utilizing 5-fold cross-validation to assess the model’s performance. The experimental results show that the proposed method achieves a classification accuracy of 87.41%, sensitivity of 86.05%, specificity of 88.59%, F1 score of 0.8638, and AUC of 0.94. Grad-CAM visualization techniques indicate that the model’s attention is attention to the tracheal region. The results demonstrate that the proposed method outperforms current approaches, delivering objective and accurate airway classification outcomes, which serve as a valuable reference for evaluating the difficulty of endotracheal intubation and providing guidance for clinicians. Full article
Show Figures

Figure 1

24 pages, 43005 KB  
Article
Accurate Estimation of Spring Maize Aboveground Biomass in Arid Regions Based on Integrated UAV Remote Sensing Feature Selection
by Fengxiu Li, Yanzhao Guo, Yingjie Ma, Ning Lv, Zhijian Gao, Guodong Wang, Zhitao Zhang, Lei Shi and Chongqi Zhao
Agronomy 2026, 16(2), 219; https://doi.org/10.3390/agronomy16020219 - 16 Jan 2026
Abstract
Maize is one of the top three crops globally, ranking only behind rice and wheat, making it an important crop of interest. Aboveground biomass is a key indicator for assessing maize growth and its yield potential. This study developed an efficient and stable [...] Read more.
Maize is one of the top three crops globally, ranking only behind rice and wheat, making it an important crop of interest. Aboveground biomass is a key indicator for assessing maize growth and its yield potential. This study developed an efficient and stable biomass prediction model to estimate the aboveground biomass (AGB) of spring maize (Zea mays L.) under subsurface drip irrigation in arid regions, based on UAV multispectral remote sensing and machine learning techniques. Focusing on typical subsurface drip-irrigated spring maize in arid Xinjiang, multispectral images and field-measured AGB data were collected from 96 sample points (selected via stratified random sampling across 24 plots) over four key phenological stages in 2024 and 2025. Sixteen vegetation indices were calculated and 40 texture features were extracted using the gray-level co-occurrence matrix method, while an integrated feature-selection strategy combining Elastic Net and Random Forest was employed to effectively screen key predictor variables. Based on the selected features, six machine learning models were constructed, including Elastic Net Regression (ENR), Gradient Boosting Decision Trees (GBDT), Gaussian Process Regression (GPR), Partial Least Squares Regression (PLSR), Random Forest (RF), and Extreme Gradient Boosting (XGB). Results showed that the fused feature set comprised four vegetation indices (GRDVI, RERVI, GRVI, NDVI) and five texture features (R_Corr, NIR_Mean, NIR_Vari, B_Mean, B_Corr), thereby retaining red-edge and visible-light texture information highly sensitive to AGB. The GPR model based on the fused features exhibited the best performance (test set R2 = 0.852, RMSE = 2890.74 kg ha−1, MAE = 1676.70 kg ha−1), demonstrating high fitting accuracy and stable predictive ability across both the training and test sets. Spatial inversions over the two growing seasons of 2024 and 2025, derived from the fused-feature GPR optimal model at four key phenological stages, revealed pronounced spatiotemporal heterogeneity and stage-dependent dynamics of spring maize AGB: the biomass accumulates rapidly from jointing to grain filling, slows thereafter, and peaks at maturity. At a constant planting density, AGB increased markedly with nitrogen inputs from N0 to N3 (420 kg N ha−1), with the high-nitrogen N3 treatment producing the greatest biomass; this successfully captured the regulatory effect of the nitrogen gradient on maize growth, provided reliable data for variable-rate fertilization, and is highly relevant for optimizing water–fertilizer coordination in subsurface drip irrigation systems. Future research may extend this integrated feature selection and modeling framework to monitor the growth and estimate the yield of other crops, such as rice and cotton, thereby validating its generalizability and robustness in diverse agricultural scenarios. Full article
Show Figures

Figure 1

15 pages, 2108 KB  
Article
[18F]FDG PET/MRI in Endometrial Cancer: Prospective Evaluation of Preoperative Staging, Molecular Characterization and Prognostic Assessment
by Carolina Bezzi, Gabriele Ironi, Tommaso Russo, Giorgio Candotti, Federico Fallanca, Carlotta Sabini, Ana Maria Samanes Gajate, Samuele Ghezzo, Alice Bergamini, Miriam Sant’Angelo, Luca Bocciolone, Giorgio Brembilla, Paola Scifo, GianLuca Taccagni, Onofrio Antonio Catalano, Giorgia Mangili, Massimo Candiani, Francesco De Cobelli, Arturo Chiti, Paola Mapelli and Maria Picchioadd Show full author list remove Hide full author list
Cancers 2026, 18(2), 280; https://doi.org/10.3390/cancers18020280 - 16 Jan 2026
Abstract
Background/Objectives: Early and accurate characterization of endometrial cancer (EC) is crucial for patient management, but current imaging modalities lack in diagnostic accuracy and ability to assess molecular profiles. The aim of this study is to evaluate hybrid [18F]FDG PET/MRI’s diagnostic accuracy [...] Read more.
Background/Objectives: Early and accurate characterization of endometrial cancer (EC) is crucial for patient management, but current imaging modalities lack in diagnostic accuracy and ability to assess molecular profiles. The aim of this study is to evaluate hybrid [18F]FDG PET/MRI’s diagnostic accuracy in EC staging and role in predicting tumor aggressiveness, molecular characterization, and recurrence. Methods: A prospective study (ClinicalTrials.gov, ID:NCT04212910) evaluating EC patients undergoing [18F]FDG PET/MRI before surgery (2018–2024). Histology, immunohistochemistry, and patients’ follow-up (mean FU time: 3.13y) were used as the reference standard. [18F]FDG PET/MRI, PET only, and MRI only were independently reviewed to assess the diagnostic accuracy (ACC), sensitivity (SN), specificity (SP), and positive/negative predictive value (PPV, NPV). Imaging parameters were extracted from [18F]FDG PET and pcT1w, T2w, DWI, and DCE MRI. Spearman’s correlations, Fisher’s exact test, ROC-AUC analysis, Kaplan–Meier survival curves, log-rank tests and Cox proportional hazards models were applied. Results: Eighty participants with primary EC (median age 63 ± 12 years) were enrolled, with 17% showing LN involvement. [18F]FDG PET/MRI provided ACC = 98.75%, SN = 98.75%, and PPV = 100% for primary tumor detection, and ACC = 92.41%, SN = 84.62%, SP = 93.94%, PPV = 73.33%, and NPV = 96.88% for LN detection. PET/MRI parameters predicted LN involvement (AUC = 79.49%), deep myometrial invasion (79.78%), lymphovascular space invasion (82.00%), p53abn (71.47%), MMRd (74.51%), relapse (82.00%), and postoperative administration of adjuvant therapy (79.64%). Patients with a tumor cranio-caudal diameter ≥ 43 mm and MTV ≥ 13.5 cm3 showed increased probabilities of recurrence (p < 0.001). Conclusions: [18F]FDG PET/MR showed exceptional accuracy in EC primary tumor and LN detection. Derived parameters demonstrated potential ability in defining features of aggressiveness, molecular alterations, and tumor recurrence. Full article
(This article belongs to the Special Issue Molecular Biology, Diagnosis and Management of Endometrial Cancer)
Show Figures

Figure 1

31 pages, 1485 KB  
Article
Explainable Multi-Modal Medical Image Analysis Through Dual-Stream Multi-Feature Fusion and Class-Specific Selection
by Naeem Ullah, Ivanoe De Falco and Giovanna Sannino
AI 2026, 7(1), 30; https://doi.org/10.3390/ai7010030 - 16 Jan 2026
Abstract
Effective and transparent medical diagnosis relies on accurate and interpretable classification of medical images across multiple modalities. This paper introduces an explainable multi-modal image analysis framework based on a dual-stream architecture that fuses handcrafted descriptors with deep features extracted from a custom MobileNet. [...] Read more.
Effective and transparent medical diagnosis relies on accurate and interpretable classification of medical images across multiple modalities. This paper introduces an explainable multi-modal image analysis framework based on a dual-stream architecture that fuses handcrafted descriptors with deep features extracted from a custom MobileNet. Handcrafted descriptors include frequency-domain and texture features, while deep features are summarized using 26 statistical metrics to enhance interpretability. In the fusion stage, complementary features are combined at both the feature and decision levels. Decision-level integration combines calibrated soft voting, weighted voting, and stacking ensembles with optimized classifiers, including decision trees, random forests, gradient boosting, and logistic regression. To further refine performance, a hybrid class-specific feature selection strategy is proposed, combining mutual information, recursive elimination, and random forest importance to select the most discriminative features for each class. This hybrid selection approach eliminates redundancy, improves computational efficiency, and ensures robust classification. Explainability is provided through Local Interpretable Model-Agnostic Explanations, which offer transparent details about the ensemble model’s predictions and link influential handcrafted features to clinically meaningful image characteristics. The framework is validated on three benchmark datasets, i.e., BTTypes (brain MRI), Ultrasound Breast Images, and ACRIMA Retinal Fundus Images, demonstrating generalizability across modalities (MRI, ultrasound, retinal fundus) and disease categories (brain tumor, breast cancer, glaucoma). Full article
(This article belongs to the Special Issue Digital Health: AI-Driven Personalized Healthcare and Applications)
Show Figures

Figure 1

18 pages, 1144 KB  
Article
Hypersector-Based Method for Real-Time Classification of Wind Turbine Blade Defects
by Lesia Dubchak, Bohdan Rusyn, Carsten Wolff, Tomasz Ciszewski, Anatoliy Sachenko and Yevgeniy Bodyanskiy
Energies 2026, 19(2), 442; https://doi.org/10.3390/en19020442 - 16 Jan 2026
Abstract
This paper presents a novel hypersector-based method with Fuzzy Learning Vector Quantization (FLVQ) for the real-time classification of wind turbine blade defects using data acquired by unmanned aerial vehicles (UAVs). Unlike conventional prototype-based FLVQ approaches that rely on Euclidean distance in the feature [...] Read more.
This paper presents a novel hypersector-based method with Fuzzy Learning Vector Quantization (FLVQ) for the real-time classification of wind turbine blade defects using data acquired by unmanned aerial vehicles (UAVs). Unlike conventional prototype-based FLVQ approaches that rely on Euclidean distance in the feature space, the proposed method models each defect class as a hypersector on an n-dimensional hypersphere, where class boundaries are defined by angular similarity and fuzzy membership transitions. This geometric reinterpretation of FLVQ constitutes the core innovation of the study, enabling improved class separability, robustness to noise, and enhanced interpretability under uncertain operating conditions. Feature vectors extracted via the pre-trained SqueezeNet convolutional network are normalized onto the hypersphere, forming compact directional clusters that serve as the geometric foundation of the FLVQ classifier. A fuzzy softmax membership function and an adaptive prototype-updating mechanism are introduced to handle class overlap and improve learning stability. Experimental validation on a custom dataset of 900 UAV-acquired images achieved 95% classification accuracy on test data and 98.3% on an independent dataset, with an average F1-score of 0.91. Comparative analysis with the classical FLVQ prototype demonstrated superior performance and noise robustness. Owing to its low computational complexity and transparent geometric decision structure, the developed model is well-suited for real-time deployment on UAV embedded systems. Furthermore, the proposed hypersector FLVQ framework is generic and can be extended to other renewable-energy diagnostic tasks, including solar and hydropower asset monitoring, contributing to enhanced energy security and sustainability. Full article
(This article belongs to the Special Issue Modeling, Control and Optimization of Wind Power Systems)
Show Figures

Figure 1

24 pages, 3292 KB  
Article
Comparing Emerging and Hybrid Quantum–Kolmogorov Architectures for Image Classification
by Lelio Campanile, Mariarosaria Castaldo, Stefano Marrone and Fabio Napoli
Computers 2026, 15(1), 65; https://doi.org/10.3390/computers15010065 - 16 Jan 2026
Abstract
The rapid evolution of Artificial Intelligence has enabled significant progress in image classification, with emerging approaches extending traditional deep learning paradigms. This article presents an extended version of a paper originally introduced at ICCSA 2025, providing a broader comparative analysis of classical, spline-based, [...] Read more.
The rapid evolution of Artificial Intelligence has enabled significant progress in image classification, with emerging approaches extending traditional deep learning paradigms. This article presents an extended version of a paper originally introduced at ICCSA 2025, providing a broader comparative analysis of classical, spline-based, and quantum machine learning architectures. The study evaluates Convolutional Neural Networks (CNNs), Kolmogorov–Arnold Networks (KANs), Convolutional KANs (CKANs), and Quantum Convolutional Neural Networks (QCNNs) on the Labeled Faces in the Wild dataset. In addition to these baselines, two novel architectures are introduced: a fully quantum Kolmogorov–Arnold model (F-QKAN) and a hybrid KAN–Quantum network (H-QKAN) that combines spline-based feature extraction with variational quantum classification. Rather than targeting state-of-the-art performance, the evaluation focuses on analyzing the behaviour of these architectures in terms of accuracy, computational efficiency, and interpretability under a unified experimental protocol. Results show that the fully quantum F-QKAN achieves a test accuracy above 80%. The hybrid H-QKAN obtains the best overall performance, exceeding 92% accuracy with rapid convergence and stable training dynamics. Classical CNNs models remain state-of-the-art in terms of predictive performance, whereas CKANs offer a favorable balance between accuracy and efficiency. QCNNs show potential in ideal noise-free settings but are significantly affected by realistic noise conditions, motivating further investigation into hybrid quantum–classical designs. Full article
(This article belongs to the Special Issue Computational Science and Its Applications 2025 (ICCSA 2025))
Show Figures

Figure 1

16 pages, 3916 KB  
Article
Real-Time Detection of Electrohydrodynamic Atomization Modes via a YOLOv8-Based Deep Learning Model
by Xiong Ran, Heming Xu, Xiangfei Wei, Jinxin Wang and Wei-Cheng Yan
Processes 2026, 14(2), 313; https://doi.org/10.3390/pr14020313 - 15 Jan 2026
Abstract
A YOLOv8-based deep learning model was developed to address real-time detection and dynamic regulation needs of the electrohydrodynamic atomization process. An EHDA experimental system was built to obtain images of six typical atomization modes, forming a dataset with 6000 images. After annotation and [...] Read more.
A YOLOv8-based deep learning model was developed to address real-time detection and dynamic regulation needs of the electrohydrodynamic atomization process. An EHDA experimental system was built to obtain images of six typical atomization modes, forming a dataset with 6000 images. After annotation and mosaic augmentation, the dataset served as the training data for the model. The YOLOv8 adopts a “backbone-neck-head” architecture to extract and fuse features, decouple classification and detection, and optimize performance. Experimental results demonstrate that on the test set, the model attains a precision value, recall rate, and mAP50 of 0.995, alongside an mAP50-95 of 0.8. Additionally, its prediction accuracy exceeds 0.99 across all operational modes. Compared with 10 models, it has the best precision and mAP50, as well as low computational complexity, combining high accuracy and lightweight advantages, which can be effectively used for real-time detection of EHDA modes. Full article
17 pages, 1776 KB  
Article
Multi-Scale Adaptive Light Stripe Center Extraction for Line-Structured Light Vision Based Online Wheelset Measurement
by Saisai Liu, Qixin He, Wenjie Fu, Boshi Du and Qibo Feng
Sensors 2026, 26(2), 600; https://doi.org/10.3390/s26020600 - 15 Jan 2026
Abstract
The extraction of the light stripe center is a pivotal step in line-structured light vision measurement. This paper addresses a key challenge in the online measurement of train wheel treads, where the diverse and complex profile characteristics of the tread surface lead to [...] Read more.
The extraction of the light stripe center is a pivotal step in line-structured light vision measurement. This paper addresses a key challenge in the online measurement of train wheel treads, where the diverse and complex profile characteristics of the tread surface lead to uneven gray-level distribution and varying width features in the stripe image, ultimately degrading the accuracy of center extraction. To solve this problem, a region-adaptive multiscale method for light stripe center extraction is proposed. First, potential light stripe regions are identified and enhanced based on the gray-gradient features of the image, enabling precise segmentation. Subsequently, by normalizing the feature responses under Gaussian kernels with different scales, the locally optimal scale parameter (σ) is determined adaptively for each stripe region. Sub-pixel center extraction is then performed using the Hessian matrix corresponding to this optimal σ. Experimental results demonstrate that under on-site conditions featuring uneven wheel surface reflectivity, the proposed method can reliably extract light stripe centers with high stability. It achieves a repeatability of 0.10 mm, with mean measurement errors of 0.12 mm for flange height and 0.10 mm for flange thickness, thereby enhancing both stability and accuracy in industrial measurement environments. The repeatability and reproducibility of the method were further validated through repeated testing of multiple wheels. Full article
(This article belongs to the Special Issue Intelligent Sensors and Signal Processing in Industry)
Show Figures

Figure 1

38 pages, 7681 KB  
Article
A Sequential GAN–CNN–FUZZY Framework for Robust Face Recognition and Attentiveness Analysis in E-Learning
by Chaimaa Khoudda, Yassine El Harrass, Kaoutar Tazi, Salma Azzouzi and Moulay El Hassan Charaf
Appl. Sci. 2026, 16(2), 909; https://doi.org/10.3390/app16020909 - 15 Jan 2026
Abstract
In modern e-learning environments, ensuring both student identity verification and concentration monitoring during online examinations has become increasingly important. This paper introduces a robust sequential framework that integrates Generative Adversarial Networks (GANs), Convolutional Neural Networks (CNNs) and fuzzy logic to achieve reliable face [...] Read more.
In modern e-learning environments, ensuring both student identity verification and concentration monitoring during online examinations has become increasingly important. This paper introduces a robust sequential framework that integrates Generative Adversarial Networks (GANs), Convolutional Neural Networks (CNNs) and fuzzy logic to achieve reliable face recognition and interpretable attentiveness assessment. Images from the Extended Yale B (cropped) dataset are preprocessed through grayscale normalization and resizing, while GANs generate synthetic variations in pose, illumination, and occlusion to enrich the training set and improve generalization. The CNN extracts discriminative facial features for identity recognition, and a fuzzy inference system transforms the CNN’s confidence scores into human-interpretable concentration levels. To stabilize learning and prevent overfitting, the model incorporates dropout regularization, batch normalization, and extensive data augmentation. Comprehensive evaluations using confusion matrices, ROC–AUC, and precision–recall analyses demonstrate an accuracy of 98.42%. The proposed framework offers a scalable and interpretable solution for secure and reliable online exam proctoring. Full article
Show Figures

Figure 1

29 pages, 12605 KB  
Article
YOLOv11n-CGSD: Lightweight Detection of Dairy Cow Body Temperature from Infrared Thermography Images in Complex Barn Environments
by Zhongwei Kang, Hang Song, Hang Xue, Miao Wu, Derui Bao, Chuang Yan, Hang Shi, Jun Hu and Tomas Norton
Agriculture 2026, 16(2), 229; https://doi.org/10.3390/agriculture16020229 - 15 Jan 2026
Abstract
Dairy cow body temperature is a key physiological indicator that reflects metabolic level, immune status, and environmental stress responses, and it has been widely used for early disease recognition. Infrared thermography (IRT), as a non-contact imaging technique capable of remotely acquiring the surface [...] Read more.
Dairy cow body temperature is a key physiological indicator that reflects metabolic level, immune status, and environmental stress responses, and it has been widely used for early disease recognition. Infrared thermography (IRT), as a non-contact imaging technique capable of remotely acquiring the surface radiation temperature distribution of animals, is regarded as a powerful alternative to traditional temperature measurement methods. Under practical cowshed conditions, IRT images of dairy cows are easily affected by complex background interference and generally suffer from low resolution, poor contrast, indistinct boundaries, weak structural perception, and insufficient texture information, which lead to significant degradation in target detection and temperature extraction performance. To address these issues, a lightweight detection model named YOLOv11n-CGSD is proposed for dairy cow IRT images, aiming to improve the accuracy and robustness of region of interest (ROI) detection and body temperature extraction under complex background conditions. At the architectural level, a C3Ghost lightweight module based on the Ghost concept is first constructed to reduce redundant feature extraction while lowering computational cost and enhancing the network capability for preserving fine-grained features during feature propagation. Subsequently, a space-to-depth convolution module is introduced to perform spatial rearrangement of feature maps and achieve channel compression via non-strided convolution, thereby improving the sensitivity of the model to local temperature variations and structural details. Finally, a dynamic sampling mechanism is embedded in the neck of the network, where the upsampling and scale alignment processes are adaptively driven by feature content, enhancing the model response to boundary temperature changes and weak-texture regions. Experimental results indicate that the YOLOv11n-CGSD model can effectively shift attention from irrelevant background regions to ROI contour boundaries and increase attention coverage within the ROI. Under complex IRT conditions, the model achieves P, R, and mAP50 values of 89.11%, 86.80%, and 91.94%, which represent improvements of 3.11%, 5.14%, and 4.08%, respectively, compared with the baseline model. Using Tmax as the temperature extraction parameter, the maximum error (Max. Error) and mean error (MAE. Error) in the lower udder region are reduced by 33.3% and 25.7%, respectively, while in the around the anus region, the Max. Error and MAE. Error are reduced by 87.5% and 95.0%, respectively. These findings demonstrate that, under complex backgrounds and low-quality IRT imaging conditions, the proposed model achieves lightweight and high-performance detection for both lower udder (LU) and around the anus (AA) regions and provides a methodological reference and technical support for non-contact body temperature measurement of dairy cows in practical cowshed production environments. Full article
(This article belongs to the Section Farm Animal Production)
Show Figures

Figure 1

24 pages, 3045 KB  
Article
A Dual Stream Deep Learning Framework for Alzheimer’s Disease Detection Using MRI Sonification
by Nadia A. Mohsin and Mohammed H. Abdul Ameer
J. Imaging 2026, 12(1), 46; https://doi.org/10.3390/jimaging12010046 - 15 Jan 2026
Abstract
Alzheimer’s Disease (AD) is an advanced brain illness that affects millions of individuals across the world. It causes gradual damage to the brain cells, leading to memory loss and cognitive dysfunction. Although Magnetic Resonance Imaging (MRI) is widely used in AD diagnosis, the [...] Read more.
Alzheimer’s Disease (AD) is an advanced brain illness that affects millions of individuals across the world. It causes gradual damage to the brain cells, leading to memory loss and cognitive dysfunction. Although Magnetic Resonance Imaging (MRI) is widely used in AD diagnosis, the existing studies rely solely on the visual representations, leaving alternative features unexplored. The objective of this study is to explore whether MRI sonification can provide complementary diagnostic information when combined with conventional image-based methods. In this study, we propose a novel dual-stream multimodal framework that integrates 2D MRI slices with their corresponding audio representations. MRI images are transformed into audio signals using a multi-scale, multi-orientation Gabor filtering, followed by a Hilbert space-filling curve to preserve spatial locality. The image and sound modalities are processed using a lightweight CNN and YAMNet, respectively, then fused via logistic regression. The experimental results of the multimodal achieved the highest accuracy in distinguishing AD from Cognitively Normal (CN) subjects at 98.2%, 94% for AD vs. Mild Cognitive Impairment (MCI), and 93.2% for MCI vs. CN. This work provides a new perspective and highlights the potential of audio transformation of imaging data for feature extraction and classification. Full article
(This article belongs to the Section AI in Imaging)
23 pages, 5052 KB  
Article
Exploratory Study on Hybrid Systems Performance: A First Approach to Hybrid ML Models in Breast Cancer Classification
by Francisco J. Rojas-Pérez, José R. Conde-Sánchez, Alejandra Morlett-Paredes, Fernando Moreno-Barbosa, Julio C. Ramos-Fernández, José Luna-Muñoz, Genaro Vargas-Hernández, Blanca E. Jaramillo-Loranca, Juan M. Xicotencatl-Pérez and Eucario G. Pérez-Pérez
AI 2026, 7(1), 29; https://doi.org/10.3390/ai7010029 - 15 Jan 2026
Abstract
The classification of breast cancer using machine learning techniques has become a critical tool in modern medical diagnostics. This study analyzes the performance of hybrid models that combine traditional machine learning algorithms (TMLAs) with a convolutional neural network (CNN)-based VGG16 model for feature [...] Read more.
The classification of breast cancer using machine learning techniques has become a critical tool in modern medical diagnostics. This study analyzes the performance of hybrid models that combine traditional machine learning algorithms (TMLAs) with a convolutional neural network (CNN)-based VGG16 model for feature extraction to improve accuracy for classifying eight breast cancer subtypes (BCS). The methodology consists of three steps. First, image preprocessing is performed on the BreakHis dataset at 400× magnification, which contains 1820 histopathological images classified into eight BCS. Second, the CNN VGG16 is modified to function as a feature extractor that converts images into representative vectors. These vectors constitute the training set for TMLAs, such as Random Forest (RF), Support Vector Machine (SVM), K-Nearest Neighbors (KNN), and Naive Bayes (NB), leveraging VGG16’s ability to capture relevant features. Third, k-fold cross-validation is applied to evaluate the model’s performance by averaging the metrics obtained across all folds. The results reveal that hybrid models leveraging a CNN-based VGG16 model for feature extraction, followed by TMLAs, achieve accuracy outstanding experimental accuracy. The KNN-based hybrid model stood out with a precision of 0.97, accuracy of 0.96, sensitivity of 0.96, specificity of 0.99, F1-score of 0.96, and ROC-AUC of 0.97. These findings suggest that, with an appropriate methodology, hybrid models based on TMLA have strong potential in classification tasks, offering a balance between performance and predictive capability. Full article
Show Figures

Figure 1

Back to TopTop