Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (951)

Search Parameters:
Keywords = improved VGG16

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 4990 KB  
Article
Curvature Radius Measurement Based on Interferogram Analysis and Deep Learning Model
by Yan-Yi Li, Chuen-Lin Tien, Hsi-Fu Shih, Han-Yen Tu and Chih-Cheng Chen
Photonics 2026, 13(5), 416; https://doi.org/10.3390/photonics13050416 - 24 Apr 2026
Abstract
Accurate estimation of curvature radius from interference fringes is critical in optical metrology and precision manufacturing. Conventional interferogram analytical approaches often require manual intervention and are sensitive to fringe variations related to noise and environmental vibrations. To address these limitations, we combine an [...] Read more.
Accurate estimation of curvature radius from interference fringes is critical in optical metrology and precision manufacturing. Conventional interferogram analytical approaches often require manual intervention and are sensitive to fringe variations related to noise and environmental vibrations. To address these limitations, we combine an improved Twyman–Green interferometer with different artificial intelligence (AI) deep learning models and utilize a self-developed MATLAB analysis program to propose a non-destructive and rapid measurement system for optical coating substrates. The proposed AI-assisted Twyman–Green interferometric system differs fundamentally from conventional wavefront sensing techniques in both principle and implementation. This paper utilizes the Twyman–Green interferometer to generate interference fringe datasets on B270 glass and sapphire substrates, and employs convolutional neural network (CNN), ResNet-18, and VGG-16 models for training and evaluation. The proposed method integrates image enhancement, fringe pattern clustering, and analysis and validation based on fast Fourier transform (FFT). Experimental results show that ResNet-18 outperforms other models, with a mean absolute percentage error of 5.44% on sapphire substrates and 3.40% on B270 glass substrates. These findings highlight the effectiveness and robustness of deep learning models, especially residual networks, in automatic ROC prediction for optical measurement applications. Full article
Show Figures

Figure 1

18 pages, 9518 KB  
Article
A Multi-Scale Deep Network for Aircraft Wake Vortex Recognition Using Lidar Radial Velocity Fields
by Xuan Wang, Shangjun Li, Xiqiao Dai, Weijun Pan and Yuanfei Leng
Appl. Sci. 2026, 16(9), 4121; https://doi.org/10.3390/app16094121 - 23 Apr 2026
Abstract
Aircraft wake vortices pose significant threats to following aircraft during takeoff and landing phases. Coherent Doppler lidar provides an effective remote sensing technique for wake vortex monitoring through radial velocity measurements. However, reliable identification of wake vortices from lidar observations remains challenging due [...] Read more.
Aircraft wake vortices pose significant threats to following aircraft during takeoff and landing phases. Coherent Doppler lidar provides an effective remote sensing technique for wake vortex monitoring through radial velocity measurements. However, reliable identification of wake vortices from lidar observations remains challenging due to noise and the complex multi-scale evolution of vortex structures. In this study, we propose a physics-guided multi-scale deep network (HMNet) for aircraft wake vortex identification. First, we propose a denoising module (DE) to suppress noise in radial velocity fields. Subsequently, we design a hybrid multi-scale backbone network containing a hybrid multi-scale feature extraction module (HMFE) to capture vortex structures at different spatial scales. Furthermore, we propose a feature gradient guidance module (FGGM) to incorporate physically meaningful gradient cues and enhance vortex-sensitive features. HMNet is evaluated and tested on 1401 radial velocity field data samples collected on the runway at Shenzhen Bao’an Airport. The experimental results show that HMNet achieves 97.15% accuracy, 95.83% recall, and 96.84% F1 score. Compared with the baseline VGG16 and Random Forest, HMNet improves accuracy by 6.18% and 11.88%, respectively. These results demonstrate that HMNet provides an effective solution for lidar-based wake vortex identification and can support the development of intelligent air traffic management. Full article
36 pages, 5739 KB  
Article
Multi-Scale Atrous Feature Fusion Based on a VGG19-UNet Encoder for Brain Tumor Segmentation
by Shoffan Saifullah and Rafał Dreżewski
Appl. Sci. 2026, 16(8), 3971; https://doi.org/10.3390/app16083971 - 19 Apr 2026
Viewed by 136
Abstract
Accurate brain tumor segmentation from magnetic resonance imaging (MRI) remains challenging due to heterogeneous tumor morphology, intensity variability, and multi-scale structural complexity. This study proposes a DeepLabV3+-based segmentation framework integrating a VGG19-UNet encoder, Atrous Spatial Pyramid Pooling (ASPP), and low-level feature refinement to [...] Read more.
Accurate brain tumor segmentation from magnetic resonance imaging (MRI) remains challenging due to heterogeneous tumor morphology, intensity variability, and multi-scale structural complexity. This study proposes a DeepLabV3+-based segmentation framework integrating a VGG19-UNet encoder, Atrous Spatial Pyramid Pooling (ASPP), and low-level feature refinement to simultaneously capture hierarchical semantics and boundary-sensitive spatial details. The architecture enhances receptive field coverage without additional downsampling while preserving fine-grained contour information during reconstruction. Extensive evaluation was conducted on the Figshare Brain Tumor Segmentation (FBTS) dataset and the BraTS 2021 and BraTS 2018 benchmarks, focusing on Whole Tumor segmentation across multiple MRI modalities and tumor grades. Under five-fold cross-validation, the proposed model achieved a mean Dice Similarity Coefficient of 0.9717 and Jaccard Index of 0.9456 on FBTS, with stable and competitive performance across FLAIR, T1, T2, and T1CE modalities in both HGG and LGG cases. Boundary-level analysis further confirmed controlled Hausdorff Distance and low Average Symmetric Surface Distance. Statistical validation and ablation analysis demonstrate consistent improvements over baseline U-Net configurations. The proposed framework provides a robust and computationally efficient solution for automated brain tumor segmentation across heterogeneous datasets. Full article
(This article belongs to the Special Issue Research on Artificial Intelligence in Healthcare)
Show Figures

Figure 1

35 pages, 5529 KB  
Article
Occasion-Based Clothing Classification Using Vision Transformer and Traditional Machine Learning Models
by Hanaa Alzahrani, Maram Almotairi and Arwa Basbrain
Computers 2026, 15(4), 249; https://doi.org/10.3390/computers15040249 - 17 Apr 2026
Viewed by 254
Abstract
Clothing classification by occasion is an important area in computer vision and artificial intelligence (AI). This task is particularly challenging because of the subtle visual similarities among clothing categories such as formal, party, and casual attire. Variations in color, fabric, patterns, and lighting [...] Read more.
Clothing classification by occasion is an important area in computer vision and artificial intelligence (AI). This task is particularly challenging because of the subtle visual similarities among clothing categories such as formal, party, and casual attire. Variations in color, fabric, patterns, and lighting further increase the complexity of this task. To address this challenge, we used the Fashionpedia dataset to create a balanced subset of 15,000 images. Specifically, we adopted two different methods for labeling these images: automated classification, which relies on category identifications (IDs) and components, and manual labeling performed by human annotators. We then implemented our preprocessing pipeline, which includes several steps: resizing, image normalization, background removal using segmentation masks, and class balancing. We benchmarked traditional models, including artificial neural networks (ANNs), support vector machines (SVMs), and k-nearest neighbors (KNNs), which use a histogram of oriented gradient (HOG) features, as well as deep learning models such as convolutional neural networks (CNNs), the Visual Geometry Group 16 (VGG16) model utilizing transfer learning, and the vision transformer (ViT) model, all evaluated using identical data splits and preprocessing procedures. The traditional models achieved moderate accuracy, ranging from 54% to 66%. In contrast, the ViT model achieved an accuracy of 81.78% with automated classification and 98.09% with manual labeling. This indicates that a higher label accuracy, along with the preprocessing steps used, significantly enhances the performance. Together, these factors improve the effectiveness of ViT in context-aware apparel classification and establish a reliable baseline for future research. Full article
(This article belongs to the Special Issue Machine Learning: Innovation, Implementation, and Impact)
Show Figures

Figure 1

24 pages, 13348 KB  
Article
Morphological Convolutional Neural Network for Efficient Facial Expression Recognition
by Robert, Sarifuddin Madenda, Suryadi Harmanto, Michel Paindavoine and Dina Indarti
J. Imaging 2026, 12(4), 171; https://doi.org/10.3390/jimaging12040171 - 15 Apr 2026
Viewed by 261
Abstract
This study proposes a morphological convolutional neural network (MCNN) architecture that integrates morphological operations with CNN layers for facial expression recognition (FER). Conventional CNN-based FER models primarily rely on appearance features and may be sensitive to illumination and demographic variations. This work investigates [...] Read more.
This study proposes a morphological convolutional neural network (MCNN) architecture that integrates morphological operations with CNN layers for facial expression recognition (FER). Conventional CNN-based FER models primarily rely on appearance features and may be sensitive to illumination and demographic variations. This work investigates whether morphological structural representations provide complementary information to convolutional features. A multi-source and multi-ethnic FER dataset was constructed by combining CK+, JAFFE, KDEF, TFEID, and a newly collected Indonesian Facial Expression dataset, resulting in 3684 images from 326 subjects across seven expression classes. Subject-independent data splitting with 10-fold cross-validation was applied to ensure reliable evaluation. Experimental results show that the proposed MCNN1 model achieves an average accuracy of 88.16%, while the best MCNN2 variant achieves 88.7%, demonstrating competitive performance compared to MobileNetV2 (88.27%), VGG19 (87.58%), and the morphological baseline MNN (50.73%). The proposed model also demonstrates improved computational efficiency, achieving lower inference latency (21%) and reduced GPU memory usage (64%) compared to baseline models. These results indicate that integrating morphological representations into convolutional architectures provides a modest but consistent improvement in FER performance while enhancing generalization and efficiency under heterogeneous data conditions. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

20 pages, 4643 KB  
Article
Deep Learning-Assisted Early Detection of Skin Cancer from Dermoscopic Images in Underserved Clinical Settings
by Anchal Kumari, Punam Rattan, Anand Kumar Shukla, Sita Rani, Aman Kataria, Hong Min and Taeho Kim
Bioengineering 2026, 13(4), 456; https://doi.org/10.3390/bioengineering13040456 - 13 Apr 2026
Viewed by 410
Abstract
Skin cancer is caused by aberrant cells that proliferate uncontrollably after unrepaired DNA damage results in mutations in the epidermis. The majority of skin cancer is caused by high UV exposure from the sun, tanning beds, or sunlamps. Due to sociocultural hurdles, limited [...] Read more.
Skin cancer is caused by aberrant cells that proliferate uncontrollably after unrepaired DNA damage results in mutations in the epidermis. The majority of skin cancer is caused by high UV exposure from the sun, tanning beds, or sunlamps. Due to sociocultural hurdles, limited access to specialized dermatological care, and low public knowledge, many nations, including India, have higher mortality rates and late-stage presentations. The unequal distribution of specialized dermatological treatments, particularly in rural and underdeveloped areas, makes detection and treatment more difficult. For skin cancer, one of the most prevalent malignancies with a high death rate, early detection is crucial. This study gathered 1200 dermoscopic images from two clinics in Himachal Pradesh in order to solve these problems. In order to automatically classify dermoscopic clinical images into melanoma and non-melanoma skin cancer categories, this study compares VGG16 with ResNet-50. Preprocessing, lesion segmentation, and classification are all part of the suggested approach. A collection of 1200 dermoscopic images with clinical annotations was used to improve the models. ResNet-50 outperformed VGG16 in tests, with 93% accuracy and 96% AUC-ROC as opposed to 89% and 94%, respectively. These results emphasize how crucial model selection and preprocessing are to diagnostic performance. Ensemble methods, multi-class classification, explainability integration, and clinical validation will be investigated in order to facilitate the implementation of AI-assisted dermatological diagnostic tools. Full article
Show Figures

Figure 1

22 pages, 2255 KB  
Article
Distributed Stochastic Multi-GPU Hyperparameter Optimization for Transfer Learning-Based Vehicle Detection Under Degraded Visual Conditions
by Zhi-Ren Tsai and Jeffrey J. P. Tsai
Algorithms 2026, 19(4), 296; https://doi.org/10.3390/a19040296 - 10 Apr 2026
Viewed by 180
Abstract
Robust vehicle detection in real-world traffic surveillance remains challenging due to degraded imagery caused by motion blur, adverse weather, and low illumination, which significantly increases detector sensitivity to hyperparameter configurations. This study proposes a “Frugal AI” distributed multi-GPU framework that optimizes hyperparameters via [...] Read more.
Robust vehicle detection in real-world traffic surveillance remains challenging due to degraded imagery caused by motion blur, adverse weather, and low illumination, which significantly increases detector sensitivity to hyperparameter configurations. This study proposes a “Frugal AI” distributed multi-GPU framework that optimizes hyperparameters via a stochastic simplex-based search coupled with five-fold cross-validation. Utilizing three low-cost NVIDIA GTX 1050 Ti GPUs, the framework performs parallel candidate exploration with an asynchronous model-level exchange mechanism to escape local optima without the overhead of gradient synchronization. Seven CNN backbones—VGG16, VGG19, GoogLeNet, MobileNetV2, ResNet18, ResNet50, and ResNet101—were evaluated within YOLOv2 and Faster R-CNN detectors. To address memory constraints (4 GB VRAM), YOLOv2 was selected for extensive benchmarking. Performance was measured using a harmonic precision–recall-based cost metric to strictly penalize imbalanced outcomes. Experimental results demonstrate that under identical wall-clock time budgets, the proposed framework achieves an average 1.38% reduction in aggregated cost across all models, with the highly sensitive VGG19 backbone showing a 4.00% improvement. Benchmarking against Bayesian optimization, genetic algorithms, and random search confirms that our method achieves superior optimization quality with statistical significance (p < 0.05). Under a rigorous IoU = 0.75 threshold, the optimized models consistently yielded F1-scores 0.8444 ± 0.0346. Ablation studies further validate that the collaborative model exchange is essential for accelerating convergence in rugged loss landscapes. This research offers a practical, scalable, and cost-efficient solution for deploying robust AI surveillance in resource-constrained smart city infrastructure. Full article
(This article belongs to the Special Issue Advances in Deep Learning-Based Data Analysis)
Show Figures

Figure 1

15 pages, 2849 KB  
Article
Empowering Rural Livestock Health: AI-Powered Early Detection of Cattle Diseases
by Dammavalam Srinivasa Rao, P. Chandra Sekhar Reddy, Annam Revathi, Vangipuram Sravan Kiran, Nuvvusetty Rajasekhar, Nadella Sandhya, Pulipati Venkateswara Rao, Adla Sai Karthik and Puvvala Jogeeswara Venkata Naga Sai
AI 2026, 7(4), 137; https://doi.org/10.3390/ai7040137 - 9 Apr 2026
Viewed by 570
Abstract
This paper presents a novel approach for the early detection of cattle diseases. We present a uniquely integrated image classification-based project for real-time cattle disease diagnosis that combines image classification models to identify diseases accurately; a seamless, user-friendly dashboard for real-time monitoring with [...] Read more.
This paper presents a novel approach for the early detection of cattle diseases. We present a uniquely integrated image classification-based project for real-time cattle disease diagnosis that combines image classification models to identify diseases accurately; a seamless, user-friendly dashboard for real-time monitoring with data visualization and instant predictions; and a mobile application that acts as a data source. The mobile application enables real-time collection of farmer and cattle-related data, including age, number of cattle, vaccination cycles, cattle images, and location metadata. Our AI-based cattle health monitoring project enables the early, efficient, scalable, and timely detection of Lumpy Skin Disease (LSD) and Foot and Mouth Disease (FMD) in cattle with high accuracy. A dataset of approximately 1600 LSD/non-LSD images and 840 FMD images was used to train multiple classification networks such as EfficientNetB0, ResNet50, VGG16, EfficientNetV2B0, and EfficientNetV2S, along with a soft-voting ensemble at inference. The proposed framework achieved a maximum testing accuracy of 98.36% for LSD classification and 99.84% for FMD classification under internal validation. These results indicate strong disease recognition capability, with ensemble-based prediction improving robustness, particularly for FMD classification. The proposed system enables practical, early, efficient, and scalable applications of AI research to improve livestock health monitoring and support the early prevention of widespread disease outbreaks. Full article
Show Figures

Graphical abstract

25 pages, 15195 KB  
Article
An Interpretable Deep Learning Approach for Brain Tumor Classification Using a Bangladeshi Brain MRI Dataset
by Md. Saymon Hosen Polash, Md. Tamim Hasan Saykat, Md. Ehsanul Haque, Md. Maniruzzaman, Mahe Zabin and Jia Uddin
BioMedInformatics 2026, 6(2), 19; https://doi.org/10.3390/biomedinformatics6020019 - 7 Apr 2026
Viewed by 1063
Abstract
Magnetic resonance imaging (MRI) is a critical clinical tool that requires precise and reliable interpretation for effective brain tumor diagnosis and timely treatment planning. Deep learning methods have advanced automated tumor classification greatly in the last few years, but many of the current [...] Read more.
Magnetic resonance imaging (MRI) is a critical clinical tool that requires precise and reliable interpretation for effective brain tumor diagnosis and timely treatment planning. Deep learning methods have advanced automated tumor classification greatly in the last few years, but many of the current methods are still challenged by a lack of interpretability, a lack of testing on region-focused data, and a lack of model robustness testing. Such limitations reduce clinical trust and limit the practice of automated diagnostic systems. To address these challenges, this study proposes an interpretable deep learning model for classifying brain tumors using the PMRAM dataset, which is a Bangladeshi brain MRI collection containing four categories: glioma, meningioma, pituitary tumor, and normal brain.. The proposed pipeline combines image preprocessing and feature enhancement methods, and then it trains a series of squeeze-and-excitation (SE)-enhanced convolutional neural networks such as VGG19, DenseNet201, MobileNetV3-Large, InceptionV3, and EfficientNetB3. The SE-enhanced EfficientNetB3 performed best, with 98.70% accuracy, 98.77% precision, 98.70% recall, and 98.70% F1-score. Cross-validation also demonstrated stable performance, with a mean accuracy of 96.89%. The model also exhibited efficient inference with low GPU memory consumption, enabling predictions in about 2–4 s per MRI image. Grad-CAM++ and saliency maps were used to improve the transparency of the results, and it was found that the network was concentrated on the clinically significant parts of the tumor, which affected the model predictions. Further robustness analysis and cross-dataset testing are additional evidence of the generalization possibility of the model. An online application was also implemented to allow real-time prediction and visual explanation of brain tumors. Overall, the proposed framework offers a precise, interpretable, and promising solution to automated brain tumor classification using MRI images. Full article
Show Figures

Figure 1

15 pages, 2437 KB  
Article
A Hybrid Self-ONN and Vision Mamba Architecture for Robust Radio Interference Recognition in GNSS Applications
by Nursultan Meirambekuly, Margulan Ibraimov, Bakyt Khaniyev, Beibit Karibayev, Alisher Skabylov, Nursultan Uzbekov, Sungat Koishybay, Timur Dautov, Ainur Khaniyeva and Bagdat Kozhakhmetova
Electronics 2026, 15(7), 1498; https://doi.org/10.3390/electronics15071498 - 3 Apr 2026
Viewed by 357
Abstract
Radio-frequency interference (RFI) poses a critical challenge for modern high-precision Global Navigation Satellite System (GNSS) applications, as both intentional and unintentional interference can significantly degrade positioning accuracy and reliability. With increasingly sophisticated interference sources, robust and computationally efficient automatic recognition methods are required [...] Read more.
Radio-frequency interference (RFI) poses a critical challenge for modern high-precision Global Navigation Satellite System (GNSS) applications, as both intentional and unintentional interference can significantly degrade positioning accuracy and reliability. With increasingly sophisticated interference sources, robust and computationally efficient automatic recognition methods are required for next-generation GNSS receivers. Although deep learning approaches show strong potential for interference detection, their high computational cost often limits deployment in resource-constrained navigation hardware. This paper proposes a hybrid deep learning architecture for radio interference recognition in high-precision GNSSs. The framework employs a dual-branch design integrating complementary signal representations. A Self-Organizing Operational Neural Network (Self-ONN) extracts nonlinear temporal features from raw one-dimensional signals, while a Vision Mamba state-space model processes two-dimensional time-frequency spectrograms obtained via Short-Time Fourier Transform (STFT). The fused features enable accurate classification of diverse interference types with high computational efficiency. Experiments on a synthetic dataset demonstrate that the proposed model achieves 99.83% accuracy and F1-score, outperforming ResNet18, VGG16, and Vision Transformer while reducing computational complexity by up to 42% and improving inference speed by up to 35%, supporting its applicability for intelligent interference monitoring in GNSS receivers. Full article
Show Figures

Figure 1

23 pages, 2936 KB  
Article
Lightweight Transient-Source Detection Method for Edge Computing
by Jiahao Zhang, Yutian Fu, Feng Dong and Lingfeng Huang
Universe 2026, 12(4), 101; https://doi.org/10.3390/universe12040101 - 1 Apr 2026
Viewed by 288
Abstract
Transient-source detection without relying on difference images still faces challenges in achieving high accuracy, especially under practical space-based astronomical survey conditions where the data volume is enormous, on-orbit transmission bandwidth is limited, and real-time response is required for rapid follow-up observations. To address [...] Read more.
Transient-source detection without relying on difference images still faces challenges in achieving high accuracy, especially under practical space-based astronomical survey conditions where the data volume is enormous, on-orbit transmission bandwidth is limited, and real-time response is required for rapid follow-up observations. To address these issues, this paper proposes a lightweight detection network that integrates multi-scale feature fusion with contextual feature extraction, enabling efficient real-time processing on resource-constrained edge devices. The proposed model enhances robustness to point-spread-function variations across observation conditions and to complex background environments, while simultaneously improving detection accuracy. To evaluate performance comprehensively, lightweight VGG and lightweight ResNet architectures and other baseline models—commonly used as baselines for transient-source detection—are adopted for comparison. Experimental results show that under the condition that the models have approximately the same number of parameters, the proposed network achieves the best accuracy, obtaining nearly 1% improvement compared with the best-performing baseline model. Based on this design, an ultra-lightweight version with only 7k parameters is further developed by incorporating a compact multi-scale module, improving accuracy by 1% over the version without the multi-scale structure. Moreover, through heterogeneous knowledge distillation and adaptive iterative training, the accuracy of the ultra-lightweight model is further increased from 93.3% to 94.0%. Finally, the model is deployed and validated on an AI hardware acceleration platform. The results demonstrate that the proposed method substantially improves inference throughput while maintaining high accuracy, providing a practical solution for real-time, low-latency, on-device transient-source detection under large data volume and limited transmission conditions. Specifically, the proposed models are trained offline on a high-performance GPU and subsequently deployed on the Fudan Microelectronics 7100 AI board to evaluate their real-world inference efficiency on resource-constrained edge devices. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Modern Astronomy)
Show Figures

Figure 1

19 pages, 4575 KB  
Article
DenseNet-BiFPN-ECA Fusion Network: An Enhanced Transfer Learning Approach for Tomato Leaf Disease Recognition
by Lina Liang, Jingnan Chen, Ying Tian, Hongyan Wang, Yiting Cai, Fenglin Zhong, Senpeng Wang, Maomao Hou and Junyang Lu
Horticulturae 2026, 12(4), 423; https://doi.org/10.3390/horticulturae12040423 - 31 Mar 2026
Viewed by 406
Abstract
Early and accurate identification of tomato leaf diseases constitutes a key safeguard for mitigating economic losses in tomato production. Conventional tomato leaf disease detection methodologies are constrained by inherent limitations, such as low operational efficiency, inadequate detection precision, and limited adaptability to environmental [...] Read more.
Early and accurate identification of tomato leaf diseases constitutes a key safeguard for mitigating economic losses in tomato production. Conventional tomato leaf disease detection methodologies are constrained by inherent limitations, such as low operational efficiency, inadequate detection precision, and limited adaptability to environmental fluctuations. In contrast, the integration of deep learning techniques has yielded improvements in this research domain. Consequently, the development of deep learning-based approaches for the rapid and precise detection of tomato leaf diseases holds considerable theoretical significance and practical application value. To improve the detection accuracy of tomato leaf diseases, this study proposes a transfer learning-based DenseNet disease recognition model named DenseNet-BiFPN-ECA Fusion Network. The bidirectional feature pyramid network (BiFPN) is introduced at the terminal of DenseNet121 to achieve multi-scale feature fusion, while the efficient channel attention (ECA) mechanism is applied to enhance the discriminative capacity of fused features. Classification is ultimately completed via a global average pooling layer and a fully connected layer. The experimental results demonstrate that the improved model achieves an accuracy of 90.63% on the small-sample tomato leaf dataset collected from complex greenhouse environments, representing an improvement of 20.32 percentage points over the original DenseNet121 model. On the large-scale open-source Plant Village dataset, the model attains an accuracy of 98.47%, significantly outperforming the baseline models. Furthermore, a comparative analysis shows that the highest accuracy achieved by DenseNet, ResNet101, and VGG16 models on the same dataset is only 83.59% (within ±0.5%). This result validates the effectiveness of DenseNet-BiFPN-ECA Fusion Network in disease recognition tasks. The model provides a reliable technical reference for the intelligent diagnosis of tomato leaf diseases. Full article
(This article belongs to the Special Issue Computer Vision and Machine Learning in Horticulture Plants)
Show Figures

Figure 1

34 pages, 6554 KB  
Article
Syncretic Grad-CAM Integrated ViT-CNN Hybrids with Inherent Explainability for Early Thyroid Cancer Diagnosis from Ultrasound
by Ahmed Y. Alhafdhi, Gibrael Abosamra and Abdulrhman M. Alshareef
Diagnostics 2026, 16(7), 999; https://doi.org/10.3390/diagnostics16070999 - 26 Mar 2026
Viewed by 371
Abstract
Background/Objectives: Accurate detection of thyroid cancer using ultrasound remains a challenge, as malignant nodules can be microscopic and heterogeneous, easily confused with point clusters and borderline-featured tissues. Current studies in deep learning demonstrate good performance with convolutional neural networks (CNNs) and clustering; however, [...] Read more.
Background/Objectives: Accurate detection of thyroid cancer using ultrasound remains a challenge, as malignant nodules can be microscopic and heterogeneous, easily confused with point clusters and borderline-featured tissues. Current studies in deep learning demonstrate good performance with convolutional neural networks (CNNs) and clustering; however, many approaches focus on local tissue and provide limited, non-quantitative interpretation, reducing clinical confidence. This study proposes an integrated framework combining enhanced convolutional feature encoders (DenseNet169 and VGG19) with an enhanced vision transformer (ViT-E) to integrate local feature and global relational context during learning, rather than delayed integration. Methods: The proposed framework integrates enhanced convolutional feature encoders (DenseNet169 and VGG19) with an enhanced vision transformer (ViT-E), enabling simultaneous learning of local feature representations and global relational context. This design allows feature fusion during the learning stage instead of delayed integration, aiming to improve diagnostic performance and interpretability in thyroid ultrasound image analysis. Results: The best-performing model, ViT-E–DenseNet169, achieved 98.5% accuracy, 98.9% sensitivity, 99.15% specificity, and 97.35% AUC, surpassing the robust basic hybrid model (CNN–XGBoost/ANN) and existing systems. A second contribution is improved interpretability, moving from mere illustration to validation. Gradient-weighted class activation mapping (Grad-CAM) maps demonstrated distinct and clinically understandable concentration patterns across various thyroid cancers: precise intralesional concentration for high-confidence malignancies (PTC = 0.968), edge/interface concentration for capsule risk patterns (PTC = 0.957), and broader-field activation consistent with infiltration concerns (PTC = 0.984), while benign scans showed low and diffuse activation (PTC = 0.002). Spatial audits reinforced this behavior (IoU/PAP: 0.72/91%, 0.65/78%, 0.58/62%). Conclusions: The integrated ViT-E–DenseNet169 framework provides highly accurate thyroid cancer detection while offering clinically meaningful interpretability through Grad-CAM-based spatial validation, supporting improved confidence in AI-assisted ultrasound diagnosis. Full article
(This article belongs to the Special Issue Deep Learning Techniques for Medical Image Analysis)
Show Figures

Figure 1

23 pages, 11748 KB  
Article
Polarization-Regularized Adversarial Pruning for Efficient Radio Frequency Fingerprint Identification on IoT Devices
by Caidan Zhao, Haoliang Jiang, Jinhui Yu, Zepeng Meng and Xuhao He
Sensors 2026, 26(6), 2005; https://doi.org/10.3390/s26062005 - 23 Mar 2026
Viewed by 428
Abstract
Radio frequency fingerprint identification (RFFI) based on physical-layer characteristics provides a reliable solution for secure authentication of Internet of Things (IoT) devices. Deep neural networks have demonstrated strong capability in improving RFFI performance; however, their high computational complexity and large parameter size pose [...] Read more.
Radio frequency fingerprint identification (RFFI) based on physical-layer characteristics provides a reliable solution for secure authentication of Internet of Things (IoT) devices. Deep neural networks have demonstrated strong capability in improving RFFI performance; however, their high computational complexity and large parameter size pose significant challenges for deployment on resource-constrained edge devices. In RFFI tasks, existing pruning methods often lack effective performance recovery strategies, which leads to noticeable degradation in identification accuracy after pruning. To address this issue, this paper proposes a pruning method based on adversarial learning and polarization regularization. Polarization regularization is applied to learnable soft masks to effectively distinguish channels to be pruned from those to be retained. In addition, an adversarial learning-based performance recovery strategy is introduced to align the output feature distributions between the baseline network and the pruning network, thereby improving identification accuracy after pruning. Experimental results on multiple RFFI datasets demonstrate that the proposed method can effectively prune ResNet18 and VGG16, achieving substantial reductions in model complexity with only minor losses in identification performance. Full article
Show Figures

Figure 1

22 pages, 8609 KB  
Article
Integrating SimAM Attention and S-DRU Feature Reconstruction for Sentinel-2 Imagery-Based Soybean Planting Area Extraction
by Haotong Wu, Xinwen Wan, Rong Qian, Chao Ruan, Jinling Zhao and Chuanjian Wang
Agriculture 2026, 16(6), 693; https://doi.org/10.3390/agriculture16060693 - 19 Mar 2026
Viewed by 336
Abstract
Accurate and stable acquisition of the spatial distribution of soybean planting areas is essential for supporting precision agricultural monitoring and ensuring food security. However, crop remote-sensing mapping for specific regions still faces critical data bottlenecks: high-precision, large-scale pixel-level annotation is costly, resulting in [...] Read more.
Accurate and stable acquisition of the spatial distribution of soybean planting areas is essential for supporting precision agricultural monitoring and ensuring food security. However, crop remote-sensing mapping for specific regions still faces critical data bottlenecks: high-precision, large-scale pixel-level annotation is costly, resulting in scarce available labeled samples that make it difficult to construct large-scale training datasets. Although parameter-intensive models such as FCN and SegNet can achieve sufficient end-to-end training on large-scale public remote sensing datasets like LoveDA, when directly applied to the data-limited dataset in this study area, the models are prone to overfitting, leading to a significant decline in generalization ability. To address these issues, this study proposes a lightweight U-shaped semantic segmentation model, SimSDRU-Net. The model utilizes a pre-trained VGG-16 backbone to extract shallow texture and deep semantic features. The pre-trained weights mitigate the impact of overfitting in data-limited settings. In the decoding stage, a parameter-free lightweight SimAM attention module enhances effective soybean features and suppresses soil background redundancy, while an embedded S-DRU unit fuses multi-scale features for deep complementary reconstruction to improve edge detail capture. A label dataset was constructed using Sentinel-2 images as the data source and Menard County (USA) as the study area. The USDA CDL was used as a foundation for the dataset, with Google high-resolution images serving as visual interpretation aids. In the context of the experiment, Deeplabv3+ and U-Net++ were compared with U-Net under identical conditions. The results demonstrated that SimSDRU-Net exhibited optimal performance, with MIoU of 89.03%, MPA of 93.81%, and OA of 95.96%. Specifically, SimSDRU-Net uses the SimAM attention module to generate spatial attention weights by analyzing feature statistical differences through an energy function, so as to adaptively enhance soybean texture features. Meanwhile, the S-DRU unit groups, dynamically weights, and cross-branch reconstructs multi-scale convolutional features to preserve fine boundary details and achieve accurate segmentation of soybean plots. The present study demonstrates that SimSDRU-Net integrates lightweight design and high precision in data-limited scenarios, thereby providing effective technical support for the rapid extraction of soybean planting areas in North America. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

Back to TopTop