Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (6,128)

Search Parameters:
Keywords = YOLOv4

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 1335 KB  
Article
Experimental Analysis of Animal Behavior for Biomedical Applications
by Florin Rotaru, Silviu-Ioan Bejinariu, Hariton-Nicolae Costin, Ramona Luca, Mihaela Luca, Cristina Diana Nita, Diana Costin, Bogdan-Ionel Tamba, Ivona Costachescu, Gabriela-Dumitrita Stanciu and Gabriela-Gladiola Petroiu
Appl. Sci. 2026, 16(9), 4488; https://doi.org/10.3390/app16094488 (registering DOI) - 2 May 2026
Abstract
This study addresses the problem of robust video-based tracking of laboratory rats in open-field and Y-maze experiments under challenging acquisition conditions, including non-uniform illumination, low contrast, and heterogeneous recording setups. Existing approaches based on classical image processing or deep learning often fail to [...] Read more.
This study addresses the problem of robust video-based tracking of laboratory rats in open-field and Y-maze experiments under challenging acquisition conditions, including non-uniform illumination, low contrast, and heterogeneous recording setups. Existing approaches based on classical image processing or deep learning often fail to maintain stable localization under such conditions or require large, annotated datasets. We propose a hybrid tracking framework that combines an improved motion–appearance voting mechanism with consistency-constrained optimization for open-field experiments, together with a comparative deep learning-based detection strategy for Y-maze analysis. The proposed method introduces (i) adaptive dual-threshold motion extraction, (ii) directionally constrained temporal validation, and (iii) a robustness-driven fusion of motion and appearance cues. Experimental results demonstrate that the proposed approach achieves reliable tracking with a maximum localization error below 10 pixels under severe illumination variations. In the Y-maze scenario, a comparative evaluation of multiple detectors (YOLOv5, YOLOv9, YOLO12, Faster R-CNN) highlights the trade-off between accuracy and inference time, with YOLOv9 providing the best balance. The main contribution consists of enabling robust behavioral quantification in low-quality experimental conditions using limited training data, bridging the gap between classical tracking robustness and deep learning flexibility. Full article
(This article belongs to the Section Biomedical Engineering)
24 pages, 22833 KB  
Article
DAER-YOLO: Defect-Aware and Edge-Reconstruction Enhanced YOLO for Surface Defect Detection of Varistors
by Wu Xie, Shushuo Yao, Tao Zhang, Gaoxue Qiu, Dong Li, Fuxian Luo and Yong Fan
J. Imaging 2026, 12(5), 198; https://doi.org/10.3390/jimaging12050198 (registering DOI) - 2 May 2026
Abstract
Varistors are critical overvoltage protection components in modern power electronic systems. They effectively absorb and dissipate surge energy to ensure the safe and stable operation of electrical equipment. However, surface defects can lead to substandard performance or even trigger equipment failure, compromising overall [...] Read more.
Varistors are critical overvoltage protection components in modern power electronic systems. They effectively absorb and dissipate surge energy to ensure the safe and stable operation of electrical equipment. However, surface defects can lead to substandard performance or even trigger equipment failure, compromising overall system stability. Therefore, high-precision surface defect detection is essential for quality assurance. To address these challenges, we propose a lightweight model termed Defect-Aware and Edge-Reconstruction Enhanced YOLO (DAER-YOLO) for efficient varistor inspection. First, we construct a C3k2-based defect-aware enhancement module (C3k2-iEMA). This module tackles the difficulty of extracting features from small or morphologically complex defects. By integrating multi-scale feature extraction, an attention mechanism, and efficient nonlinear mapping, it strengthens the perception of defect details. Second, to enhance the reconstruction capability for edge damage and small-object defects, we introduce the Efficient Up-Convolution Block (EUCB). This block improves multi-level feature fusion and generates clearer enhanced feature maps. Based on these improvements, DAER-YOLO outperforms the YOLOv11n baseline on a custom varistor dataset, with mAP@50 and mAP@50:95 increasing by 1.6% and 2.3%, respectively. Experimental results demonstrate that the model effectively improves detection accuracy while exhibiting significant potential for real-time industrial applications. Full article
(This article belongs to the Section Computer Vision and Pattern Recognition)
Show Figures

Figure 1

29 pages, 19729 KB  
Article
Deep Learning-Based Multistage Peach Ripeness Detection with Data Leakage Mitigation and Real-World Validation
by Salvador Castro-Tapia, Germán Díaz-Florez, Rafael Reveles-Martínez, Héctor A. Guerrero-Osuna, Luis F. Luque-Vega, Humberto Morales-Magallanes, Jorge Pablo Vega-Borrego, Gilberto Vázquez-García and Carlos A. Olvera-Olvera
Appl. Sci. 2026, 16(9), 4484; https://doi.org/10.3390/app16094484 (registering DOI) - 2 May 2026
Abstract
Accurate fruit ripeness assessment is essential for optimizing harvest timing and ensuring postharvest quality, particularly in climacteric fruits such as peaches, which exhibit rapid and heterogeneous ripening. This study proposes a deep learning-based approach for multistage peach ripeness classification across five maturity levels [...] Read more.
Accurate fruit ripeness assessment is essential for optimizing harvest timing and ensuring postharvest quality, particularly in climacteric fruits such as peaches, which exhibit rapid and heterogeneous ripening. This study proposes a deep learning-based approach for multistage peach ripeness classification across five maturity levels (green, green-blushed, blushed, yellow-blushed, and fully yellow). Four datasets were constructed using controlled image acquisition, segmentation, data augmentation, and perceptual hashing to mitigate data leakage. The performance of AlexNet, EfficientNet-B0, and three YOLO (You Only Look Once) architectures (YOLOv8, YOLOv11, and YOLOv12) was evaluated using standard metrics, including accuracy, precision, recall, F1 score, mAP, and inference speed. Results show that YOLO-based models significantly outperform classical networks, achieving accuracies between 95.25% and 98.3% and mAP@0.5 above 98.25%, while also reducing inference time to 8.1–12.7 ms compared with 722.23 ms for AlexNet and 171.87 ms for EfficientNet-B0. In a practical sorting experiment with 214 peaches, YOLOv12 achieved 92.06% accuracy, demonstrating robust real-world performance. Misclassifications were primarily observed between adjacent ripeness stages. These findings indicate that YOLO-based models provide an effective and scalable solution for real-time fruit sorting, while the use of perceptual hashing enhances dataset reliability and model generalization for deployment in agricultural quality control systems. Full article
(This article belongs to the Special Issue Intelligent Systems: Design and Engineering Applications)
29 pages, 31629 KB  
Article
Quantification of Opercular Pigmentation Changes in Farmed Atlantic Salmon: A Novel Application for Computer Vision in Fish Welfare Assessment
by Talha Laique, Mikkel Gunnes, Ole Folkedal, Jonatan Nilsson, Evelina A. L. Green, Hannah Normann Gundersen, Øyvind Øverli and Habib Ullah
Fishes 2026, 11(5), 271; https://doi.org/10.3390/fishes11050271 (registering DOI) - 2 May 2026
Abstract
Intensive salmon farming is associated with high mortality rates, highlighting the need for new welfare indicators that can detect adverse conditions earlier and less invasively than many current approaches. Existing animal-based indicators used in the industry typically depend on subjective scoring and provide [...] Read more.
Intensive salmon farming is associated with high mortality rates, highlighting the need for new welfare indicators that can detect adverse conditions earlier and less invasively than many current approaches. Existing animal-based indicators used in the industry typically depend on subjective scoring and provide information mostly after welfare problems have already developed, thereby raising questions about their efficacy. Examples include emaciation, wounds, or scale loss, etc. Preliminary data and ongoing investigation suggest that melanin-based skin pigmentation may change dynamically with stress and condition in salmonid fishes. In this study, we present a semi-automated methodology for assessing changes in the grayscale intensity of melanin-based skin spots within the operculum region of adult Atlantic salmon (Salmo salar) kept in seawater. The pipeline combines computer vision models to detect the operculum, segment individual spots, and extract grayscale-based features for spot-level analysis over time. The method was applied to out-of-water images collected before and after exposure to a confinement episode. The results showed an overall shift in grayscale intensity from black to pigmentation fading after the challenge, although responses varied among individuals. These findings indicate that the proposed methodology can detect temporal changes in opercular melanin-based spots under applied experimental conditions. We therefore present this work as proof of principle for using computer vision to quantify changes in melanin-based skin spots as a potentially useful, non-invasive indicator of stress and welfare in Atlantic Salmon. Full article
(This article belongs to the Special Issue Computer Vision Applications for Fisheries and Aquaculture)
Show Figures

Figure 1

21 pages, 5108 KB  
Article
Lightweight Detection and Adaptive Path Planning for Selective Hotan Rose Harvesting
by Jijing Lin, Yuhang Yang, Baojian Ma, Zhenghao Wu and Bangbang Chen
Sensors 2026, 26(9), 2848; https://doi.org/10.3390/s26092848 (registering DOI) - 2 May 2026
Abstract
Selective harvesting of Hotan roses requires distinguishing between buds and blooms for different industrial uses. However, balancing detection accuracy and computational efficiency for edge deployment remains a challenge. This study proposes an integrated framework combining a lightweight detection model, Rose_YOLO, with an adaptive [...] Read more.
Selective harvesting of Hotan roses requires distinguishing between buds and blooms for different industrial uses. However, balancing detection accuracy and computational efficiency for edge deployment remains a challenge. This study proposes an integrated framework combining a lightweight detection model, Rose_YOLO, with an adaptive path-planning algorithm, the ROSE algorithm, to address these issues. The Rose_YOLO model optimizes the YOLOv8n architecture by incorporating the C2f-Faster-CGLU module and a Rose_Head detection head to enhance feature extraction while reducing redundancy. The ROSE algorithm integrates an improved genetic algorithm (GA) with a reciprocating search mechanism to dynamically optimize picking sequences based on scene complexity. Experimental results demonstrate that Rose_YOLO achieves a precision of 90.4% and a mAP@0.5 of 96.6% for blooms and a precision of 88.4% with a mAP@0.5 of 91.7% for buds. Compared to the baseline YOLOv8n, the model reduces parameters by 47.46% to 1.579 million, compresses the size to 3.19 MB, and lowers computational complexity to 4.6 GFLOPs. For path planning, the ROSE algorithm generates optimal paths with an average length of 2796.94 pixels, which is 73.1% shorter than the reciprocating algorithm and 51.6% shorter than the standard GA. Furthermore, it achieves an average runtime of only 7.33 ms, significantly outperforming traditional methods with respect to computational speed. In conclusion, the proposed framework achieves a superior balance between lightweight design and detection performance. The successful deployment on edge devices validates its effectiveness in providing real-time visual guidance and efficient path planning, offering a robust technical solution for the automated selective harvesting of roses in complex field environments. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

16 pages, 14902 KB  
Article
AI-Augmented Partial Discharge Analytics for TEV-Based Monitoring
by Hsien Wee Bryan Sim, Sivaneasan Bala Krishnan and Kai Xian Lai
Appl. Sci. 2026, 16(9), 4462; https://doi.org/10.3390/app16094462 (registering DOI) - 2 May 2026
Abstract
Transient earth voltage (TEV) measurements are widely used for partial discharge (PD) monitoring in medium-voltage equipment due to their non-intrusive nature and suitability for field deployment. However, TEV-based PD analytics remain challenging in practical environments because PD signatures often overlap with noise characteristics. [...] Read more.
Transient earth voltage (TEV) measurements are widely used for partial discharge (PD) monitoring in medium-voltage equipment due to their non-intrusive nature and suitability for field deployment. However, TEV-based PD analytics remain challenging in practical environments because PD signatures often overlap with noise characteristics. Recent advances in deep learning enable the image-based analysis of phase-resolved partial discharge (PRPD) representations, offering improved robustness compared to conventional signal-based methods. This paper presents a structured experimental investigation of classification-based and localisation-based deep learning models for TEV PRPD analytics. A multi-column convolutional neural network (MCNN) is evaluated as a classification model, while YOLOv5 and YOLOv12, based on the You Only Look Once (YOLO) framework, are investigated as object detection frameworks capable of PD localisation. Experiments are conducted using both original and expanded datasets, with additional analysis on the impact of hyperparameter optimisation. The results show that dataset expansion and hyperparameter tuning improve classification and detection performance for several models. The MCNN achieves strong classification accuracy for PD/noise screening, while YOLOv5 demonstrates substantial improvement in localisation performance after optimisation. In contrast, YOLOv12 maintains stable detection performance under the evaluated training configurations. The results highlight the trade-offs between classification accuracy, localisation capability, and annotation complexity when selecting deep learning models for PD analytics. These findings provide practical insights into the deployment of deep learning techniques for TEV-based PD monitoring systems and highlight the complementary roles of classification and localisation frameworks in PRPD analysis. Full article
Show Figures

Figure 1

22 pages, 7661 KB  
Article
YOLOv11-SMS: An Improved Algorithm for Impurity Detection in Seed Cotton
by Wenyan Yuan, Laigang Zhang, Donghe Wang and Zhijun Guo
Sensors 2026, 26(9), 2835; https://doi.org/10.3390/s26092835 - 1 May 2026
Abstract
To enhance the precision of cottonseed impurity detection and address issues such as high miss-detection rates and suboptimal performance, this paper introduces an improved YOLOv11 algorithm, termed YOLOv11-SMS. Initially, the algorithm integrates a local self-attention mechanism (LRSA) to design the C2PSA-SL module, which [...] Read more.
To enhance the precision of cottonseed impurity detection and address issues such as high miss-detection rates and suboptimal performance, this paper introduces an improved YOLOv11 algorithm, termed YOLOv11-SMS. Initially, the algorithm integrates a local self-attention mechanism (LRSA) to design the C2PSA-SL module, which augments the model’s ability to learn local information while maintaining global feature awareness. Furthermore, the feature extraction stage and the network head incorporate a multi-branch reparameterized convolution (MBRConv) module, enhancing feature extraction capabilities while preserving the model’s lightweight properties. Lastly, a spatial adaptive modulation (SAFM) module is introduced to optimize the detection of small targets. Experimental results demonstrate that YOLOv11-SMS outperforms the baseline model, with mAP@50–95 increasing from 79.42% to 82.49%, an improvement of 3.07 percentage points. The average mIOU increased from 90.98% to 94.18%, representing a 3.2 percentage point improvement. Moreover, the model achieves an impressive real-time inference speed of 178.63 frames per second (FPS), effectively balancing detection accuracy and speed, offering an efficient and precise solution for cottonseed impurity detection. Full article
Show Figures

Figure 1

20 pages, 30829 KB  
Article
Crop-IRM: An Intelligent Recognition and Management System for Organ Characteristics of Crop Germplasm Resources
by Jie Zhang, Chenyao Yang, Hailin Peng, Xintong Wei, Jiaqi Zou, Shiyu Wang, Zhaohong Lu, Xianming Tan and Feng Yang
Agriculture 2026, 16(9), 996; https://doi.org/10.3390/agriculture16090996 - 30 Apr 2026
Abstract
The traditional methods of field-based phenotypic data collection for crop germplasm resources are often inefficient and highly subjective. As the foundation for breeding innovation, these resources require precise identification of phenotypic traits for effective evaluation and utilization. Therefore, efficient and standardized management of [...] Read more.
The traditional methods of field-based phenotypic data collection for crop germplasm resources are often inefficient and highly subjective. As the foundation for breeding innovation, these resources require precise identification of phenotypic traits for effective evaluation and utilization. Therefore, efficient and standardized management of germplasm data is critical during the breeding process. To address this, we have developed an intelligent recognition and management system focused on the crop’s organ characteristics. The system consists of a web client for overall project management and data download, and a WeChat Mini Program for data collection and uploading. Both components are integrated with image analysis models. Using a soybean variety screening experiment as a case study, we have constructed multiple high-definition datasets for soybean phenotypic traits, and employed YOLOv11 series models for object detection, image classification, instance segmentation, and pose estimation to build analytical models for each of these traits. All models achieved a mean average precision (mAP@0.5) exceeding 94%, along with a top1_accuracy of 0.999. In practical evaluations, all models took between 0.71 and 3.03 s to make predictions for 100 images, achieving an accuracy rate of over 98%. This system delivers a comprehensive solution for field phenotypic identification of crop germplasm resources, substantially enhancing the efficiency and objectivity of data collection and analysis. It serves as a valuable decision-support tool for precision breeding and digital agriculture. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

23 pages, 138069 KB  
Article
Instance Segmentation of Ship Images Based on Multi-Branch Adaptive Feature Fusion and Occluded Region Decoupling in Occluded Scenes
by Yuwei Zhu, Wentao Xue, Wei Liu, Hui Ye and Yaohua Shen
J. Mar. Sci. Eng. 2026, 14(9), 841; https://doi.org/10.3390/jmse14090841 - 30 Apr 2026
Abstract
Instance segmentation accurately extracts the position and outline of ships, serving as the foundation for maritime safety tasks such as multi-object tracking, sensor fusion, and collision warning. This study focuses on single-frame segmentation and aims to address the challenge of multi-scale ship occlusion [...] Read more.
Instance segmentation accurately extracts the position and outline of ships, serving as the foundation for maritime safety tasks such as multi-object tracking, sensor fusion, and collision warning. This study focuses on single-frame segmentation and aims to address the challenge of multi-scale ship occlusion in congested ports, providing reliable observational data through high-precision recognition to ensure navigation safety. Existing methods suffer from performance degradation in complex maritime environments due to factors such as multi-scale distribution, low resolution of distant targets, and frequent occlusions. Among these, ship occlusion is particularly problematic as it leads to feature confusion between adjacent instances and inaccurate boundary segmentation. To address these challenges, we propose a novel instance segmentation algorithm (MAF-ORDNet) based on Multi-branch Adaptive Feature Fusion and Occluded Region Decoupling. Firstly, a multi-branch adaptive feature fusion module is designed to capture contextual information through different receptive fields and dynamically fuse multi-scale features, thereby restoring occluded semantics and enhancing robustness. Secondly, an occlusion region decoupling module is constructed to accurately localize occluded regions and enhance contour responses via adaptive sampling, achieving refined boundary processing. In addition, we constructed and annotated the Occlusion ShipSeg dataset, which contains 1969 real occlusion images, 2150 simulated occlusion images, and 1132 images under adverse weather conditions, totaling 17,352 fine instance annotations. Experimental results show that, compared with PatchDCT, YOLOv11s, and Mask2Former, our method improves AP by 2.7%, 3.2%, and 2.4%, respectively, while maintaining a comparable inference speed to YOLOv8s. These results confirm that MAF-ORDNet achieves a favorable balance between accuracy and efficiency in multi-scale occluded ship segmentation tasks. Full article
(This article belongs to the Section Ocean Engineering)
28 pages, 2497 KB  
Article
Research on the Application of Time-Frequency Characteristics of GPR in Railway Mud Pumping Intelligent Detection
by Wenxing Shi, Shilei Wang, Feng Yang, Chi Zhang, Fanruo Li and Suping Peng
Remote Sens. 2026, 18(9), 1393; https://doi.org/10.3390/rs18091393 - 30 Apr 2026
Abstract
Ground penetrating radar (GPR), as an efficient non-destructive testing technique, plays a crucial role in the structural condition assessment and defect identification of railway ballast. Typical defects such as mud pumping generally exhibit characteristics in B-scan images including weak reflections, blurred boundaries, and [...] Read more.
Ground penetrating radar (GPR), as an efficient non-destructive testing technique, plays a crucial role in the structural condition assessment and defect identification of railway ballast. Typical defects such as mud pumping generally exhibit characteristics in B-scan images including weak reflections, blurred boundaries, and irregular structures, which pose significant challenges for stable detection and precise localization using existing methods that rely primarily on spatial feature modeling. Most current deep learning approaches focus on modeling spatial or temporal information, while lacking effective utilization of frequency-domain features, thereby limiting their discriminative capability under complex electromagnetic environments. To address these issues, this paper proposes a single-stage object detection framework, termed YOLO-DGW, based on time-frequency collaborative modeling. Built upon YOLOv8, the proposed method introduces a structure-aware spatial enhancement module to improve the representation of continuous GPR echo structures. Meanwhile, frequency-domain information is incorporated as a modulation prior to guide spatial feature learning, enhancing the model’s sensitivity to weak reflections and complex-shaped targets. In addition, A-CIoU loss function is designed to improve localization accuracy and stability for defect regions of varying scales. Experimental results demonstrate that YOLO-DGW achieves an F1-score of 63.06% and an AP@0.50 of 62.07%, representing improvements of approximately 7.41% and 2.8%, respectively, over the strongest baseline method. Compared with several mainstream object detection models, the proposed approach exhibits superior performance in both detection accuracy and cross-region generalization capability. These findings indicate that integrating frequency-domain information into spatial feature learning through a modulation mechanism can effectively enhance the model’s ability to discriminate weak-reflection anomalies, providing a novel time-frequency collaborative modeling paradigm for railway GPR defect detection. Full article
16 pages, 2473 KB  
Article
Incorporating Crop-Centric Segmentation and Enhanced YOLOv10 for Indirect Weed Detection in Bok Choy Fields
by Weili Li, Wenpeng Zhu, Qianyu Wang, Feng Gao, Kang Han and Xiaojun Jin
Agronomy 2026, 16(9), 907; https://doi.org/10.3390/agronomy16090907 - 30 Apr 2026
Abstract
Weed infestation poses a significant threat to bok choy (Brassica rapa subsp. chinensis) cultivation, reducing crop yield and quality through resource competition and pest facilitation. Traditional weed detection methods face two major bottlenecks: one is data annotation, arising from the need for [...] Read more.
Weed infestation poses a significant threat to bok choy (Brassica rapa subsp. chinensis) cultivation, reducing crop yield and quality through resource competition and pest facilitation. Traditional weed detection methods face two major bottlenecks: one is data annotation, arising from the need for extensive, species-diverse datasets, and the other is visual discrimination, due to the high morphological similarity between crops and weeds at certain growth stages. To address these challenges, this study proposed an indirect weed detection framework that combines an optimized You Only Look Once version 10 (YOLOv10) model for crop detection with Excess Green ExG-based segmentation of residual vegetation. The model incorporates RFD and C2f-WDBB modules to improve feature preservation and multi-scale fusion. Compared with baseline YOLOv10, the final proposed RCW-YOLOv10 reduced the number of parameters by 1.04 million and improved detection performance, achieving increases of 3.5%, 1.5%, and 1.1% percentage points in Precision, Recall, and mAP50, respectively, under field conditions. The system initially detected bok choy plants, subsequently localizing weeds by masking crop regions and thresholding residual ExG signals in the uncovered areas. The detected weed coordinates were used to construct a distribution map that may support targeted control in precision agriculture. This approach simplifies weed identification under the tested bok choy field conditions and may be adaptable to other crops after further validation. Full article
Show Figures

Figure 1

29 pages, 23221 KB  
Article
FMSD-YOLO12: An Efficient and Lightweight Network for Surface Defect Detection of Ferrite Permanent Magnets
by Chuanyu Zhan, Haiting Yu, Ruize Wu and Junfeng Li
Electronics 2026, 15(9), 1900; https://doi.org/10.3390/electronics15091900 - 30 Apr 2026
Abstract
To address micro-break and edge-chipping defects in ferrite magnetic sheets, as well as the difficulty of balancing detection accuracy and deployment cost under complex grinding-texture interference, this paper proposes FMSD-YOLO12, an efficient and lightweight defect detection model based on YOLOv12. The proposed method [...] Read more.
To address micro-break and edge-chipping defects in ferrite magnetic sheets, as well as the difficulty of balancing detection accuracy and deployment cost under complex grinding-texture interference, this paper proposes FMSD-YOLO12, an efficient and lightweight defect detection model based on YOLOv12. The proposed method follows a task-oriented design for three coupled challenges in ferrite magnetic sheet inspection, namely texture-interfered feature extraction, cross-scale feature inconsistency, and lightweight yet accurate defect localization. Specifically, a Spatially Re-weighted Convolution (SR-Conv) is introduced into the C3k2 backbone module to suppress repetitive grinding-texture noise and enhance the response contrast of subtle defect regions. A Context and Spatial Feature Calibration Network (CSFCN) is further developed to improve semantic consistency and spatial alignment during multi-scale feature fusion. In addition, a Lightweight Shared Detail-Enhanced Convolutional Detection head (LSDECD) is designed to strengthen weak-edge localization while reducing parameter redundancy through re-parameterization. Experimental results show that, with a comparable number of parameters, FMSD-YOLO12 improves mAP@50 by 2.40%, mAP@75 by 3.71%, and mAP@50-95 by 3.03% on the magnetic sheet defect dataset. These results demonstrate that the proposed model achieves a favorable balance between detection accuracy and computational efficiency for irregular defect detection under complex industrial backgrounds. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

32 pages, 91311 KB  
Article
From Geometric Exploration to Semantic Completion: Scene Exploration Convolution and Large Format Perception for Adverse-Weather UAV Aerial Object Detection
by Yize Zhao, Bo Wang and Jialei Zhan
Sensors 2026, 26(9), 2802; https://doi.org/10.3390/s26092802 - 30 Apr 2026
Abstract
Object detection from unmanned aerial vehicle (UAV) imagery is essential for applications such as traffic monitoring, disaster response, and urban surveillance, yet most existing methods are developed and evaluated under clear-sky conditions. In real-world UAV operations, adverse weather including fog, rain, and snow [...] Read more.
Object detection from unmanned aerial vehicle (UAV) imagery is essential for applications such as traffic monitoring, disaster response, and urban surveillance, yet most existing methods are developed and evaluated under clear-sky conditions. In real-world UAV operations, adverse weather including fog, rain, and snow introduces severe image degradation that simultaneously disrupts both the geometric and photometric properties of targets. This paper identifies two fundamental bottlenecks underlying this performance collapse: the lack of geometric invariance in standard convolutional operators and the inability of fixed receptive fields to reconstruct features corrupted by atmospheric interference. To address these bottlenecks, we propose SELPNet (Scene Exploration and Large Format Perception Network), a unified framework that integrates geometric alignment and multi-scale contextual perception into the YOLOv13 head. SELPNet consists of two key modules: (1) The Scene Exploration Convolution (SEC) leverages affine Lie group theory to construct a discrete manifold of rotation and scale transformations, actively probing multiple geometric views and selecting the most coherent response via a Maxout mechanism. (2) The Large Format Perception Module (LPM) introduces a dynamic dilation strategy with depthwise separable convolutions, progressively enlarging the receptive field from fine-grained edge preservation to scene-level contextual perception for semantic completion of degraded regions. We further construct and release AWU-OBB, a large-scale benchmark containing over 18,000 oriented bounding box-annotated UAV images across four representative scene categories. Ablation experiments demonstrate that SEC and LPM yield complementary gains, achieving a combined improvement of +4.26% mAP50 over the YOLOv13-n baseline with only 0.11 M additional parameters and 0.2 extra GFLOPs. The source code will be publicly released upon acceptance of this paper. Full article
(This article belongs to the Section Intelligent Sensors)
16 pages, 13549 KB  
Article
YOLO-ALD: An Efficient and Robust Lightweight Model for Apple Leaf Disease Detection in Complex Orchard Environments
by Lei Liu, Yinyin Li, Qingyu Liu, Huihui Sun, Yeguo Sun and Xiaobo Shen
Horticulturae 2026, 12(5), 550; https://doi.org/10.3390/horticulturae12050550 - 30 Apr 2026
Abstract
Real-time detection of apple leaf diseases in orchard environments faces ongoing challenges, particularly in preserving fine-grained disease features with limited computing resources. To address these issues, we propose a high-precision lightweight model based on YOLOv10n, called YOLO-ALD. First, we introduce Spatial and Channel [...] Read more.
Real-time detection of apple leaf diseases in orchard environments faces ongoing challenges, particularly in preserving fine-grained disease features with limited computing resources. To address these issues, we propose a high-precision lightweight model based on YOLOv10n, called YOLO-ALD. First, we introduce Spatial and Channel Reconstruction Convolution into deeper backbone networks to replace standard downsampling layers and convolutions. This suppresses spatial and channel redundancy caused by environmental noise and optimizes feature representation. Second, we design a new C2f-Faster-SimAM module for the neck network. This module combines the inference efficiency of FasterNet with a parameter-free 3D attention mechanism to adaptively focus on early lesions, effectively distinguishing them from leaf veins without increasing model complexity. Third, in the detection head section, we use the Focaler-ShapeIoU loss function to optimize bounding box regression. It utilizes a dynamic focusing mechanism and geometric constraints to ensure the localization accuracy of irregular shapes and hard-to-detect samples. Experimental results on our self-built dataset covering four specific diseases and healthy leaves showed that, compared with YOLOv10n, the mAP@0.5 of YOLO-ALD reached 92.1%, achieving a 2.1% increase. In addition, the model has an inference speed of 105 FPS, with only 2.1 M parameters and 5.6 GFLOPs. Therefore, YOLO-ALD achieves a good balance between efficiency and robustness, showing strong theoretical potential for resource-constrained mobile agriculture diagnosis. Full article
(This article belongs to the Special Issue Emerging Technologies in Smart Agriculture)
Show Figures

Figure 1

24 pages, 4665 KB  
Article
Human Fall Detection with Infrared Imaging: A Comparison of Graph Convolutional Networks and YOLO
by Karol Perliński, Artur Faltyński and Aleksandra Świetlicka
Sensors 2026, 26(9), 2794; https://doi.org/10.3390/s26092794 - 30 Apr 2026
Abstract
This paper presents a comparative study of two artificial intelligence approaches—graph convolutional networks (GCNs) and the YOLO object detection algorithm—for analyzing human fall events using infrared imaging. From the AI perspective, the study introduces a GCN model that achieves over 99% classification accuracy [...] Read more.
This paper presents a comparative study of two artificial intelligence approaches—graph convolutional networks (GCNs) and the YOLO object detection algorithm—for analyzing human fall events using infrared imaging. From the AI perspective, the study introduces a GCN model that achieves over 99% classification accuracy by modeling 2D and 3D skeletal data as graph structures and evaluates the real-time detection capabilities of YOLOv8 on infrared video frames. On the engineering side, the research addresses practical challenges in elderly care and healthcare monitoring systems by demonstrating how these AI methods can accurately detect and classify fall directions under infrared conditions. The results highlight each model’s strengths and propose a hybrid framework combining YOLO’s spatial localization with GCN’s motion-pattern analysis for future real-world applications. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Back to TopTop