Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (166)

Search Parameters:
Keywords = SPPF

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 9448 KB  
Article
SeaLSOD-YOLO: A Lightweight Framework for Maritime Small Object Detection Using YOLOv11
by Jinjia Ruan, Jin He and Yao Tong
Sensors 2026, 26(7), 2017; https://doi.org/10.3390/s26072017 - 24 Mar 2026
Viewed by 283
Abstract
Maritime small object detection is critical for UAV-based sea surveillance but remains challenging due to the small size of targets and interference from sea reflections and waves. This paper proposes SeaLSOD-YOLO, a lightweight detection algorithm based on YOLOv11, designed to improve small object [...] Read more.
Maritime small object detection is critical for UAV-based sea surveillance but remains challenging due to the small size of targets and interference from sea reflections and waves. This paper proposes SeaLSOD-YOLO, a lightweight detection algorithm based on YOLOv11, designed to improve small object detection accuracy while maintaining real-time performance. The method incorporates four key modules: Shallow Multi-scale Output Reconstruction, which fuses shallow and mid-level features to preserve fine-grained details; SPPF-FD, which combines spatial pyramid pooling with frequency-domain adaptive convolution to enhance sensitivity to high-frequency textures and suppress sea-surface interference; attention-based feature fusion, which emphasizes small object features through channel and spatial attention; and dynamic multi-scale sampling, which optimizes feature representation across different scales. Experiments on the SeaDroneSee dataset demonstrate that, compared with YOLOv11s, the proposed method improves precision from 75.6% to 81.9%, recall from 62.6% to 73.5%, and mAP@0.5 from 67.9% to 77.0%. The mAP@0.5:0.95 also increases from 41.1% to 44.9%. The model achieves an inference speed of 256 FPS. Although the parameter size increases from 18.2 MB to 30.8 MB, the method maintains a favorable balance between detection accuracy and computational efficiency. Comparative evaluation further shows superior performance in detecting small maritime objects such as buoys and lifeboats. These results indicate that SeaLSOD-YOLO effectively balances accuracy, efficiency, and real-time capability in complex maritime environments. Future work will focus on further optimization of attention mechanisms and upsampling strategies to enhance the detection of extremely small targets. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

11 pages, 2065 KB  
Article
Detection of Estrus in Dairy Cows Based on CE-YOLO
by Junjie Zhao, Huijing Zhang and Lei Liu
Electronics 2026, 15(6), 1269; https://doi.org/10.3390/electronics15061269 - 18 Mar 2026
Viewed by 190
Abstract
Accurate estrus detection is essential for dairy farm productivity, yet traditional manual and wearable methods remain limited by high labor costs, delayed responses, and animal stress. To address these challenges, we propose CE-YOLO, a lightweight YOLOv11n-based vision model tailored for edge deployment, which [...] Read more.
Accurate estrus detection is essential for dairy farm productivity, yet traditional manual and wearable methods remain limited by high labor costs, delayed responses, and animal stress. To address these challenges, we propose CE-YOLO, a lightweight YOLOv11n-based vision model tailored for edge deployment, which detects mounting behavior by integrating a Channel-Aware Downsampling (CA-Down) module to preserve small-scale features, a SimSPPF module for efficient contextual fusion, and a DySample module for dynamic spatial alignment. Experiments on a curated estrus behavior dataset demonstrate that CE-YOLO achieves a precision of 94.9% and an mAP50 of 98.2%, significantly outperforming the baseline by 3.9% and 4.6% respectively. These results validate the model as an efficient, non-intrusive solution for real-time estrus monitoring, strongly supporting the advancement of smart livestock management. Full article
(This article belongs to the Special Issue Advances in Imaging Technologies for Precision Agriculture)
Show Figures

Figure 1

0 pages, 4366 KB  
Article
Intelligent Detection of Asphalt Pavement Cracks Based on Improved YOLOv8s
by Jinfei Su, Jicong Xu, Chuqiao Shi, Yuhan Wang, Shihao Dong and Xue Zhang
Coatings 2026, 16(3), 359; https://doi.org/10.3390/coatings16030359 - 12 Mar 2026
Viewed by 306
Abstract
The intelligent detection of asphalt pavement cracks has become increasingly important for ensuring service performance of road infrastructure. Traditional manual detection has significant safety hazards and insufficient accuracy. Furthermore, existing deep learning models still face challenges, including missed detection, false alarms, and poor [...] Read more.
The intelligent detection of asphalt pavement cracks has become increasingly important for ensuring service performance of road infrastructure. Traditional manual detection has significant safety hazards and insufficient accuracy. Furthermore, existing deep learning models still face challenges, including missed detection, false alarms, and poor performance in small target detection under complex conditions. This investigation adopts unmanned aerial vehicles (UAVs) to acquire pavement distress information and develops an intelligent detection approach for asphalt pavement crack based on improved YOLOv8s. First, the Spatial Pyramid Pooling Fast (SPPF) module is replaced with the Spatial Pyramid Pooling Fast with Cross Stage Partial Connections (SPPFCSPC) module in the backbone network to enhance the multi-scale feature fusion capability. Secondly, the Convolutional Block Attention Module (CBAM) module is introduced to the neck network to optimize the feature weights in both channel and spatial attention. Meanwhile, the Efficient Intersection over Union (EIoU) loss is adopted to improve accuracy. Finally, the Crack_Dataset is established, and the ablation experiments are conducted to verify the reliability of the detection model. The research indicates that the improved model achieves Precision, Recall, and mAP@0.5 of 83.9%, 79.6%, and 83.9%, respectively, representing increases of 1.5%, 1.3%, and 1.4%, compared with the baseline model. In comparison with mainstream object detection algorithms such as YOLOv5s and YOLOv8s, the proposed method attains an F1-score, mAP@0.5, and mAP@[0.5–0.95] of 0.82, 83.9%, and 46.6%, respectively, demonstrating a performance improvement. Based on the improved detection model, a pavement crack detection system was designed and implemented using PyQt5. This system supports image, video, and real-time camera input and detection. Full article
(This article belongs to the Special Issue Pavement Surface Status Evaluation and Smart Perception)
Show Figures

Figure 1

17 pages, 4021 KB  
Article
Dangerous Goods Detection in X-Ray Security Inspection Images Based on Improved YOLOv8-seg
by Ting Wang, Pengfei Yuan and Aili Wang
Electronics 2026, 15(5), 1112; https://doi.org/10.3390/electronics15051112 - 7 Mar 2026
Viewed by 300
Abstract
In X-ray security inspection imagery, hazardous object detection is challenged by severe object overlap/occlusion, ambiguous boundaries of small objects, and complex texture representations caused by material diversity. Although YOLOv8-seg provides real-time instance segmentation capability, it still has clear limitations in this application scenario. [...] Read more.
In X-ray security inspection imagery, hazardous object detection is challenged by severe object overlap/occlusion, ambiguous boundaries of small objects, and complex texture representations caused by material diversity. Although YOLOv8-seg provides real-time instance segmentation capability, it still has clear limitations in this application scenario. Specifically, the original SPPF module has limited ability to model long-range spatial dependencies, making it difficult to accurately separate boundaries of densely overlapped objects, while the C2f module is insufficient for multi-scale feature parsing of hazardous items with diverse sizes and materials and introduces feature redundancy, which degrades segmentation accuracy in occluded scenes. To address these issues, this paper proposes an improved YOLOv8-seg framework for X-ray hazardous object detection, termed LM-YOLOv8. For feature enhancement, an SPPF-LSKA module is constructed by integrating large-kernel separable attention with dynamic receptive-field adjustment, thereby improving global contextual modeling and alleviating boundary ambiguity. For multi-scale feature fusion, a C2f-MSC module is designed by combining multi-branch dilated convolutions with the C2f structure to enhance complex contour parsing and cross-scale feature interaction. Experiments on the PIDray dataset show that the proposed method achieves 84.8% mAP50 in instance segmentation, representing an improvement of approximately 4.0 percentage points over the baseline YOLOv8-seg. In addition, the method demonstrates stronger robustness on challenging hard/hidden subsets, validating its effectiveness for X-ray security inspection hazardous object detection. Full article
(This article belongs to the Special Issue Image Processing, Target Tracking and Recognition System Design)
Show Figures

Figure 1

20 pages, 5981 KB  
Article
YOLO11-MSCAM UAV Remote Sensing-Based Detection of Illegal Rare-Earth Mining with Multi-Scale Convolution and Attention Module
by Hengkai Li, Yingming Cai, Shengdong Nie and Kunming Liu
Remote Sens. 2026, 18(5), 738; https://doi.org/10.3390/rs18050738 - 28 Feb 2026
Viewed by 298
Abstract
Ion-adsorption rare-earth mining in southern China often leaves small, fragmented disturbances in rugged, forested terrain, making UAV-based enforcement challenging due to confusion with bare ground, canopy gaps, and shadows. We propose YOLO11-MSCAM, an enhanced YOLO11vm detector in which the original SPPF at the [...] Read more.
Ion-adsorption rare-earth mining in southern China often leaves small, fragmented disturbances in rugged, forested terrain, making UAV-based enforcement challenging due to confusion with bare ground, canopy gaps, and shadows. We propose YOLO11-MSCAM, an enhanced YOLO11vm detector in which the original SPPF at the backbone–neck junction is replaced by a Multi-Scale Convolution–Attention Module that cascades channel attention, spatial attention, and multi-scale residual convolutions to enhance context aggregation and suppress background clutter. We build a field-acquired UAV dataset, SIMA (0.05 m GSD; September–November 2023), generating 1630 non-overlapping 640 × 640 orthomosaic tiles split into 1320/147/163 for training/validation/testing; five-lens raw images (nadir + oblique) are additionally used as auxiliary training samples and for post-detection verification. On the test set, YOLO11-MSCAM achieves mAP@0.5 = 83.24%, mAP@0.5:0.95 = 58.29%, and F1 = 79.92%, outperforming YOLOv11m and other detectors (YOLOv5m/6m/8m/9m/10m and Faster R-CNN with ResNet-50). With 19.67 M parameters, 67.34 GFLOPs@640, and 45.86 FPS, it supports tile-based batch screening to prioritize suspicious sites for field checks and evidence collection. Full article
Show Figures

Figure 1

25 pages, 22841 KB  
Article
SEAF-Net: A Sustainable and Lightweight Attention-Enhanced Detection Network for Underwater Fish Species Recognition
by Yu-Shan Han, Sheng-Lun Zhao, Chu Chen, Kangning Cui, Pingfan Hu and Rui-Feng Wang
J. Mar. Sci. Eng. 2026, 14(4), 351; https://doi.org/10.3390/jmse14040351 - 12 Feb 2026
Viewed by 502
Abstract
This study presents SEAF-Net, a lightweight and efficient detection network designed for low-contrast and highly dynamic underwater environments. Built upon YOLOv11n, SEAF-Net introduces three complementary structural enhancements: (1) Omni-Dimensional Dynamic Convolution (ODConv) to improve adaptive modeling of multi-scale and directional texture variations; (2) [...] Read more.
This study presents SEAF-Net, a lightweight and efficient detection network designed for low-contrast and highly dynamic underwater environments. Built upon YOLOv11n, SEAF-Net introduces three complementary structural enhancements: (1) Omni-Dimensional Dynamic Convolution (ODConv) to improve adaptive modeling of multi-scale and directional texture variations; (2) SimA-SPPF, which embeds the SimAM attention mechanism into the SPPF module to enable neuron-level saliency reweighting and effective suppression of complex background interference; and (3) GhostC3k2 to reduce redundant computation while preserving sufficient representational capacity. Evaluated on a standardized 13-class underwater fish dataset under a unified training and evaluation protocol, SEAF-Net achieves 6.1 GFLOPs, 92.683% Precision, 88.459% Recall, 93.333% mAP50, 73.445% mAP, and a 90.522% F1-score. Compared with the YOLOv11n baseline, SEAF-Net improves F1-score and Recall by 0.510% and 0.575%, respectively, while reducing computational cost by approximately 6%, demonstrating a favorable accuracy–efficiency trade-off under lightweight constraints. Ablation results further confirm that SimA-SPPF plays a dominant role in background suppression, ODConv consistently enhances deformation and directional texture modeling, and GhostC3k2 effectively controls computational overhead without degrading detection accuracy. To assess deployment feasibility, additional test set evaluations were conducted under deployment-oriented conditions using resource-limited hardware, yielding an F1-score of 88.54%. This result confirms that the proposed model maintains stable detection performance and robustness beyond training and validation stages. Overall, SEAF-Net provides an effective balance of accuracy, efficiency, and robustness, offering practical support for low-carbon, scalable, and sustainable intelligent aquaculture monitoring and underwater ecological assessment in real-world environments. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

19 pages, 10980 KB  
Article
Landmine Press Kinematics Measured with an Enhanced YOLOv8 Model and Mathematical Modeling
by Rui Zhao, Rong Cong, Ruijie Zhou, Kelong Lin, Jianke Yang, Tongchun Kui, Jiajin Zhang, Ran Wang and Rou Dong
Sensors 2026, 26(4), 1161; https://doi.org/10.3390/s26041161 - 11 Feb 2026
Viewed by 313
Abstract
The landmine press is a reliable and valid test for assessing upper-body push strength. However, its application is constrained by the limitations of current mainstream monitoring technologies, such as linear position transducers (LPTs). These devices require physical attachment to the barbell, they rely [...] Read more.
The landmine press is a reliable and valid test for assessing upper-body push strength. However, its application is constrained by the limitations of current mainstream monitoring technologies, such as linear position transducers (LPTs). These devices require physical attachment to the barbell, they rely on proprietary software, and their measurement accuracy can degrade under high-load conditions due to sensor drift and electromechanical noise. To address these limitations, this study developed a markerless, non-contact, and vision-based system using an enhanced YOLOv8-OBB model and a mathematical modeling framework to measure four kinematic indicators during the concentric phase of the landmine press. By integrating a polarized self-attention mechanism, an improved C3k2 module, and an optimized SPPF structure, the system significantly enhanced detection accuracy and robustness for the small targets at both ends of the barbell, achieving an mAP@0.5 of 0.995 on the test set. A method comparison study was conducted against a widely used LPT device (GymAware) across four loads (20–35 kg) in 247 trials. The results showed strong correlations (r > 0.85) for peak velocity, mean velocity, peak power, and mean power. Although the vision-based method systematically overestimated velocity metrics, the bias was predictable. Notably, under the highest load (35 kg), where LPT limitations are pronounced, the vision system demonstrated comparative stability, suggesting its potential advantage in mitigating sensor-related errors. The findings demonstrate that this vision-based system offers a reliable and practical alternative for monitoring landmine press kinematics, suitable for both training and scientific research. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

24 pages, 69667 KB  
Article
YOLO-ELS: A Lightweight Cherry Tomato Maturity Detection Algorithm
by Zhimin Tong, Yu Zhou, Changhao Li, Changqing Cai and Lihong Rong
Appl. Sci. 2026, 16(2), 1043; https://doi.org/10.3390/app16021043 - 20 Jan 2026
Viewed by 411
Abstract
Within the domain of intelligent picking robotics, fruit recognition and positioning are essential. Challenging conditions such as varying light, occlusion, and limited edge-computing power compromise fruit maturity detection. To tackle these issues, this paper proposes a lightweight algorithm YOLO-ELS based on YOLOv8n. Specifically, [...] Read more.
Within the domain of intelligent picking robotics, fruit recognition and positioning are essential. Challenging conditions such as varying light, occlusion, and limited edge-computing power compromise fruit maturity detection. To tackle these issues, this paper proposes a lightweight algorithm YOLO-ELS based on YOLOv8n. Specifically, we reconstruct the backbone by replacing the bottlenecks in the C2f structure with Edge-Information-Enhanced Modules (EIEM) to prioritize morphological cues and filter background redundancy. Furthermore, a Large Separable Kernel Attention (LSKA) mechanism is integrated into the SPPF layer to expand the effective receptive field for multi-scale targets. To mitigate occlusion-induced errors, a Spatially Enhanced Attention Module (SEAM) is incorporated into the decoupled detection head to enhance feature responses in obscured regions. Finally, the Inner-GIoU loss is adopted to refine bounding box regression and accelerate convergence. Experimental results demonstrate that compared to the YOLOv8n baseline, the proposed YOLO-ELS achieves a 14.8% reduction in GFLOPs and a 2.3% decrease in parameters, while attaining a precision, recall, and mAP@50% of 92.7%, 83.9%, and 92.0%, respectively. When compared with mainstream models such as DETR, Faster-RCNN, SSD, TOOD, YOLOv5s, and YOLO11n, the mAP@50% is improved by 7.0%, 4.7%, 11.4%, 8.6%, 3.1%, and 3.2%. Deployment tests on the NVIDIA Jetson Orin Nano Super edge platform yield an inference latency of 25.2 ms and a detection speed of 28.2 FPS, successfully meeting the real-time operational requirements of automated harvesting systems. These findings confirm that YOLO-ELS effectively balances high detection accuracy with lightweight architecture, providing a robust technical foundation for intelligent fruit picking in resource-constrained greenhouse environments. Full article
(This article belongs to the Section Agricultural Science and Technology)
Show Figures

Figure 1

18 pages, 4205 KB  
Article
Research on Field Weed Target Detection Algorithm Based on Deep Learning
by Ziyang Chen, Le Wu, Zhenhong Jia, Jiajia Wang, Gang Zhou and Zhensen Zhang
Sensors 2026, 26(2), 677; https://doi.org/10.3390/s26020677 - 20 Jan 2026
Viewed by 364
Abstract
Weed detection algorithms based on deep learning are considered crucial for smart agriculture, with the YOLO series algorithms being widely adopted due to their efficiency. However, existing YOLO algorithms struggle to maintain high accuracy, while low parameter requirements and computational efficiency are achieved [...] Read more.
Weed detection algorithms based on deep learning are considered crucial for smart agriculture, with the YOLO series algorithms being widely adopted due to their efficiency. However, existing YOLO algorithms struggle to maintain high accuracy, while low parameter requirements and computational efficiency are achieved when weeds with occlusion or overlap are detected. To address this challenge, a target detection algorithm called SSS-YOLO based on YOLOv9t is proposed in this paper. First, the SCB (Spatial Channel Conv Block) module is introduced, in which large kernel convolution is employed to capture long-range dependencies, occluded weed regions are bypassed by being associated with unobstructed areas, and features of unobstructed regions are enhanced through inter-channel relationships. Second, the SPPF EGAS (Spatial Pyramid Pooling Fast Edge Gaussian Aggregation Super) module is proposed, where multi-scale max pooling is utilized to extract hierarchical contextual features, large receptive fields are leveraged to acquire background information around occluded objects, and features of weed regions obscured by crops are inferred. Finally, the EMSN (Efficient Multi-Scale Spatial-Feedforward Network) module is developed, through which semantic information of occluded regions is reconstructed by contextual reasoning and background vegetation interference is effectively suppressed while visible regional details are preserved. To validate the performance of this method, experiments are conducted on both our self-built dataset and the publicly available Cotton WeedDet12 dataset. The results demonstrate that compared to existing algorithms, significant performance improvements are achieved by the proposed method. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

16 pages, 6142 KB  
Article
Research on Image Detection of Thin-Vein Precious Metal Ores and Rocks Based on Improved YOLOv8n
by Heyan Zhou, Yuanhui Li, Yunsen Wang, Hong Zhou and Kunmeng Li
Appl. Sci. 2026, 16(2), 988; https://doi.org/10.3390/app16020988 - 19 Jan 2026
Viewed by 299
Abstract
To address the high-dilution issues arising from efficient mining methods such as medium-deep drilling for underground thin veins of precious metals, detecting raw rock fragments after blasting for subsequent sorting has become a cutting-edge research focus. With the continuous advancement of artificial intelligence, [...] Read more.
To address the high-dilution issues arising from efficient mining methods such as medium-deep drilling for underground thin veins of precious metals, detecting raw rock fragments after blasting for subsequent sorting has become a cutting-edge research focus. With the continuous advancement of artificial intelligence, deep learning offers novel applications for rock detection. Accordingly, this study employs an improved lightweight YOLOv8n model to detect two typical thin-vein precious metal ores: gold ore and wolframite. In consideration of the computational resource constraints in underground environments, a triple optimization strategy is proposed. First, GhostConv and C2f-Ghost modules were introduced into the backbone network to reduce redundant computations while preserving feature representation capabilities. Second, the VoVGSCSP module was incorporated into the neck to further decrease model parameters and computational load. Finally, the ECA mechanism was embedded before the SPPF pooling layer to enhance feature extraction for ores and rocks, thereby improving detection accuracy. The results demonstrate that the GVE-YOLOv8 model contains only 2.28 million parameters—a 24.3% reduction compared to the original YOLOv8n. FLOPs decrease from 8.1 G to 5.6 G, and the model size reduces from 6.3 MB to 4.9 MB, while detection accuracy improves to 98.3% mAP50 and 95.3% mAP50-95. This enhanced model meets the performance requirements for accurately detecting raw ore and rock fragments after underground blasting, thereby providing a novel research method for thin-vein mining. Full article
Show Figures

Figure 1

25 pages, 9860 KB  
Article
Symmetry-Aware SXA-YOLO: Enhancing Tomato Leaf Disease Recognition with Bidirectional Feature Fusion and Task Decoupling
by Guangyue Du, Shuyu Fang, Lianbin Zhang, Wanlu Ren and Biao He
Symmetry 2026, 18(1), 178; https://doi.org/10.3390/sym18010178 - 18 Jan 2026
Viewed by 354
Abstract
Tomatoes are an important economic crop in China, and crop diseases often lead to a decline in their yield. Deep learning-based visual recognition methods have become an approach for disease identification; however, challenges remain due to complex background interference in the field and [...] Read more.
Tomatoes are an important economic crop in China, and crop diseases often lead to a decline in their yield. Deep learning-based visual recognition methods have become an approach for disease identification; however, challenges remain due to complex background interference in the field and the diversity of disease manifestations. To address these issues, this paper proposes the SXA-YOLO (an improvement based on YOLO, where S stands for the SAAPAN architecture, X represents the XIoU loss function, and A denotes the AsDDet module) symmetric perception recognition model. First, a comprehensive symmetry architecture system is established. The backbone network creates a hierarchical feature foundation through C3k2 (Cross-stage Partial Concatenated Bottleneck Convolution with Dual-kernel Design) and SPPF (the Fast Pyramid Pooling module) modules; the neck employs a SAAPAN (Symmetry-Aware Adaptive Path Aggregation Architecture) bidirectional feature pyramid architecture, utilizing multiple modules to achieve equal fusion of multi-scale features; and the detection head is based on the AsDDet (Adaptive Symmetry-aware Decoupled Detection Head) module for functional decoupling, combining dynamic label assignment and the XIoU (Extended Intersection over Union) loss function to collaboratively optimize classification, regression, and confidence prediction. Ultimately, a complete recognition framework is formed through triple symmetric optimization of “feature hierarchy, fusion path, and task functionality.” Experimental results indicate that this method effectively enhances the model’s recognition performance, achieving a P (Precision) value of 0.992 and an mAP50 (mean Average Precision at 50% IoU threshold) of 0.993. Furthermore, for ten categories of diseases, the SXA-YOLO symmetric perception recognition model outperforms other comparative models in both p value and mAP50. The improved algorithm enhances the recognition of foliar diseases in tomatoes, achieving a high level of accuracy. Full article
Show Figures

Figure 1

28 pages, 3390 KB  
Article
SDC-YOLOv8: An Improved Algorithm for Road Defect Detection Through Attention-Enhanced Feature Learning and Adaptive Feature Reconstruction
by Hao Yang, Yulong Song, Yue Liang, Enhao Tang and Danyang Cao
Sensors 2026, 26(2), 609; https://doi.org/10.3390/s26020609 - 16 Jan 2026
Viewed by 577
Abstract
Road defect detection is essential for timely road damage repair and traffic safety assurance. However, existing object detection algorithms suffer from insufficient accuracy in detecting small road surface defects and are prone to missed detections and false alarms under complex lighting and background [...] Read more.
Road defect detection is essential for timely road damage repair and traffic safety assurance. However, existing object detection algorithms suffer from insufficient accuracy in detecting small road surface defects and are prone to missed detections and false alarms under complex lighting and background conditions. To address these challenges, this study proposes SDC-YOLOv8, an improved YOLOv8-based algorithm for road defect detection that employs attention-enhanced feature learning and adaptive feature reconstruction. The model incorporates three key innovations: (1) an SPPF-LSKA module that integrates Fast Spatial Pyramid Pooling with Large Separable Kernel Attention to enhance multi-scale feature representation and irregular defect modeling capabilities; (2) DySample dynamic upsampling that replaces conventional interpolation methods for adaptive feature reconstruction with reduced computational cost; and (3) a Coordinate Attention module strategically inserted to improve spatial localization accuracy under complex conditions. Comprehensive experiments on a public pothole dataset demonstrate that SDC-YOLOv8 achieves 78.0% mAP@0.5, 81.0% Precision, and 70.7% Recall while maintaining real-time performance at 85 FPS. Compared to the baseline YOLOv8n model, the proposed method improves mAP@0.5 by 2.0 percentage points, Precision by 3.3 percentage points, and Recall by 1.8 percentage points, yielding an F1 score of 75.5%. These results demonstrate that SDC-YOLOv8 effectively enhances small-target detection accuracy while preserving real-time processing capability, offering a practical and efficient solution for intelligent road defect detection applications. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

23 pages, 91075 KB  
Article
Improved Lightweight Marine Oil Spill Detection Using the YOLOv8 Algorithm
by Jianting Shi, Tianyu Jiao, Daniel P. Ames, Yinan Chen and Zhonghua Xie
Appl. Sci. 2026, 16(2), 780; https://doi.org/10.3390/app16020780 - 12 Jan 2026
Viewed by 458
Abstract
Marine oil spill detection using Synthetic Aperture Radar (SAR) is crucial but challenged by dynamic marine conditions, diverse spill scales, and limitations in existing algorithms regarding model size and real-time performance. To address these challenges, we propose LSFE-YOLO, a YOLOv8s-optimized (You Only Look [...] Read more.
Marine oil spill detection using Synthetic Aperture Radar (SAR) is crucial but challenged by dynamic marine conditions, diverse spill scales, and limitations in existing algorithms regarding model size and real-time performance. To address these challenges, we propose LSFE-YOLO, a YOLOv8s-optimized (You Only Look Once version 8) lightweight model with an original, domain-tailored synergistic integration of FasterNet, GN-LSC Head (GroupNorm Lightweight Shared Convolution Head), and C2f_MBE (C2f Mobile Bottleneck Enhanced). FasterNet serves as the backbone (25% neck width reduction), leveraging partial convolution (PConv) to minimize memory access and redundant computations—overcoming traditional lightweight backbones’ high memory overhead—laying the foundation for real-time deployment while preserving feature extraction. The proposed GN-LSC Head replaces YOLOv8’s decoupled head: its shared convolutions reduce parameter redundancy by approximately 40%, and GroupNorm (Group Normalization) ensures stable accuracy under edge computing’s small-batch constraints, outperforming BatchNorm (Batch Normalization) in resource-limited scenarios. The C2f_MBE module integrates EffectiveSE (Effective Squeeze and Excitation)-optimized MBConv (Mobile Inverted Bottleneck Convolution) into C2f: MBConv’s inverted-residual design enhances multi-scale feature capture, while lightweight EffectiveSE strengthens discriminative oil spill features without extra computation, addressing the original C2f’s scale variability insufficiency. Additionally, an SE (Squeeze and Excitation) attention mechanism embedded upstream of SPPF (Spatial Pyramid Pooling Fast) suppresses background interference (e.g., waves, biological oil films), synergizing with FasterNet and C2f_MBE to form a cascaded feature optimization pipeline that refines representations throughout the model. Experimental results show that LSFE-YOLO improves mAP (mean Average Precision) by 1.3% and F1 score by 1.7% over YOLOv8s, while achieving substantial reductions in model size (81.9%), parameter count (82.9%), and computational cost (84.2%), alongside a 20 FPS (Frames Per Second) increase in detection speed. LSFE-YOLO offers an efficient and effective solution for real-time marine oil spill detection. Full article
Show Figures

Figure 1

26 pages, 4991 KB  
Article
GBDR-Net: A YOLOv10-Derived Lightweight Model with Multi-Scale Feature Fusion for Accurate, Real-Time Detection of Grape Berry Diseases
by Pan Li, Jitao Zhou, Huihui Sun, Penglin Li and Xi Chen
Horticulturae 2026, 12(1), 38; https://doi.org/10.3390/horticulturae12010038 - 28 Dec 2025
Viewed by 531
Abstract
Grape berries are highly susceptible to diseases during growth and harvest, which severely impacts yield and postharvest quality. While rapid and accurate disease detection is essential for real-time control and optimized management, it remains challenging due to complex symptom patterns, occlusions in dense [...] Read more.
Grape berries are highly susceptible to diseases during growth and harvest, which severely impacts yield and postharvest quality. While rapid and accurate disease detection is essential for real-time control and optimized management, it remains challenging due to complex symptom patterns, occlusions in dense clusters, and orchard environmental interference. Although deep learning presents a viable solution, robust methods specifically for detecting grape berry diseases under dense clustering conditions are still lacking. To bridge this gap, we propose GBDR-Net—a high-accuracy, lightweight, and deployable model based on YOLOv10. It incorporates four key enhancements: (1) an SDF-Fusion module replaces the original C2f module in deeper backbone layers to improve global context and subtle lesion feature extraction; (2) an additional Detect-XSmall head is integrated at the neck, with cross-concatenated outputs from SPPF and PSA modules, to enhance sensitivity to small disease spots; (3) the nearest-neighbor upsampling is substituted with a lightweight content-aware feature reassembly operator (LCFR-Op) for efficient and semantically aligned multi-scale feature enhancement; and (4) the conventional bounding box loss function is replaced with Inner-SIoU loss to accelerate convergence and improve localization accuracy. Evaluated on the Grape Berry Disease Visual Analysis (GBDVA) dataset, GBDR-Net achieves a precision of 93.4%, recall of 89.6%, mAP@0.5 of 90.2%, and mAP@0.5:0.95 of 86.4%, with a model size of only 4.83 MB, computational cost of 20.5 GFLOPs, and a real-time inference speed of 98.2 FPS. It outperforms models such as Faster R-CNN, SSD, YOLOv6s, and YOLOv8s across key metrics, effectively balancing detection accuracy with computational efficiency. This work provides a reliable technical solution for the intelligent monitoring of grape berry diseases in horticultural production. The proposed lightweight architecture and its design focus on dense, small-target detection offer a valuable framework that could inform the development of similar systems for other cluster-growing fruits and vegetables. Full article
(This article belongs to the Section Viticulture)
Show Figures

Figure 1

22 pages, 3870 KB  
Article
Accurate Pose Detection Method for Rail Fastener Clips Based on Improved YOLOv8-Pose
by Defang Lv, Jianjun Meng, Zhenhan Ren, Liqing Yao and Gengqi Liu
Appl. Sci. 2026, 16(1), 276; https://doi.org/10.3390/app16010276 - 26 Dec 2025
Viewed by 304
Abstract
Minor displacements and deflections of rail fastener clips pose a critical risk to railway safety, which are difficult to quantify accurately using traditional object detection methods. This paper proposes an improved YOLOv8-pose-based method, You Only Look Once version 8-pose with GAM, SPPF-Attention, and [...] Read more.
Minor displacements and deflections of rail fastener clips pose a critical risk to railway safety, which are difficult to quantify accurately using traditional object detection methods. This paper proposes an improved YOLOv8-pose-based method, You Only Look Once version 8-pose with GAM, SPPF-Attention, and Wise-IoU (YOLOv8-pose-GSW) for automated and quantitative pose detection of fastener clips. Firstly, a high-precision keypoint detection network is constructed by integrating a Global Attention Mechanism (GAM) into the neck, enhancing the Spatial Pyramid Pooling Fast (SPPF) module to Spatial Pyramid Pooling Fast with Attention (SPPF-Attention) in the backbone, and adopting the Wise Intersection over Union (Wise-IoU) loss function. Subsequently, a posterior verification mechanism based on spatial constraint error is designed to eliminate unreliable detections by leveraging the inherent geometric priors of fasteners. Finally, the deflection angle, longitudinal displacement, and lateral displacement of the clip are calculated from the verified keypoints. Experimental results demonstrate that the proposed method achieves an Average Precision at IoU threshold from 0.5 to 0.95 (AP@0.5:0.95) of 77.5%, representing a 3.6% improvement over the baseline YOLOv8s-pose model, effectively balancing detection accuracy and computational efficiency. This work provides a reliable technical solution for the refined maintenance of rail fasteners. Full article
Show Figures

Figure 1

Back to TopTop