Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,786)

Search Parameters:
Keywords = Yolo11

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
31 pages, 12246 KB  
Article
DVIF-Net: A Small-Target Detection Network for UAV Aerial Images Based on Visible and Infrared Fusion
by Xiaofeng Zhao, Hui Zhang, Chenxiao Li, Kehao Wang and Zhili Zhang
Remote Sens. 2025, 17(20), 3411; https://doi.org/10.3390/rs17203411 (registering DOI) - 11 Oct 2025
Abstract
During UAV aerial photography tasks, influenced by flight altitude and imaging mechanisms, the target in images often exhibits characteristics such as small size, complex backgrounds, and small inter-class differences. Under single optical modality, the weak and less discriminative feature representation of targets in [...] Read more.
During UAV aerial photography tasks, influenced by flight altitude and imaging mechanisms, the target in images often exhibits characteristics such as small size, complex backgrounds, and small inter-class differences. Under single optical modality, the weak and less discriminative feature representation of targets in drone-captured images makes them easily overwhelmed by complex background noise, leading to low detection accuracy, high missed-detection and false-detection rates in current object detection networks. Moreover, such methods struggle to meet all-weather and all-scenario application requirements. To address these issues, this paper proposes DVIF-Net, a visible-infrared fusion network for small-target detection in UAV aerial images, which leverages the complementary characteristics of visible and infrared images to enhance detection capability in complex environments. Firstly, a dual-branch feature extraction structure is designed based on YOLO architecture to separately extract features from visible and infrared images. Secondly, a P4-level cross-modal fusion strategy is proposed to effectively integrate features from both modalities while reducing computational complexity. Meanwhile, we design a novel dual context-guided fusion module to capture complementary features through channel attention of visible and infrared images during fusion and enhance interaction between modalities via element-wise multiplication. Finally, an edge information enhancement module based on cross stage partial structure is developed to improve sensitivity to small-target edges. Experimental results on two cross-modal datasets, DroneVehicle and VEDAI, demonstrate that DVIF-Net achieves detection accuracies of 85.8% and 62%, respectively. Compared with YOLOv10n, it has improved by 21.7% and 10.5% in visible modality, and by 7.4% and 30.5% in infrared modality, while maintaining a model parameter count of only 2.49 M. Furthermore, compared with 15 other algorithms, the proposed DVIF-Net attains SOTA performance. These results indicate that the method significantly enhances the detection capability for small targets in UAV aerial images, offering a high-precision and lightweight solution for real-time applications in complex aerial scenarios. Full article
25 pages, 4958 KB  
Article
YOLO-DPDG: A Dual-Pooling Dynamic Grouping Network for Small and Long-Distance Traffic Sign Detection
by Ruishi Liang, Minjie Jiang and Shuaibing Li
Appl. Sci. 2025, 15(20), 10921; https://doi.org/10.3390/app152010921 (registering DOI) - 11 Oct 2025
Abstract
Traffic sign detection is a crucial task for autonomous driving perception systems, as it directly impacts vehicle path planning and safety decisions. Existing algorithms face challenges such as feature information attenuation and model lightweighting requirements in the detection of small traffic signs at [...] Read more.
Traffic sign detection is a crucial task for autonomous driving perception systems, as it directly impacts vehicle path planning and safety decisions. Existing algorithms face challenges such as feature information attenuation and model lightweighting requirements in the detection of small traffic signs at long distances. To address these issues, this paper proposes a dual-pooling dynamic grouping (DPDG) module. This module dynamically adjusts the number of groups to adapt to different input features, combines global average pooling and max pooling to enhance channel attention representation, and uses a lightweight 3 × 3 convolution-based spatial branch to generate spatial weights. Based on a hierarchical optimization strategy, the DPDG module is integrated into the YOLOv10n network. Experimental results on the traffic sign dataset demonstrate a significant improvement in the performance of the YOLO-DPDG network: Compared to the baseline YOLOv10n model, mAP@0.5 and mAP@0.5:0.95 improved by 8.77% and 10.56%, respectively, while precision and recall were enhanced by 6.16% and 6.62%, respectively. Additionally, inference speed (FPS) increased by 11.1%, with only a 4.89% increase in model parameters. Compared to the YOLOv10-Small model, this method achieves a similar detection accuracy while reducing the number of model parameters by 64.83%. This study provides a more efficient and lightweight solution for edge-based traffic sign detection. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

18 pages, 9861 KB  
Article
EH-YOLO: Dimensional Transformation and Hierarchical Feature Fusion-Based PCB Surface Defect Detection
by Chengzhi Deng, You Zhang, Zhaoming Wu, Yingbo Wu, Xiaowei Sun and Shengqian Wang
Appl. Sci. 2025, 15(20), 10895; https://doi.org/10.3390/app152010895 - 10 Oct 2025
Abstract
Small surface defects in printed circuit boards (PCBs) severely affect the reliability of electronic devices, making PCB surface defect detection crucial for ensuring the quality of electronic products. However, the existing detection methods often struggle with insufficient accuracy and the inherent trade-off between [...] Read more.
Small surface defects in printed circuit boards (PCBs) severely affect the reliability of electronic devices, making PCB surface defect detection crucial for ensuring the quality of electronic products. However, the existing detection methods often struggle with insufficient accuracy and the inherent trade-off between detection precision and inference speed. To address these problems, we propose a novel ESDM-HNN-YOLO (EH-YOLO) network based on the improved YOLOv10 for efficient detection of small PCB defects. Firstly, an enhanced spatial-depth module (ESDM) is designed, which transforms spatial-dimensional features into depth-dimensional representations while integrating spatial attention module (SAM) and channel attention module (CAM) to highlight critical features. This dual mechanism not only effectively suppresses feature loss in micro-defects but also significantly enhances detection accuracy. Secondly, a hybrid neck network (HNN) is designed, which optimizes the speed–accuracy balance through hierarchical architecture. The hierarchical structure uses a computationally efficient weighted bidirectional feature pyramid network (BiFPN) to enhance multi-scale feature fusion of small objects in the shallow layer and uses a path aggregation network (PAN) to prevent feature loss in the deeper layer. Comprehensive evaluations on benchmark datasets (PCB_DATASET and DeepPCB) demonstrate the superior performance of EH-YOLO, achieving mAP@50-95 scores of 45.3% and 78.8% with inference speeds of 166.67 FPS and 158.73 FPS, respectively. These results significantly outperform existing approaches in both accuracy and processing efficiency. Full article
Show Figures

Figure 1

24 pages, 7207 KB  
Article
YOLO–LaserGalvo: A Vision–Laser-Ranging System for High-Precision Welding Torch Localization
by Jiajun Li, Tianlun Wang and Wei Wei
Sensors 2025, 25(20), 6279; https://doi.org/10.3390/s25206279 - 10 Oct 2025
Abstract
A novel closed loop visual positioning system, termed YOLO–LaserGalvo (YLGS), is proposed for precise localization of welding torch tips in industrial welding automation. The proposed system integrates a monocular camera, an infrared laser distance sensor with a galvanometer scanner, and a customized deep [...] Read more.
A novel closed loop visual positioning system, termed YOLO–LaserGalvo (YLGS), is proposed for precise localization of welding torch tips in industrial welding automation. The proposed system integrates a monocular camera, an infrared laser distance sensor with a galvanometer scanner, and a customized deep learning detector based on an improved YOLOv11 model. In operation, the vision subsystem first detects the approximate image location of the torch tip using the YOLOv11-based model. Guided by this detection, the galvanometer steers the IR laser beam to that point and measures the distance to the torch tip. The distance feedback is then fused with the vision coordinates to compute the precise 3D position of the torch tip in real-time. Under complex illumination, the proposed YLGS system exhibits superior robustness compared with color-marker and ArUco baselines. Experimental evaluation shows that the system outperforms traditional color-marker and ArUco-based methods in terms of accuracy, robustness, and processing speed. This marker-free method provides high-precision torch positioning without requiring structured lighting or artificial markers. Its pedagogical implications in engineering education are also discussed. Potential future work includes extending the method to full 6-DOF pose estimation and integrating additional sensors for enhanced performance. Full article
(This article belongs to the Section Navigation and Positioning)
Show Figures

Figure 1

21 pages, 771 KB  
Article
LLM-Driven Offloading Decisions for Edge Object Detection in Smart City Deployments
by Xingyu Yuan and He Li
Smart Cities 2025, 8(5), 169; https://doi.org/10.3390/smartcities8050169 - 10 Oct 2025
Abstract
Object detection is a critical technology for smart city development. As request volumes surge, inference is increasingly offloaded from centralized clouds to user-proximal edge sites to reduce latency and backhaul traffic. However, heterogeneous workloads, fluctuating bandwidth, and dynamic device capabilities make offloading and [...] Read more.
Object detection is a critical technology for smart city development. As request volumes surge, inference is increasingly offloaded from centralized clouds to user-proximal edge sites to reduce latency and backhaul traffic. However, heterogeneous workloads, fluctuating bandwidth, and dynamic device capabilities make offloading and scheduling difficult to optimize in edge environments. Deep reinforcement learning (DRL) has proved effective for this problem, but in practice, it relies on manually engineered reward functions that must be redesigned whenever service objectives change. To address this limitation, we introduce an LLM-driven framework that retargets DRL policies for edge object detection directly through natural language instructions. By leveraging understanding of the text and encoding capabilities of large language models (LLMs), our system (i) interprets the current optimization objective; (ii) generates an executable, environment-compatible reward function code; and (iii) iteratively refines the reward via closed-loop simulation feedback. In simulations for a real-world dataset, policies trained with LLM-generated rewards adapt from prompts alone and outperform counterparts trained with expert-designed rewards, while eliminating manual reward engineering. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

24 pages, 76400 KB  
Article
MBD-YOLO: An Improved Lightweight Multi-Scale Small-Object Detection Model for UAVs Based on YOLOv8
by Bo Xu, Di Cai, Kelin Sui, Zheng Wang, Chuangchuang Liu and Xiaolong Pei
Appl. Sci. 2025, 15(20), 10877; https://doi.org/10.3390/app152010877 - 10 Oct 2025
Abstract
To address the challenges of low detection accuracy and weak generalization in UAV aerial imagery caused by complex ground environments, significant scale variations among targets, dense small objects, and background interference, this paper proposes an improved lightweight multi-scale small-object detection model, MBD-YOLO (MBFF [...] Read more.
To address the challenges of low detection accuracy and weak generalization in UAV aerial imagery caused by complex ground environments, significant scale variations among targets, dense small objects, and background interference, this paper proposes an improved lightweight multi-scale small-object detection model, MBD-YOLO (MBFF module, BiMS-FPN, and Dual-Stream Head). Specifically, to enhance multi-scale feature extraction capabilities, we introduce the Multi-Branch Feature Fusion (MBFF) module, which dynamically adjusts receptive fields through parallel branches and adaptive depthwise convolutions, expanding the receptive field while preserving detail perception. We further design a lightweight Bidirectional Multi-Scale Feature Aggregation Pyramid Network (BiMS-FPN), integrating bidirectional propagation paths and a Multi-Scale Feature Aggregation (MSFA) module to mitigate feature spatial misalignment and improve small-target detection. Additionally, the Dual-Stream Head with NMS-free architecture leverages a task-aligned architecture and dynamic matching strategies to boost inference speed without compromising accuracy. Experiments on the VisDrone2019 dataset demonstrate that MBD-YOLO-n surpasses YOLOv8n by 6.3% in mAP50 and 8.2% in mAP50–95, with accuracy gains of 17.96–55.56% for several small-target categories, while increasing parameters by merely 3.1%. Moreover, MBD-YOLO-s achieves superior detection accuracy, efficiency, and generalization with only 12.1 million parameters, outperforming state-of-the-art models and proving suitable for resource-constrained embedded deployment scenarios. The superior performance of MBD-YOLO, which harmonizes high precision with low computational demand, fulfills the critical requirements for real-time deployment on resource-limited UAVs, showing great promise for applications in traffic monitoring, urban security, and agricultural surveying. Full article
Show Figures

Figure 1

23 pages, 2122 KB  
Article
PSD-YOLO: An Enhanced Real-Time Framework for Robust Worker Detection in Complex Offshore Oil Platform Environments
by Yikun Qin, Jiawen Dong, Wei Li, Linxin Zhang, Ke Feng and Zijia Wang
Sensors 2025, 25(20), 6264; https://doi.org/10.3390/s25206264 - 10 Oct 2025
Abstract
To address the safety challenges for personnel in the complex and hazardous environments of offshore drilling platforms, this paper introduces the Platform Safety Detection YOLO (PSD-YOLO), an enhanced, real-time object detection framework based on YOLOv10s. The framework integrates several key innovations to improve [...] Read more.
To address the safety challenges for personnel in the complex and hazardous environments of offshore drilling platforms, this paper introduces the Platform Safety Detection YOLO (PSD-YOLO), an enhanced, real-time object detection framework based on YOLOv10s. The framework integrates several key innovations to improve detection robustness: first, the Channel Attention-Aware (CAA) mechanism is incorporated into the backbone network to effectively suppress complex background noise interference; second, a novel C2fCIB_Conv2Former module is designed in the neck to strengthen multi-scale feature fusion for small and occluded targets; finally, the Soft-NMS algorithm is employed in place of traditional NMS to significantly reduce missed detections in dense scenes. Experimental results on a custom offshore platform personnel dataset show that PSD-YOLO achieves a mean Average Precision (mAP@0.5) of 82.5% at an inference speed of 232.56 FPS. The efficient and accurate detection framework proposed in this study provides reliable technical support for automated safety monitoring systems, holds significant practical implications for reducing accident rates and safeguarding personnel by enabling real-time warnings of hazardous situations, fills a critical gap in intelligent sensor monitoring for offshore platforms and makes a significant contribution to advancing their safety monitoring systems. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

20 pages, 2524 KB  
Article
YOLO-PFA: Advanced Multi-Scale Feature Fusion and Dynamic Alignment for SAR Ship Detection
by Shu Liu, Peixue Liu, Zhongxun Wang, Mingze Sun and Pengfei He
J. Mar. Sci. Eng. 2025, 13(10), 1936; https://doi.org/10.3390/jmse13101936 - 9 Oct 2025
Abstract
Maritime ship detection faces challenges due to complex object poses, variable target scales, and background interference. This paper introduces YOLO-PFA, a novel SAR ship detection model that integrates multi-scale feature fusion and dynamic alignment. By leveraging the Bidirectional Feature Pyramid Network (BiFPN), YOLO-PFA [...] Read more.
Maritime ship detection faces challenges due to complex object poses, variable target scales, and background interference. This paper introduces YOLO-PFA, a novel SAR ship detection model that integrates multi-scale feature fusion and dynamic alignment. By leveraging the Bidirectional Feature Pyramid Network (BiFPN), YOLO-PFA enhances cross-scale weighted feature fusion, improving detection of objects of varying sizes. The C2f-Partial Feature Aggregation (C2f-PFA) module aggregates raw and processed features, enhancing feature extraction efficiency. Furthermore, the Dynamic Alignment Detection Head (DADH) optimizes classification and regression feature interaction, enabling dynamic collaboration. Experimental results on the iVision-MRSSD dataset demonstrate YOLO-PFA’s superiority, achieving an mAP@0.5 of 95%, outperforming YOLOv11 by 1.2% and YOLOv12 by 2.8%. This paper contributes significantly to automated maritime target detection. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

17 pages, 3028 KB  
Article
YOLO-Based Object and Keypoint Detection for Autonomous Traffic Cone Placement and Retrieval for Industrial Robots
by János Hollósi
Appl. Sci. 2025, 15(19), 10845; https://doi.org/10.3390/app151910845 - 9 Oct 2025
Abstract
The accurate and efficient placement of traffic cones is a critical safety and logistical requirement in diverse industrial environments. This study introduces a novel dataset specifically designed for the near-overhead detection of traffic cones, containing both bounding box annotations and apex keypoints. Leveraging [...] Read more.
The accurate and efficient placement of traffic cones is a critical safety and logistical requirement in diverse industrial environments. This study introduces a novel dataset specifically designed for the near-overhead detection of traffic cones, containing both bounding box annotations and apex keypoints. Leveraging this dataset, we systematically evaluated whether classical object detection methods or keypoint-based detection methods are more effective for the task of cone apex localization. Several state-of-the-art YOLO-based architectures (YOLOv8, YOLOv11, YOLOv12) were trained and tested under identical conditions. The comparative experiments showed that both approaches can achieve high accuracy, but they differ in their trade-offs between robustness, computational cost, and suitability for real-time embedded deployment. These findings highlight the importance of dataset design for specialized viewpoints and confirm that lightweight YOLO models are particularly well-suited for resource-constrained robotic platforms. The key contributions of this work are the introduction of a new annotated dataset for overhead cone detection and a systematic comparison of object detection and keypoint detection paradigms for apex localization in real-world robotic applications. Full article
(This article belongs to the Special Issue Sustainable Mobility and Transportation (SMTS 2025))
Show Figures

Figure 1

19 pages, 24139 KB  
Article
EnhancedMulti-Scenario Pig Behavior Recognition Based on YOLOv8n
by Panqi Pu, Junge Wang, Geqi Yan, Hongchao Jiao, Hao Li and Hai Lin
Animals 2025, 15(19), 2927; https://doi.org/10.3390/ani15192927 - 9 Oct 2025
Abstract
Advances in smart animal husbandry necessitate efficient pig behavior monitoring, yet traditional approaches suffer from operational inefficiency and animal stress. We address these limitations through a lightweight YOLOv8n architecture enhanced with SPD-Conv for feature preservation during downsampling, LSKBlock attention for contextual feature fusion, [...] Read more.
Advances in smart animal husbandry necessitate efficient pig behavior monitoring, yet traditional approaches suffer from operational inefficiency and animal stress. We address these limitations through a lightweight YOLOv8n architecture enhanced with SPD-Conv for feature preservation during downsampling, LSKBlock attention for contextual feature fusion, and a dedicated small-target detection head. Experimental validation demonstrates superior performance: the optimized model achieves a 92.4% mean average precision (mAP@0.5) and 87.4% recall, significantly outperforming baseline YOLOv8n by 3.7% in AP while maintaining minimal parameter growth (3.34M). Controlled illumination tests confirm enhanced robustness under strong and warm lighting conditions, with performance gains of 1.5% and 0.7% in AP, respectively. This high-precision framework enables real-time recognition of standing, prone lying, lateral lying, and feeding behaviors in commercial piggeries, supporting early health anomaly detection through non-invasive monitoring. Full article
(This article belongs to the Section Pigs)
Show Figures

Figure 1

21 pages, 3712 KB  
Article
CISC-YOLO: A Lightweight Network for Micron-Level Defect Detection on Wafers via Efficient Cross-Scale Feature Fusion
by Yulun Chi, Xingyu Gong, Bing Zhao and Lei Yao
Electronics 2025, 14(19), 3960; https://doi.org/10.3390/electronics14193960 - 9 Oct 2025
Viewed by 68
Abstract
With the development of the semiconductor manufacturing process towards miniaturization and high integration, the detection of microscopic defects on wafer surfaces faces the challenge of balancing precision and efficiency. Therefore, this study proposes a lightweight inspection model based on the YOLOv8 framework, aiming [...] Read more.
With the development of the semiconductor manufacturing process towards miniaturization and high integration, the detection of microscopic defects on wafer surfaces faces the challenge of balancing precision and efficiency. Therefore, this study proposes a lightweight inspection model based on the YOLOv8 framework, aiming to achieve an optimal balance between inspection accuracy, model complexity, and inference speed. First, we design a novel lightweight module called IRB-GhostConv-C2f (IGC) to replace the C2f module in the backbone, thereby significantly minimizing redundant feature computations. Second, a CNN-based cross-scale feature fusion neck network, the CCFF-ISC neck, is proposed to reduce the redundant computation of low-level features and enhance the expression of multi-scale semantic information. Meanwhile, the novel IRB-SCSA-C2f (ISC) module replaces the C2f in the neck to further improve the efficiency of feature fusion. In addition, a novel dynamic head network, DyHeadv3, is integrated into the head structure, aiming to improve the small-scale target detection performance by dynamically adjusting the feature interaction mechanism. Finally, so as to comprehensively assess the proposed algorithm’s performance, an industrial dataset of wafer defects, WSDD, is constructed, which covers “broken edges”, “scratches”, “oil pollution”, and “minor defects”. The experimental results demonstrate that the CISC-YOLO model attains an mAP50 of 93.7%, and the parameter amount is reduced to 1.92 M, outperforming other mainstream leading algorithms in the field. The proposed approach provides a high-precision and low-latency real-time defect detection solution for semiconductor industry scenarios. Full article
Show Figures

Figure 1

24 pages, 2777 KB  
Article
LightSeek-YOLO: A Lightweight Architecture for Real-Time Trapped Victim Detection in Disaster Scenarios
by Xiaowen Tian, Yubi Zheng, Liangqing Huang, Rengui Bi, Yu Chen, Shiqi Wang and Wenkang Su
Mathematics 2025, 13(19), 3231; https://doi.org/10.3390/math13193231 - 9 Oct 2025
Viewed by 134
Abstract
Rapid and accurate detection of trapped victims is vital in disaster rescue operations, yet most existing object detection methods cannot simultaneously deliver high accuracy and fast inference under resource-constrained conditions. To address this limitation, we propose the LightSeek-YOLO, a lightweight, real-time victim detection [...] Read more.
Rapid and accurate detection of trapped victims is vital in disaster rescue operations, yet most existing object detection methods cannot simultaneously deliver high accuracy and fast inference under resource-constrained conditions. To address this limitation, we propose the LightSeek-YOLO, a lightweight, real-time victim detection framework for disaster scenarios built upon YOLOv11. Our LightSeek-YOLO integrates three core innovations. First, it employs HGNetV2 as the backbone, whose HGStem and HGBlock modules leverage depthwise separable convolutions to markedly reduce computational cost while preserving feature extraction. Secondly, it introduces Seek-DS (Seek-DownSampling), a dual-branch downsampling module that preserves key feature extrema through a MaxPool branch while capturing spatial patterns via a progressive convolution branch, thereby effectively mitigating background interference. Third, it incorporates Seek-DH (Seek Detection Head), a lightweight detection head that processes features through a unified pipeline, enhancing scale adaptability while reducing parameter redundancy. Evaluated on the common C2A disaster dataset, LightSeek-YOLO achieves 0.478 AP@small for small-object detection, demonstrating strong robustness in challenging conditions such as rubble and smoke. Moreover, on the COCO, it reaches 0.473 mAP@[0.5:0.95], matching YOLOv8n while achieving superior computational efficiency through 38.2% parameter reduction and 39.5% FLOP reduction, and achieving 571.72 FPS on desktop hardware, with computational efficiency improvements suggesting potential for edge deployment pending validation. Full article
(This article belongs to the Special Issue Machine Learning Applications in Image Processing and Computer Vision)
Show Figures

Figure 1

31 pages, 6076 KB  
Article
MSWindD-YOLO: A Lightweight Edge-Deployable Network for Real-Time Wind Turbine Blade Damage Detection in Sustainable Energy Operations
by Pan Li, Jitao Zhou, Jian Zeng, Qian Zhao and Qiqi Yang
Sustainability 2025, 17(19), 8925; https://doi.org/10.3390/su17198925 - 8 Oct 2025
Viewed by 135
Abstract
Wind turbine blade damage detection is crucial for advancing wind energy as a sustainable alternative to fossil fuels. Existing methods based on image processing technologies face challenges such as limited adaptability to complex environments, trade-offs between model accuracy and computational efficiency, and inadequate [...] Read more.
Wind turbine blade damage detection is crucial for advancing wind energy as a sustainable alternative to fossil fuels. Existing methods based on image processing technologies face challenges such as limited adaptability to complex environments, trade-offs between model accuracy and computational efficiency, and inadequate real-time inference capabilities. In response to these limitations, we put forward MSWindD-YOLO, a lightweight real-time detection model for wind turbine blade damage. Building upon YOLOv5s, our work introduces three key improvements: (1) the replacement of the Focus module with the Stem module to enhance computational efficiency and multi-scale feature fusion, integrating EfficientNetV2 structures for improved feature extraction and lightweight design, while retaining the SPPF module for multi-scale context awareness; (2) the substitution of the C3 module with the GBC3-FEA module to reduce computational redundancy, coupled with the incorporation of the CBAM attention mechanism at the neck network’s terminus to amplify critical features; and (3) the adoption of Shape-IoU loss function instead of CIoU loss function to facilitate faster model convergence and enhance localization accuracy. Evaluated on the Wind Turbine Blade Damage Visual Analysis Dataset (WTBDVA), MSWindD-YOLO achieves a precision of 95.9%, a recall of 96.3%, an mAP@0.5 of 93.7%, and an mAP@0.5:0.95 of 87.5%. With a compact size of 3.12 MB and 22.4 GFLOPs inference cost, it maintains high efficiency. After TensorRT acceleration on Jetson Orin NX, the model attains 43 FPS under FP16 quantization for real-time damage detection. Consequently, the proposed MSWindD-YOLO model not only elevates detection accuracy and inference efficiency but also achieves significant model compression. Its deployment-compatible performance in edge environments fulfills stringent industrial demands, ultimately advancing sustainable wind energy operations through lightweight lifecycle maintenance solutions for wind farms. Full article
Show Figures

Figure 1

24 pages, 6407 KB  
Article
Lightweight SCC-YOLO for Winter Jujube Detection and 3D Localization with Cross-Platform Deployment Evaluation
by Meng Zhou, Yaohua Hu, Anxiang Huang, Yiwen Chen, Xing Tong, Mengfei Liu and Yunxiao Pan
Agriculture 2025, 15(19), 2092; https://doi.org/10.3390/agriculture15192092 - 8 Oct 2025
Viewed by 96
Abstract
Harvesting winter jujubes is a key step in production, yet traditional manual approaches are labor-intensive and inefficient. To overcome these challenges, we propose SCC-YOLO, a lightweight method for winter jujube detection, 3D localization, and cross-platform deployment, aiming to support intelligent harvesting. In this [...] Read more.
Harvesting winter jujubes is a key step in production, yet traditional manual approaches are labor-intensive and inefficient. To overcome these challenges, we propose SCC-YOLO, a lightweight method for winter jujube detection, 3D localization, and cross-platform deployment, aiming to support intelligent harvesting. In this study, RGB-D cameras were integrated with an improved YOLOv11 network optimized by ShuffleNetV2, CBAM, and a redesigned C2f_WTConv module, which enables joint spatial–frequency feature modeling and enhances small-object detection in complex orchard conditions. The model was trained on a diversified dataset with extensive augmentation to ensure robustness. In addition, the original localization loss was replaced with DIoU to improve bounding box regression accuracy. A robotic harvesting system was developed, and an Eye-to-Hand calibration-based 3D localization pipeline was implemented to map fruit coordinates to the robot workspace for accurate picking. To validate engineering applicability, the SCC-YOLO model was deployed on both desktop (PyTorch and ONNX Runtime) and mobile (NCNN with Vulkan+FP16) platforms, and FPS, latency, and stability were comparatively analyzed. Experimental results showed that SCC-YOLO improved mAP by 5.6% over YOLOv11, significantly enhanced detection precision and robustness, and achieved real-time performance on mobile devices while maintaining peak throughput on high-performance desktops. Field and laboratory tests confirmed the system’s effectiveness for detection, localization, and harvesting efficiency, demonstrating its adaptability to diverse deployment environments and its potential for broader agricultural applications. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

21 pages, 6386 KB  
Article
SPMF-YOLO-Tracker: A Method for Quantifying Individual Activity Levels and Assessing Health in Newborn Piglets
by Jingge Wei, Yurong Tang, Jinxin Chen, Kelin Wang, Peng Li, Mingxia Shen and Longshen Liu
Agriculture 2025, 15(19), 2087; https://doi.org/10.3390/agriculture15192087 - 7 Oct 2025
Viewed by 147
Abstract
This study proposes a behavioral monitoring framework for newborn piglets based on SPMF-YOLO object detection and ByteTrack multi-object tracking, which enables precise quantification of early postnatal activity levels and health assessment. The method enhances small-object detection performance by incorporating the SPDConv module, the [...] Read more.
This study proposes a behavioral monitoring framework for newborn piglets based on SPMF-YOLO object detection and ByteTrack multi-object tracking, which enables precise quantification of early postnatal activity levels and health assessment. The method enhances small-object detection performance by incorporating the SPDConv module, the MFM module, and the NWD loss function into YOLOv11. When combined with the ByteTrack algorithm, it achieves stable tracking and maintains trajectory continuity for multiple targets. An annotated dataset containing both detection and tracking labels was constructed using video data from 10 piglet pens for evaluation. Experimental results indicate that SPMF-YOLO achieved a recognition accuracy rate of 95.3% for newborn piglets. When integrated with ByteTrack, it achieves 79.1% HOTA, 92.2% MOTA, and 84.7% IDF1 in multi-object tracking tasks, thereby outperforming existing methods. Building upon this foundation, this study further quantified the cumulative movement distance of each newborn piglet within 30 min after birth and proposed a health-assessment method based on statistical thresholds. The results demonstrated an overall consistency rate of 98.2% across pens and an accuracy rate of 92.9% for identifying abnormal individuals. The results validated the effectiveness of this method for quantifying individual behavior and assessing health status in newborn piglets within complex farming environments, providing a feasible technical pathway and scientific basis for health management and early intervention in precision animal husbandry. Full article
(This article belongs to the Special Issue Modeling of Livestock Breeding Environment and Animal Behavior)
Show Figures

Figure 1

Back to TopTop