Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (5)

Search Parameters:
Keywords = RFCBAMConv

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 3534 KiB  
Article
I-YOLOv11n: A Lightweight and Efficient Small Target Detection Framework for UAV Aerial Images
by Yukai Ma, Caiping Xi, Ting Ma, Han Sun, Huiyang Lu, Xiang Xu and Chen Xu
Sensors 2025, 25(15), 4857; https://doi.org/10.3390/s25154857 - 7 Aug 2025
Viewed by 324
Abstract
UAV small target detection in urban security, disaster monitoring, agricultural inspection, and other fields faces the challenge of increasing accuracy and real-time requirements. However, existing detection algorithms still have weak small target representation ability, extensive computational resource overhead, and poor deployment adaptability. Therefore, [...] Read more.
UAV small target detection in urban security, disaster monitoring, agricultural inspection, and other fields faces the challenge of increasing accuracy and real-time requirements. However, existing detection algorithms still have weak small target representation ability, extensive computational resource overhead, and poor deployment adaptability. Therefore, this paper proposes a lightweight algorithm, I-YOLOv11n, based on YOLOv11n, which is systematically improved in terms of both feature enhancement and structure compression. The RFCBAMConv module that combines deformable convolution and channel–spatial attention is designed to adjust the receptive field and strengthen the edge features dynamically. The multiscale pyramid of STCMSP context and the lightweight Transformer–DyHead hybrid detection head are designed by combining the multiscale hole feature pyramid (DFPC), which realizes the cross-scale semantic modeling and adaptive focusing of the target area. A collaborative lightweight strategy is proposed. Firstly, the semantic discrimination ability of the teacher model for small targets is transferred to guide and protect the subsequent compression process by integrating the mixed knowledge distillation of response alignment, feature imitation, and structure maintenance. Secondly, the LAMP–Taylor channel pruning mechanism is used to compress the model redundancy, mainly to protect the key channels sensitive to shallow small targets. Finally, K-means++ anchor frame optimization based on IoU distance is implemented to adapt the feature structure retained after pruning and the scale distribution of small targets of UAV. While significantly reducing the model size (parameter 3.87 M, calculation 14.7 GFLOPs), the detection accuracy of small targets is effectively maintained and improved. Experiments on VisDrone, AI-TOD, and SODA-A datasets show that the mAP@0.5 and mAP@0.5:0.95 of I-YOLOv11n are 7.1% and 4.9% higher than the benchmark model YOLOv11 n, respectively, while maintaining real-time processing capabilities, verifying its comprehensive advantages in accuracy, light weight, and deployment. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

21 pages, 4936 KiB  
Article
A Lightweight Pavement Defect Detection Algorithm Integrating Perception Enhancement and Feature Optimization
by Xiang Zhang, Xiaopeng Wang and Zhuorang Yang
Sensors 2025, 25(14), 4443; https://doi.org/10.3390/s25144443 - 17 Jul 2025
Viewed by 343
Abstract
To address the current issue of large computations and the difficulty in balancing model complexity and detection accuracy in pavement defect detection models, a lightweight pavement defect detection algorithm, PGS-YOLO, is proposed based on YOLOv8, which integrates perception enhancement and feature optimization. The [...] Read more.
To address the current issue of large computations and the difficulty in balancing model complexity and detection accuracy in pavement defect detection models, a lightweight pavement defect detection algorithm, PGS-YOLO, is proposed based on YOLOv8, which integrates perception enhancement and feature optimization. The algorithm first designs the Receptive-Field Convolutional Block Attention Module Convolution (RFCBAMConv) and the Receptive-Field Convolutional Block Attention Module C2f-RFCBAM, based on which we construct an efficient Perception Enhanced Feature Extraction Network (PEFNet) that enhances multi-scale feature extraction capability by dynamically adjusting the receptive field. Secondly, the dynamic upsampling module DySample is introduced into the efficient feature pyramid, constructing a new feature fusion pyramid (Generalized Dynamic Sampling Feature Pyramid Network, GDSFPN) to optimize the multi-scale feature fusion effect. In addition, a shared detail-enhanced convolution lightweight detection head (SDCLD) was designed, which significantly reduces the model’s parameters and computation while improving localization and classification performance. Finally, Wise-IoU was introduced to optimize the training performance and detection accuracy of the model. Experimental results show that PGS-YOLO increases mAP50 by 2.8% and 2.9% on the complete GRDDC2022 dataset and the Chinese subset, respectively, outperforming the other detection models. The number of parameters and computations are reduced by 10.3% and 9.9%, respectively, compared to the YOLOv8n model, with an average frame rate of 69 frames per second, offering good real-time performance. In addition, on the CRACK500 dataset, PGS-YOLO improved mAP50 by 2.3%, achieving a better balance between model complexity and detection accuracy. Full article
(This article belongs to the Topic Applied Computing and Machine Intelligence (ACMI))
Show Figures

Figure 1

29 pages, 57066 KiB  
Article
URT-YOLOv11: A Large Receptive Field Algorithm for Detecting Tomato Ripening Under Different Field Conditions
by Di Mu, Yuping Guou, Wei Wang, Ran Peng, Chunjie Guo, Francesco Marinello, Yingjie Xie and Qiang Huang
Agriculture 2025, 15(10), 1060; https://doi.org/10.3390/agriculture15101060 - 14 May 2025
Cited by 1 | Viewed by 846
Abstract
This study proposes an improved YOLOv11 model to address the limitations of traditional tomato recognition algorithms in complex agricultural environments, such as lighting changes, occlusion, scale variations, and complex backgrounds. These factors often hinder accurate feature extraction, leading to recognition errors and reduced [...] Read more.
This study proposes an improved YOLOv11 model to address the limitations of traditional tomato recognition algorithms in complex agricultural environments, such as lighting changes, occlusion, scale variations, and complex backgrounds. These factors often hinder accurate feature extraction, leading to recognition errors and reduced computational efficiency. To overcome these challenges, the model integrates several architectural enhancements. First, the UniRepLKNet block replaces the C3k2 module in the standard network, improving computational efficiency, expanding the receptive field, and enhancing multi-scale target recognition. Second, the RFCBAMConv module in the neck integrates channel and spatial attention mechanisms, boosting small-object detection and robustness under varying lighting conditions. Finally, the TADDH module optimizes the detection head by balancing classification and regression tasks through task alignment strategies, further improving detection accuracy across different target scales. Ablation experiments confirm the contribution of each module to overall performance improvement. Our experimental results demonstrate that the proposed model exhibits enhanced stability under special conditions, such as similar backgrounds, lighting variations, and object occlusion, while significantly improving both accuracy and computational efficiency. The model achieves an accuracy of 85.4%, recall of 80.3%, and mAP@50 of 87.3%. Compared to the baseline YOLOv11, the improved model increases mAP@50 by 2.2% while reducing parameters to 2.16 M, making it well-suited for real-time applications in resource-constrained environments. This study provides an efficient and practical solution for intelligent agriculture, enhancing real-time tomato detection and laying a solid foundation for future crop monitoring systems. Full article
(This article belongs to the Special Issue Innovations in Precision Farming for Sustainable Agriculture)
Show Figures

Figure 1

21 pages, 9976 KiB  
Article
RLRD-YOLO: An Improved YOLOv8 Algorithm for Small Object Detection from an Unmanned Aerial Vehicle (UAV) Perspective
by Hanyun Li, Yi Li, Linsong Xiao, Yunfeng Zhang, Lihua Cao and Di Wu
Drones 2025, 9(4), 293; https://doi.org/10.3390/drones9040293 - 10 Apr 2025
Cited by 2 | Viewed by 2921
Abstract
In Unmanned Aerial Vehicle (UAV) target detection tasks, issues such as missing and erroneous detections frequently occur owing to the small size of the targets and the complexity of the image background. To improve these issues, an improved target detection algorithm named RLRD-YOLO, [...] Read more.
In Unmanned Aerial Vehicle (UAV) target detection tasks, issues such as missing and erroneous detections frequently occur owing to the small size of the targets and the complexity of the image background. To improve these issues, an improved target detection algorithm named RLRD-YOLO, based on You Only Look Once version 8 (YOLOv8), is proposed. First, the backbone network initially integrates the Receptive Field Attention Convolution (RFCBAMConv) Module, which combines the Convolutional Block Attention Module (CBAM) and Receptive Field Attention Convolution (RFAConv). This integration improves the issue of shared attention weights in receptive field features. It also combines attention mechanisms across both channel and spatial dimensions, enhancing the capability of feature extraction. Subsequently, Large-Scale Kernel Attention (LSKA) is integrated to further optimize the Spatial Pyramid Pooling Fast (SPPF) layer. This enhancement employs a large-scale convolutional kernel to improve the capture of intricate small target features and minimize background interference. To enhance feature fusion and effectively integrate low-level details with high-level semantic information, the Reparameterized Generalized Feature Pyramid Network (RepGFPN) replaces the original architecture in the neck network. Additionally, a small-target detection layer is added to enhance the model’s ability to perceive small targets. Finally, the detecting head is replaced with the Dynamic Head, designed to improve the localization accuracy of small targets in complex scenarios by optimizing for Scale Awareness, Spatial Awareness, and Task Awareness. The experimental results showed that RLRD-YOLO outperformed YOLOv8 on the VisDrone2019 dataset, achieving improvements of 12.2% in mAP@0.5 and 8.4% in mAP@0.5:0.95. It also surpassed other widely used object detection methods. Furthermore, experimental results on the HIT-HAV dataset demonstrate that RLRD-YOLO sustains excellent precision in infrared UAV imagery, validating its generalizability across diverse scenarios. Finally, RLRD-YOLO was deployed and validated on the typical airborne platform, Jetson Nano, providing reliable technical support for the improvement of detection algorithms in aerial scenarios and their practical applications. Full article
Show Figures

Figure 1

19 pages, 5561 KiB  
Article
CVW-Etr: A High-Precision Method for Estimating the Severity Level of Cotton Verticillium Wilt Disease
by Pan Pan, Qiong Yao, Jiawei Shen, Lin Hu, Sijian Zhao, Longyu Huang, Guoping Yu, Guomin Zhou and Jianhua Zhang
Plants 2024, 13(21), 2960; https://doi.org/10.3390/plants13212960 - 23 Oct 2024
Cited by 1 | Viewed by 1310
Abstract
Cotton verticillium wilt significantly impacts both cotton quality and yield. Selecting disease-resistant varieties and using their resistance genes in breeding is an effective and economical control measure. Accurate severity estimation of this disease is crucial for breeding resistant cotton varieties. However, current methods [...] Read more.
Cotton verticillium wilt significantly impacts both cotton quality and yield. Selecting disease-resistant varieties and using their resistance genes in breeding is an effective and economical control measure. Accurate severity estimation of this disease is crucial for breeding resistant cotton varieties. However, current methods fall short, slowing the breeding process. To address these challenges, this paper introduces CVW-Etr, a high-precision method for estimating the severity of cotton verticillium wilt. CVW-Etr classifies severity into six levels (L0 to L5) based on the proportion of segmented diseased leaves to lesions. Upon integrating YOLOv8-Seg with MobileSAM, CVW-Etr demonstrates excellent performance and efficiency with limited samples in complex field conditions. It incorporates the RFCBAMConv, C2f-RFCBAMConv, AWDownSample-Lite, and GSegment modules to handle blurry transitions between healthy and diseased regions and variations in angle and distance during image collection, and to optimize the model’s parameter size and computational complexity. Our experimental results show that CVW-Etr effectively segments diseased leaves and lesions, achieving a mean average precision (mAP) of 92.90% and an average severity estimation accuracy of 92.92% with only 2.6M parameters and 10.1G FLOPS. Through experiments, CVW-Etr proves robust in estimating cotton verticillium wilt severity, offering valuable insights for disease-resistant cotton breeding applications. Full article
(This article belongs to the Special Issue Integrated Pest Management and Plants Health)
Show Figures

Figure 1

Back to TopTop