Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (131)

Search Parameters:
Keywords = WIoU

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 4203 KiB  
Article
SRW-YOLO: A Detection Model for Environmental Risk Factors During the Grid Construction Phase
by Yu Zhao, Fei Liu, Qiang He, Fang Liu, Xiaohu Sun and Jiyong Zhang
Remote Sens. 2025, 17(15), 2576; https://doi.org/10.3390/rs17152576 - 24 Jul 2025
Viewed by 244
Abstract
With the rapid advancement of UAV-based remote sensing and image recognition techniques, identifying environmental risk factors from aerial imagery has emerged as a focal point in intelligent inspection during the power transmission and distribution projects construction phase. The uneven spatial distribution of risk [...] Read more.
With the rapid advancement of UAV-based remote sensing and image recognition techniques, identifying environmental risk factors from aerial imagery has emerged as a focal point in intelligent inspection during the power transmission and distribution projects construction phase. The uneven spatial distribution of risk factors on construction sites, their weak texture signatures, and the inherently multi-scale nature of UAV imagery pose significant detection challenges. To address these issues, we propose a one-stage SRW-YOLO algorithm built upon the YOLOv11 framework. First, a P2-scale shallow feature detection layer is added to capture high-resolution fine details of small targets. Second, we integrate a reparameterized convolution based on channel shuffle (RCS) of a one-shot aggregation (RCS-OSA) module into the backbone and neck’s shallow layers, enhancing feature extraction while significantly reducing inference latency. Finally, a dynamic non-monotonic focusing mechanism WIoU v3 loss function is employed to reweigh low-quality annotations, thereby improving small-object localization accuracy. Experimental results demonstrate that SRW-YOLO achieves an overall precision of 80.6% and mAP of 79.1% on the State Grid dataset, and exhibits similarly superior performance on the VisDrone2019 dataset. Compared with other one-stage detectors, SRW-YOLO delivers markedly higher detection accuracy, offering critical technical support for multi-scale, heterogeneous environmental risk monitoring during the power transmission and distribution projects construction phase, and establishes the theoretical foundation for rapid and accurate inspection using UAV-based intelligent imaging. Full article
Show Figures

Figure 1

25 pages, 9517 KiB  
Article
YOLOv8n-SSDW: A Lightweight and Accurate Model for Barnyard Grass Detection in Fields
by Yan Sun, Hanrui Guo, Xiaoan Chen, Mengqi Li, Bing Fang and Yingli Cao
Agriculture 2025, 15(14), 1510; https://doi.org/10.3390/agriculture15141510 - 13 Jul 2025
Cited by 1 | Viewed by 299
Abstract
Barnyard grass is a major noxious weed in paddy fields. Accurate and efficient identification of barnyard grass is crucial for precision field management. However, existing deep learning models generally suffer from high parameter counts and computational complexity, limiting their practical application in field [...] Read more.
Barnyard grass is a major noxious weed in paddy fields. Accurate and efficient identification of barnyard grass is crucial for precision field management. However, existing deep learning models generally suffer from high parameter counts and computational complexity, limiting their practical application in field scenarios. Moreover, the morphological similarity, overlapping, and occlusion between barnyard grass and rice pose challenges for reliable detection in complex environments. To address these issues, this study constructed a barnyard grass detection dataset using high-resolution images captured by a drone equipped with a high-definition camera in rice experimental fields in Haicheng City, Liaoning Province. A lightweight field barnyard grass detection model, YOLOv8n-SSDW, was proposed to enhance detection precision and speed. Based on the baseline YOLOv8n model, a novel Separable Residual Coord Conv (SRCConv) was designed to replace the original convolution module, significantly reducing parameters while maintaining detection accuracy. The Spatio-Channel Enhanced Attention Module (SEAM) was introduced and optimized to improve sensitivity to barnyard grass edge features. Additionally, the lightweight and efficient Dysample upsampling module was incorporated to enhance feature map resolution. A new WIoU loss function was developed to improve bounding box classification and regression accuracy. Comprehensive performance analysis demonstrated that YOLOv8n-SSDW outperformed state-of-the-art models. Ablation studies confirmed the effectiveness of each improvement module. The final fused model achieved lightweight performance while improving detection accuracy, with a 2.2% increase in mAP_50, 3.8% higher precision, 0.6% higher recall, 10.6% fewer parameters, 9.8% lower FLOPs, and an 11.1% reduction in model size compared to the baseline. Field tests using drones combined with ground-based computers further validated the model’s robustness in real-world complex paddy environments. The results indicate that YOLOv8n-SSDW exhibits excellent accuracy and efficiency. This study provides valuable insights for barnyard grass detection in rice fields. Full article
Show Figures

Figure 1

23 pages, 88853 KiB  
Article
RSW-YOLO: A Vehicle Detection Model for Urban UAV Remote Sensing Images
by Hao Wang, Jiapeng Shang, Xinbo Wang, Qingqi Zhang, Xiaoli Wang, Jie Li and Yan Wang
Sensors 2025, 25(14), 4335; https://doi.org/10.3390/s25144335 - 11 Jul 2025
Viewed by 534
Abstract
Vehicle detection in remote sensing images faces significant challenges due to small object sizes, scale variation, and cluttered backgrounds. To address these issues, we propose RSW-YOLO, an enhanced detection model built upon the YOLOv8n framework, designed to improve feature extraction and robustness against [...] Read more.
Vehicle detection in remote sensing images faces significant challenges due to small object sizes, scale variation, and cluttered backgrounds. To address these issues, we propose RSW-YOLO, an enhanced detection model built upon the YOLOv8n framework, designed to improve feature extraction and robustness against environmental noise. A Restormer module is incorporated into the backbone to model long-range dependencies via self-attention, enabling better handling of multi-scale features and complex scenes. A dedicated detection head is introduced for small objects, focusing on critical channels while suppressing irrelevant information. Additionally, the original CIoU loss is replaced with WIoU, which dynamically reweights predicted boxes based on their quality, enhancing localization accuracy and stability. Experimental results on the DJCAR dataset show mAP@0.5 and mAP@0.5:0.95 improvements of 5.4% and 6.2%, respectively, and corresponding gains of 4.3% and 2.6% on the VisDrone dataset. These results demonstrate that RSW-YOLO offers a robust and accurate solution for UAV-based vehicle detection, particularly in urban scenes with dense or small targets. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Figure 1

22 pages, 3153 KiB  
Article
Research on Road Damage Detection Algorithms for Intelligent Inspection Robots
by Hongsai Tian, Feng Zhao, Dongqing Yang, Haitao Cheng, Jiahao Zhang and Shuangshuang Song
Electronics 2025, 14(14), 2762; https://doi.org/10.3390/electronics14142762 - 9 Jul 2025
Viewed by 271
Abstract
Intelligent inspection robots are crucial tools that can be used to ensure road safety. However, current intelligent inspection robots for road damage detection are confronted by challenges, including insufficient detection accuracy and poor adaptability in complex environments. These issues exert a direct influence [...] Read more.
Intelligent inspection robots are crucial tools that can be used to ensure road safety. However, current intelligent inspection robots for road damage detection are confronted by challenges, including insufficient detection accuracy and poor adaptability in complex environments. These issues exert a direct influence on the reliability of road damage detection and the effectiveness of its practical application. To address the aforementioned issues, this study proposes a road damage detection model based on deep learning. At first, the backbone network is augmented with a multi-scale convolutional attention, which promotes more effective feature extraction and strengthens the model’s perception and representation of features at multiple scales. Secondly, the traditional SPPF module is replaced with the proposed SPPELAN module, which maintains consistent channel dimensions and facilitates improved multi-scale contextual feature extraction, thereby enhancing detection accuracy and inference efficiency under identical experimental conditions. Finally, the introduction of the WIOU loss function enhances the overall performance of the model. The experimental results on the test dataset demonstrate that the road damage detection model designed in this paper is significantly better than the original model in multiple indicators, with mAP@0.5 increased by 0.6%, accuracy increased by 1.7%, Recall increased by 1.4%, and frames per second (FPS) increased by 25.063 frames. Compared with YOLOv7 and YOLOv9, the mAP@0.5 of this model increased by 6.3% and 2.2%, the accuracy increased by 9.3% and 2.4%, and the Recall increased by 10.8% and 5.4%. In addition, the experimental results indicate that the road damage detection model designed in this study exhibits significant performance improvement in real-time road damage detection and holds promising application prospects. Full article
Show Figures

Figure 1

25 pages, 3011 KiB  
Article
An Enhanced YOLOv8 Model with Symmetry-Aware Feature Extraction for High-Accuracy Solar Panel Defect Detection
by Xiaoxia Lin, Xinyue Xiao, Lin Sun, Xiaodong Yang, Chunwei Leng, Yan Li, Zhenyu Niu, Yingzhou Meng and Weihao Gong
Symmetry 2025, 17(7), 1052; https://doi.org/10.3390/sym17071052 - 3 Jul 2025
Viewed by 414
Abstract
The growing popularity of solar panels is crucial for global decarbonization, but harsh environmental conditions can lead to defects such as cracks, fingerprints, and short circuits. Existing methods face the challenge of detecting multi-scale defects while maintaining real-time performance. This paper proposes a [...] Read more.
The growing popularity of solar panels is crucial for global decarbonization, but harsh environmental conditions can lead to defects such as cracks, fingerprints, and short circuits. Existing methods face the challenge of detecting multi-scale defects while maintaining real-time performance. This paper proposes a solar panel defect detection model, DCE-YOLO, based on YOLOv8. The model incorporates a C2f-DWR-DRB module for multi-scale feature extraction, where the parallel DRB branch models spatial symmetry through symmetric-rate dilated convolutions, improving robustness and consistency. The COT attention module strengthens long-range dependencies and fuses local and global contexts to achieve symmetric feature representation. The lightweight and efficient detection head improves detection speed and accuracy. The CIoU loss function is replaced with WIoU, and a non-monotonic dynamic focusing mechanism is used to mitigate the effect of low-quality samples. Experimental results show that compared with the YOLOv8 benchmark, DCE-YOLO achieves a 2.1% performance improvement on mAP@50 and a 4.9% performance improvement on mAP@50-95. Compared with recent methods, DCE-YOLO exhibits broader defect coverage, stronger robustness, and a better performance-efficiency balance, making it highly suitable for edge deployment. The synergistic interaction between the C2f-DWR-DRB module and COT attention enhances the detection of symmetric and multi-scale defects under real-world conditions. Full article
(This article belongs to the Section Engineering and Materials)
Show Figures

Figure 1

23 pages, 2463 KiB  
Article
MCDet: Target-Aware Fusion for RGB-T Fire Detection
by Yuezhu Xu, He Wang, Yuan Bi, Guohao Nie and Xingmei Wang
Forests 2025, 16(7), 1088; https://doi.org/10.3390/f16071088 - 30 Jun 2025
Viewed by 314
Abstract
Forest fire detection is vital for ecological conservation and disaster management. Existing visual detection methods exhibit instability in smoke-obscured or illumination-variable environments. Although multimodal fusion has demonstrated potential, effectively resolving inconsistencies in smoke features across diverse modalities remains a significant challenge. This issue [...] Read more.
Forest fire detection is vital for ecological conservation and disaster management. Existing visual detection methods exhibit instability in smoke-obscured or illumination-variable environments. Although multimodal fusion has demonstrated potential, effectively resolving inconsistencies in smoke features across diverse modalities remains a significant challenge. This issue stems from the inherent ambiguity between regions characterized by high temperatures in infrared imagery and those with elevated brightness levels in visible-light imaging systems. In this paper, we propose MCDet, an RGB-T forest fire detection framework incorporating target-aware fusion. To alleviate feature cross-modal ambiguity, we design a Multidimensional Representation Collaborative Fusion module (MRCF), which constructs global feature interactions via a state-space model and enhances local detail perception through deformable convolution. Then, a content-guided attention network (CGAN) is introduced to aggregate multidimensional features by dynamic gating mechanism. Building upon this foundation, the integration of WIoU further suppresses vegetation occlusion and illumination interference on a holistic level, thereby reducing the false detection rate. Evaluated on three forest fire datasets and one pedestrian dataset, MCDet achieves a mean detection accuracy of 77.5%, surpassing advanced methods. This performance makes MCDet a practical solution to enhance early warning system reliability. Full article
(This article belongs to the Special Issue Advanced Technologies for Forest Fire Detection and Monitoring)
Show Figures

Figure 1

21 pages, 8542 KiB  
Article
An Efficient Algorithm for Small Livestock Object Detection in Unmanned Aerial Vehicle Imagery
by Wenbo Chen, Dongliang Wang and Xiaowei Xie
Animals 2025, 15(12), 1794; https://doi.org/10.3390/ani15121794 - 18 Jun 2025
Viewed by 691
Abstract
Livestock population surveys are crucial for grassland management tasks such as health and epidemic prevention, grazing prohibition, rest grazing, and forage–livestock balance assessment. These tasks are integral to the modernization and upgrading of the livestock industry and the sustainable development of grasslands. Unmanned [...] Read more.
Livestock population surveys are crucial for grassland management tasks such as health and epidemic prevention, grazing prohibition, rest grazing, and forage–livestock balance assessment. These tasks are integral to the modernization and upgrading of the livestock industry and the sustainable development of grasslands. Unmanned aerial vehicles (UAVs) provide significant advantages in flexibility and maneuverability, making them ideal for livestock population surveys. However, grazing livestock in UAV images often appear small and densely packed, leading to identification errors. To address this challenge, we propose an efficient Livestock Network (LSNET) algorithm, a novel YOLOv7-based network. Our approach incorporates a low-level prediction head (P2) to detect small objects from shallow feature maps, while removing a deep-level prediction head (P5) to mitigate the effects of excessive down-sampling. To capture high-level semantic features, we introduce the Large Kernel Attentions Spatial Pyramid Pooling (LKASPP) module. In addition, we replaced the original CIoU with the WIoU v3 loss function. Furthermore, we developed a dataset of grazing livestock for deep learning using UAV images from the Prairie Chenbarhu Banner in Hulunbuir, Inner Mongolia. Our results demonstrate that the proposed module significantly improves the detection accuracy for small livestock objects, with the mean Average Precision (mAP) increasing by 1.47% compared to YOLOv7. Thus, this work offers a novel and practical solution for livestock detection in expansive farms. It overcomes the limitations of existing methods and contributes to more effective livestock management and advancements in agricultural technology. Full article
(This article belongs to the Section Animal System and Management)
Show Figures

Figure 1

21 pages, 4072 KiB  
Article
ST-YOLOv8: Small-Target Ship Detection in SAR Images Targeting Specific Marine Environments
by Fei Gao, Yang Tian, Yongliang Wu and Yunxia Zhang
Appl. Sci. 2025, 15(12), 6666; https://doi.org/10.3390/app15126666 - 13 Jun 2025
Viewed by 354
Abstract
Synthetic Aperture Radar (SAR) image ship detection faces challenges such as distinguishing ships from other terrains and structures, especially in specific marine complex environments. The motivation behind this work is to enhance detection accuracy while minimizing false positives, which is crucial for applications [...] Read more.
Synthetic Aperture Radar (SAR) image ship detection faces challenges such as distinguishing ships from other terrains and structures, especially in specific marine complex environments. The motivation behind this work is to enhance detection accuracy while minimizing false positives, which is crucial for applications like defense vessel monitoring and civilian search and rescue operations. To achieve this goal, we propose several architectural improvements to You Only Look Once version 8 Nano (YOLOv8n) and present Small Target-YOLOv8(ST-YOLOv8)—a novel lightweight SAR ship detection model based on the enhance YOLOv8n framework. The C2f module in the backbone’s transition sections is replaced by the Conv_Online Reparameterized Convolution (C_OREPA) module, reducing convolutional complexity and improving efficiency. The Atrous Spatial Pyramid Pooling (ASPP) module is added to the end of the backbone to extract finer features from smaller and more complex ship targets. In the neck network, the Shuffle Attention (SA) module is employed before each upsampling step to improve upsampling quality. Additionally, we replace the Complete Intersection over Union (C-IoU) loss function with the Wise Intersection over Union (W-IoU) loss function, which enhances bounding box precision. We conducted ablation experiments on two widely used multimodal SAR datasets. The proposed model significantly outperforms the YOLOv8n baseline, achieving 94.1% accuracy, 82% recall, and 87.6% F1 score on the SAR Ship Detection Dataset (SSDD), and 92.7% accuracy, 84.5% recall, and 88.1% F1 score on the SAR Ship Dataset_v0 dataset (SSDv0). Furthermore, the ST-YOLOv8 model outperforms several state-of-the-art multi-scale ship detection algorithms on both datasets. In summary, the ST-YOLOv8 model, by integrating advanced neural network architectures and optimization techniques, significantly improves detection accuracy and reduces false detection rates. This makes it highly suitable for complex backgrounds and multi-scale ship detection. Future work will focus on lightweight model optimization for deployment on mobile platforms to broaden its applicability across different scenarios. Full article
Show Figures

Figure 1

23 pages, 3908 KiB  
Article
MSUD-YOLO: A Novel Multiscale Small Object Detection Model for UAV Aerial Images
by Xiaofeng Zhao, Hui Zhang, Wenwen Zhang, Junyi Ma, Chenxiao Li, Yao Ding and Zhili Zhang
Drones 2025, 9(6), 429; https://doi.org/10.3390/drones9060429 - 13 Jun 2025
Cited by 1 | Viewed by 828
Abstract
Due to the objects in UAV aerial images often presenting characteristics of multiple scales, small objects, complex backgrounds, etc., the performance of object detection using current models is not satisfactory. To address the above issues, this paper designs a multiscale small object detection [...] Read more.
Due to the objects in UAV aerial images often presenting characteristics of multiple scales, small objects, complex backgrounds, etc., the performance of object detection using current models is not satisfactory. To address the above issues, this paper designs a multiscale small object detection model for UAV aerial images, namely MSUD-YOLO, based on YOLOv10s. First, the model uses an attention scale sequence fusion mode to achieve more efficient multiscale feature fusion. Meanwhile, a tiny prediction head is incorporated to make the model focus on the low-level features, thus improving its ability to detect small objects. Secondly, a novel feature extraction module named CFormerCGLU has been designed, which improves feature extraction capability in a lighter way. In addition, the model uses lightweight convolution instead of standard convolution to reduce the model’s computation. Finally, the WIoU v3 loss function is used to make the model more focused on low-quality examples, thereby improving the model’s object localization ability. Experimental results on the VisDrone2019 dataset show that MSUD-YOLO improves mAP50 by 8.5% compared with YOLOv10s. Concurrently, the overall model reduces parameters by 6.3%, verifying the model’s effectiveness for object detection in UAV aerial images in complex environments. Furthermore, compared with multiple latest UAV object detection algorithms, our designed MSUD-YOLO offers higher detection accuracy and lower computational cost; e.g., mAP50 reaches 43.4%, but parameters are only 6.766 M. Full article
Show Figures

Figure 1

16 pages, 3170 KiB  
Article
Improvement in Pavement Defect Scenarios Using an Improved YOLOv10 with ECA Attention, RefConv and WIoU
by Xiaolin Zhang, Lei Lu, Hanyun Luo and Lei Wang
World Electr. Veh. J. 2025, 16(6), 328; https://doi.org/10.3390/wevj16060328 - 13 Jun 2025
Viewed by 412
Abstract
This study addresses challenges such as multi-scale defects, varying lighting, and irregular shapes by proposing an improved YOLOv10 model that integrates the ECA attention mechanism, RefConv feature enhancement module, and WIoU loss function for complex pavement defect detection. The RefConv dual-branch structure achieves [...] Read more.
This study addresses challenges such as multi-scale defects, varying lighting, and irregular shapes by proposing an improved YOLOv10 model that integrates the ECA attention mechanism, RefConv feature enhancement module, and WIoU loss function for complex pavement defect detection. The RefConv dual-branch structure achieves feature complementarity between local details and global context (mAP increased by 2.1%), the ECA mechanism models channel relationships using 1D convolution (small-object recall rate increased by 27%), and the WIoU loss optimizes difficult sample regression through a dynamic weighting mechanism (location accuracy improved by 37%). Experiments show that on a dataset constructed from 23,949 high-resolution images, the improved model’s mAP reaches 68.2%, which is an increase of 6.2% compared to the baseline YOLOv10, maintaining a stable recall rate of 83.5% in highly reflective and low-light scenarios, with an inference speed of 158 FPS (RTX 4080), providing a high-precision real-time solution for intelligent road inspection. Full article
Show Figures

Figure 1

19 pages, 3527 KiB  
Article
BBW YOLO: Intelligent Detection Algorithms for Aluminium Profile Material Surface Defects
by Zijuan Yin, Haichao Li, Bo Qi and Guangyue Shan
Coatings 2025, 15(6), 684; https://doi.org/10.3390/coatings15060684 - 6 Jun 2025
Cited by 1 | Viewed by 523
Abstract
This study aims to address the issue of various defects on the surface of aluminum profile materials, which can significantly impact industrial production as well as the reliability and safety of products. An algorithmic model, BBW YOLO (YOLOv8-BiFPN-BiFormer-WIoU v3), based on an enhanced [...] Read more.
This study aims to address the issue of various defects on the surface of aluminum profile materials, which can significantly impact industrial production as well as the reliability and safety of products. An algorithmic model, BBW YOLO (YOLOv8-BiFPN-BiFormer-WIoU v3), based on an enhanced YOLOv8 model is proposed for aluminum profile material surface-defect detection. First, the model can effectively eliminate redundant feature information and enhance the feature-extraction process by incorporating a weighted Bidirectional Feature Pyramid Feature-fusion Network (BiFPN). Second, the model incorporates a dynamic sparse-attention mechanism (BiFormer) along with an efficient pyramidal network architecture, which enhances the precision and detection speed of the model. Meanwhile, the model optimizes the loss function using Wise-IoU v3 (WIoU v3), which effectively enhances the localization performance of surface-defect detection. The experimental results demonstrate that the precision and recall of the BBW YOLO model are improved by 5% and 2.65%, respectively, compared with the original YOLOv8 model. Notably, the BBW YOLO model achieved a real-time detection speed of 292.3 f/s. In addition, the model size of BBW YOLO is only 6.3 MB. At the same time, the floating-point operations of BBW YOLO are reduced to 8.3 G. As a result, the BBW YOLO model offers excellent defect detection performance and opens up new opportunities for its efficient development in the aluminum industry. Full article
(This article belongs to the Special Issue Solid Surfaces, Defects and Detection, 2nd Edition)
Show Figures

Figure 1

23 pages, 6314 KiB  
Article
For Precision Animal Husbandry: Precise Detection of Specific Body Parts of Sika Deer Based on Improved YOLO11
by Jinfan Wei, Haotian Gong, Lan Luo, Lingyun Ni, Zhipeng Li, Juanjuan Fan, Tianli Hu, Ye Mu, Yu Sun and He Gong
Agriculture 2025, 15(11), 1218; https://doi.org/10.3390/agriculture15111218 - 3 Jun 2025
Cited by 1 | Viewed by 749
Abstract
The breeding of sika deer has significant economic value in China. However, the traditional management methods have problems such as low efficiency, easy triggering of strong stress responses, and damage to animal welfare. Therefore, the development of non-contact, automated, and precise monitoring and [...] Read more.
The breeding of sika deer has significant economic value in China. However, the traditional management methods have problems such as low efficiency, easy triggering of strong stress responses, and damage to animal welfare. Therefore, the development of non-contact, automated, and precise monitoring and management technologies has become an urgent need for the sustainable development of this industry. In response to this demand, this study designed a model MFW-YOLO based on YOLO11, aiming to achieve precise detection of specific body parts of sika deer in a real breeding environment. Improvements include: designing a lightweight and efficient hybrid backbone network, MobileNetV4HybridSmall; The multi-scale fast pyramid pooling module (SPPFMscale) is proposed. The WIoU v3 loss function is used to replace the default loss function. To verify the effectiveness of the method, we constructed a sika deer dataset containing 1025 images, covering five categories. The experimental results show that the improved model performs well. Its mAP50 and MAP50-95 reached 91.9% and 64.5%, respectively. This model also demonstrates outstanding efficiency. The number of parameters is only 62% (5.9 million) of the original model, the computational load is 60% (12.8 GFLOPs) of the original model, and the average inference time is as low as 3.8 ms. This work provides strong algorithmic support for achieving non-contact intelligent monitoring of sika deer, assisting in automated management (deer antler collection and preparation), and improving animal welfare, demonstrating the application potential of deep learning technology in modern precision animal husbandry. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

18 pages, 7780 KiB  
Article
Mango Inflorescence Detection Based on Improved YOLOv8 and UAVs-RGB Images
by Linhui Wang, Jiayi Xiao, Xuxiang Peng, Yonghong Tan, Zhenqi Zhou, Lizhi Chen, Quanli Tang, Wenzhi Cheng and Xiaolin Liang
Forests 2025, 16(6), 896; https://doi.org/10.3390/f16060896 - 27 May 2025
Viewed by 427
Abstract
During the flowering period of mango trees, pests often hide in the inflorescences to suck sap, affecting fruit formation. By accurately detecting the number and location of mango inflorescences in the early stages, it can help target-specific spraying equipment to perform precise pesticide [...] Read more.
During the flowering period of mango trees, pests often hide in the inflorescences to suck sap, affecting fruit formation. By accurately detecting the number and location of mango inflorescences in the early stages, it can help target-specific spraying equipment to perform precise pesticide application. This study focuses on mango panicles and addresses challenges such as high crop planting density, poor image quality, and complex backgrounds. A series of improvements were made to the YOLOv8 model to enhance performance for this type of detection task. Firstly, a mango panicle dataset was constructed by selecting, augmenting, and correcting samples based on actual agricultural conditions. Second, the backbone network of YOLOv8 was replaced with FasterNet. Although this led to a slight decrease in accuracy, it significantly improved inference speed and reduced model parameters, demonstrating that FasterNet effectively reduced computational complexity while optimizing accuracy. Further, the GAM (Global Attention Module) attention mechanism was introduced as an attention module in the backbone network to enhance feature extraction capabilities. Experimental results indicated that the addition of GAM improved the average precision by 2.2 percentage points, outperforming other attention mechanisms such as SE, CA, and CBAM. Finally, the model’s bounding box localization ability was enhanced by replacing the loss function with WIoU, which also accelerated model convergence and improved the mAP@.5 metric by 1.1 percentage points. Our approach demonstrates a discrepancy of less than 10% compared to manual counted results. Full article
Show Figures

Figure 1

25 pages, 12873 KiB  
Article
License Plate Detection Based on Improved YOLOv8n Network
by Ruizhe Zhu, Qiyang He, Hai Jin, Yonghua Han and Kejian Jiang
Electronics 2025, 14(10), 2065; https://doi.org/10.3390/electronics14102065 - 20 May 2025
Viewed by 941
Abstract
To address the challenges of complex backgrounds, varying target scales, and dense targets in license plate detection within surveillance scenarios, we propose an enhanced method based on an improved YOLOv8n network. This approach involves redesigning key components of the YOLOv8n architecture, including the [...] Read more.
To address the challenges of complex backgrounds, varying target scales, and dense targets in license plate detection within surveillance scenarios, we propose an enhanced method based on an improved YOLOv8n network. This approach involves redesigning key components of the YOLOv8n architecture, including the C2f module, the SPPF module, and the detection head. Additionally, we optimize the WIoU loss function, replacing the original CIoU loss function, which leads to improved bounding box feature extraction and enhanced regression accuracy. To evaluate the model’s robustness in complex environments with varying lighting, backgrounds, angles, and vehicle types, we created a custom surveillance license plate dataset. Experimental results show that the improved model achieves a notable increase in detection accuracy, with mAP@0.5 rising from 90.9% in the baseline model to 94.4%, precision improving from 90.2% to 92.8%, and recall increasing from 82.9% to 87.9%. Additionally, the model’s parameters are reduced from 3.1 M to 2.1 M, significantly enhancing computational efficiency. Moreover, the model achieves an inference speed FPS of 86, maintaining high precision and meeting real-time detection requirements. This demonstrates that our method provides an efficient and reliable solution for license plate detection in surveillance scenarios. Full article
Show Figures

Figure 1

32 pages, 9504 KiB  
Article
CSSA-YOLO: Cross-Scale Spatiotemporal Attention Network for Fine-Grained Behavior Recognition in Classroom Environments
by Liuchen Zhou, Xiangpeng Liu, Xiqiang Guan and Yuhua Cheng
Sensors 2025, 25(10), 3132; https://doi.org/10.3390/s25103132 - 15 May 2025
Viewed by 722
Abstract
Under a student-centered educational paradigm, project-based learning (PBL) assessment requires accurate identification of classroom behaviors to facilitate effective teaching evaluations and the implementation of personalized learning strategies. The increasing use of visual and multi-modal sensors in smart classrooms has made it possible to [...] Read more.
Under a student-centered educational paradigm, project-based learning (PBL) assessment requires accurate identification of classroom behaviors to facilitate effective teaching evaluations and the implementation of personalized learning strategies. The increasing use of visual and multi-modal sensors in smart classrooms has made it possible to continuously capture rich behavioral data. However, challenges such as lighting variations, occlusions, and diverse behaviors complicate sensor-based behavior analysis. To address these issues, we introduce CSSA-YOLO, a novel detection network that incorporates cross-scale feature optimization. First, we establish a C2fs module that captures spatiotemporal dependencies in small-scale actions such as hand-raising through hierarchical window attention. Second, a Shuffle Attention mechanism is then integrated into the neck to suppress interference from complex backgrounds, thereby enhancing the model’s ability to focus on relevant features. Finally, to further enhance the network’s ability to detect small targets and complex boundary behaviors, we utilize the WIoU loss function, which dynamically weights gradients to optimize the localization accuracy of occluded targets. Experiments involving the SCB03-S dataset showed that CSSA-YOLO outperforms traditional methods, achieving an mAP50 of 76.0%, surpassing YOLOv8m by 1.2%, particularly in complex background and occlusion scenarios. Furthermore, it reaches 78.31 FPS, meeting the requirements for real-time application. This study offers a reliable solution for precise behavior recognition in classroom settings, supporting the development of intelligent education systems. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Back to TopTop