Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (272)

Search Parameters:
Keywords = BiFPN

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 16984 KiB  
Article
Small Ship Detection Based on Improved Neural Network Algorithm and SAR Images
by Jiaqi Li, Hongyuan Huo, Li Guo, De Zhang, Wei Feng, Yi Lian and Long He
Remote Sens. 2025, 17(15), 2586; https://doi.org/10.3390/rs17152586 - 24 Jul 2025
Abstract
Synthetic aperture radar images can be used for ship target detection. However, due to the unclear ship outline in SAR images, noise and land background factors affect the difficulty and accuracy of ship (especially small target ship) detection. Therefore, based on the YOLOv5s [...] Read more.
Synthetic aperture radar images can be used for ship target detection. However, due to the unclear ship outline in SAR images, noise and land background factors affect the difficulty and accuracy of ship (especially small target ship) detection. Therefore, based on the YOLOv5s model, this paper improves its backbone network and feature fusion network algorithm to improve the accuracy of ship detection target recognition. First, the LSKModule is used to improve the backbone network of YOLOv5s. By adaptively aggregating the features extracted by large-size convolution kernels to fully obtain context information, at the same time, key features are enhanced and noise interference is suppressed. Secondly, multiple Depthwise Separable Convolution layers are added to the SPPF (Spatial Pyramid Pooling-Fast) structure. Although a small number of parameters and calculations are introduced, features of different receptive fields can be extracted. Third, the feature fusion network of YOLOv5s is improved based on BIFPN, and the shallow feature map is used to optimize the small target detection performance. Finally, the CoordConv module is added before the detect head of YOLOv5, and two coordinate channels are added during the convolution operation to further improve the accuracy of target detection. The map50 of this method for the SSDD dataset and HRSID dataset reached 97.6% and 91.7%, respectively, and was compared with a variety of advanced target detection models. The results show that the detection accuracy of this method is higher than other similar target detection algorithms. Full article
Show Figures

Figure 1

25 pages, 9119 KiB  
Article
An Improved YOLOv8n-Based Method for Detecting Rice Shelling Rate and Brown Rice Breakage Rate
by Zhaoyun Wu, Yehao Zhang, Zhongwei Zhang, Fasheng Shen, Li Li, Xuewu He, Hongyu Zhong and Yufei Zhou
Agriculture 2025, 15(15), 1595; https://doi.org/10.3390/agriculture15151595 - 24 Jul 2025
Abstract
Accurate and real-time detection of rice shelling rate (SR) and brown rice breakage rate (BR) is crucial for intelligent hulling sorting but remains challenging because of small grain size, dense adhesion, and uneven illumination causing missed detections and blurred boundaries in traditional YOLOv8n. [...] Read more.
Accurate and real-time detection of rice shelling rate (SR) and brown rice breakage rate (BR) is crucial for intelligent hulling sorting but remains challenging because of small grain size, dense adhesion, and uneven illumination causing missed detections and blurred boundaries in traditional YOLOv8n. This paper proposes a high-precision, lightweight solution based on an enhanced YOLOv8n with improvements in network architecture, feature fusion, and attention mechanism. The backbone’s C2f module is replaced with C2f-Faster-CGLU, integrating partial convolution (PConv) local convolution and convolutional gated linear unit (CGLU) gating to reduce computational redundancy via sparse interaction and enhance small-target feature extraction. A bidirectional feature pyramid network (BiFPN) weights multiscale feature fusion to improve edge positioning accuracy of dense grains. Attention mechanism for fine-grained classification (AFGC) is embedded to focus on texture and damage details, enhancing adaptability to light fluctuations. The Detect_Rice lightweight head compresses parameters via group normalization and dynamic convolution sharing, optimizing small-target response. The improved model achieved 96.8% precision and 96.2% mAP. Combined with a quantity–mass model, SR/BR detection errors reduced to 1.11% and 1.24%, meeting national standard (GB/T 29898-2013) requirements, providing an effective real-time solution for intelligent hulling sorting. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

21 pages, 5181 KiB  
Article
TEB-YOLO: A Lightweight YOLOv5-Based Model for Bamboo Strip Defect Detection
by Xipeng Yang, Chengzhi Ruan, Fei Yu, Ruxiao Yang, Bo Guo, Jun Yang, Feng Gao and Lei He
Forests 2025, 16(8), 1219; https://doi.org/10.3390/f16081219 - 24 Jul 2025
Abstract
The accurate detection of surface defects in bamboo is critical to maintaining product quality. Traditional inspection methods rely heavily on manual labor, making the manufacturing process labor-intensive and error-prone. To overcome these limitations, TEB-YOLO is introduced in this paper, a lightweight and efficient [...] Read more.
The accurate detection of surface defects in bamboo is critical to maintaining product quality. Traditional inspection methods rely heavily on manual labor, making the manufacturing process labor-intensive and error-prone. To overcome these limitations, TEB-YOLO is introduced in this paper, a lightweight and efficient defect detection model based on YOLOv5s. Firstly, EfficientViT replaces the original YOLOv5s backbone, reducing the computational cost while improving feature extraction. Secondly, BiFPN is adopted in place of PANet to enhance multi-scale feature fusion and preserve detailed information. Thirdly, an Efficient Local Attention (ELA) mechanism is embedded in the backbone to strengthen local feature representation. Lastly, the original CIoU loss is replaced with EIoU loss to enhance localization precision. The proposed model achieves a precision of 91.7% with only 10.5 million parameters, marking a 5.4% accuracy improvement and a 22.9% reduction in parameters compared to YOLOv5s. Compared with other mainstream models including YOLOv5n, YOLOv7, YOLOv8n, YOLOv9t, and YOLOv9s, TEB-YOLO achieves precision improvements of 11.8%, 1.66%, 2.0%, 2.8%, and 1.1%, respectively. The experiment results show that TEB-YOLO significantly improves detection precision and model lightweighting, offering a practical and effective solution for real-time bamboo surface defect detection. Full article
Show Figures

Figure 1

28 pages, 15535 KiB  
Article
BCA-MVSNet: Integrating BIFPN and CA for Enhanced Detail Texture in Multi-View Stereo Reconstruction
by Ning Long, Zhengxu Duan, Xiao Hu and Mingju Chen
Electronics 2025, 14(15), 2958; https://doi.org/10.3390/electronics14152958 - 24 Jul 2025
Abstract
The 3D point cloud generated by MVSNet has good scene integrity but lacks sensitivity to details, causing holes and non-dense areas in flat and weak-texture regions. To address this problem and enhance the point cloud information of weak-texture areas, the BCA-MVSNet network is [...] Read more.
The 3D point cloud generated by MVSNet has good scene integrity but lacks sensitivity to details, causing holes and non-dense areas in flat and weak-texture regions. To address this problem and enhance the point cloud information of weak-texture areas, the BCA-MVSNet network is proposed in this paper. The network integrates the Bidirectional Feature Pyramid Network (BIFPN) into the feature processing of the MVSNet backbone network to accurately extract the features of weak-texture regions. In the feature map fusion stage, the Coordinate Attention (CA) mechanism is introduced into 3DU-Net to obtain the position information on the channel dimension related to the direction, improve the detail feature extraction, optimize the depth map and improve the depth accuracy. The experimental results show that BCA-MVSNet not only improves the accuracy of detail texture reconstruction, but also effectively controls the computational overhead. In the DTU dataset, the Overall and Comp metrics of BCA-MVSNet are reduced by 10.2% and 2.6%, respectively; in the Tanksand Temples dataset, the Mean metrics of the eight scenarios are improved by 6.51%. Three scenes are shot by binocular camera, and the reconstruction quality is excellent in the weak-texture area by combining the camera parameters and the BCA-MVSNet model. Full article
Show Figures

Figure 1

19 pages, 4861 KiB  
Article
Towards Precise Papaya Ripeness Assessment: A Deep Learning Framework with Dynamic Detection Heads
by Haohai You, Jing Fan, Dongyan Huang, Weilong Yan, Xiting Zhang, Zhenke Sun, Hongtao Liu and Jun Yuan
Agriculture 2025, 15(15), 1585; https://doi.org/10.3390/agriculture15151585 - 24 Jul 2025
Abstract
Papaya ripeness identification is a key task in orchard management. To achieve efficient deployment of this task on edge computing devices, this paper proposes a lightweight detection model, ABD-YOLO-ting, based on YOLOv8. First, the width factor of YOLOv8n is adjusted to construct a [...] Read more.
Papaya ripeness identification is a key task in orchard management. To achieve efficient deployment of this task on edge computing devices, this paper proposes a lightweight detection model, ABD-YOLO-ting, based on YOLOv8. First, the width factor of YOLOv8n is adjusted to construct a lightweight backbone network, YOLO-Ting. Second, a low-computation ADown module is introduced to replace the standard downsampling structure, aiming to enhance feature extraction efficiency. Third, an enhanced BiFPN is integrated into the neck structure to achieve efficient multi-scale feature fusion. Finally, to strengthen the model’s capability in identifying small objects, the dynamic detection head DyHead is introduced to improve ripeness recognition accuracy. On a self-constructed Japanese quince orchard dataset, ABD-YOLO-ting achieves a mAP50 of 94.7% and a mAP50–95 of 77.4%, with only 1.47 M parameters and 5.4 G FLOPs, significantly outperforming mainstream models such as YOLOv5, YOLOv8, and YOLOv11. On edge devices, the model achieves a well-balanced trade-off between detection speed and accuracy, demonstrating strong potential for practical applications in intelligent harvesting and orchard management. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

22 pages, 6496 KiB  
Article
Real-Time Search and Rescue with Drones: A Deep Learning Approach for Small-Object Detection Based on YOLO
by Francesco Ciccone and Alessandro Ceruti
Drones 2025, 9(8), 514; https://doi.org/10.3390/drones9080514 - 22 Jul 2025
Viewed by 190
Abstract
Unmanned aerial vehicles are increasingly used in civil Search and Rescue operations due to their rapid deployment and wide-area coverage capabilities. However, detecting missing persons from aerial imagery remains challenging due to small object sizes, cluttered backgrounds, and limited onboard computational resources, especially [...] Read more.
Unmanned aerial vehicles are increasingly used in civil Search and Rescue operations due to their rapid deployment and wide-area coverage capabilities. However, detecting missing persons from aerial imagery remains challenging due to small object sizes, cluttered backgrounds, and limited onboard computational resources, especially when managed by civil agencies. In this work, we present a comprehensive methodology for optimizing YOLO-based object detection models for real-time Search and Rescue scenarios. A two-stage transfer learning strategy was employed using VisDrone for general aerial object detection and Heridal for Search and Rescue-specific fine-tuning. We explored various architectural modifications, including enhanced feature fusion (FPN, BiFPN, PB-FPN), additional detection heads (P2), and modules such as CBAM, Transformers, and deconvolution, analyzing their impact on performance and computational efficiency. The best-performing configuration (YOLOv5s-PBfpn-Deconv) achieved a mAP@50 of 0.802 on the Heridal dataset while maintaining real-time inference on embedded hardware (Jetson Nano). Further tests at different flight altitudes and explainability analyses using EigenCAM confirmed the robustness and interpretability of the model in real-world conditions. The proposed solution offers a viable framework for deploying lightweight, interpretable AI systems for UAV-based Search and Rescue operations managed by civil protection authorities. Limitations and future directions include the integration of multimodal sensors and adaptation to broader environmental conditions. Full article
Show Figures

Figure 1

17 pages, 3708 KiB  
Article
YOLOv8-DBW: An Improved YOLOv8-Based Algorithm for Maize Leaf Diseases and Pests Detection
by Xiang Gan, Shukun Cao, Jin Wang, Yu Wang and Xu Hou
Sensors 2025, 25(15), 4529; https://doi.org/10.3390/s25154529 - 22 Jul 2025
Viewed by 177
Abstract
To solve the challenges of low detection accuracy of maize pests and diseases, complex detection models, and difficulty in deployment on mobile or embedded devices, an improved YOLOv8 algorithm was proposed. Based on the original YOLOv8n, the algorithm replaced the Conv module with [...] Read more.
To solve the challenges of low detection accuracy of maize pests and diseases, complex detection models, and difficulty in deployment on mobile or embedded devices, an improved YOLOv8 algorithm was proposed. Based on the original YOLOv8n, the algorithm replaced the Conv module with the DSConv module in the backbone network, which reduced the backbone network parameters and computational load and improved the detection accuracy at the same time. Additionally, BiFPN was introduced to construct a bidirectional feature pyramid structure, which realized efficient information flow and fusion between different scale features and enhanced the feature fusion ability of the model. At the same time, the Wise-IoU loss function was combined to optimize the training process, which improved the convergence speed and regression accuracy of the loss function. The experimental results showed that the precision, recall, and mAP0.5 of the improved algorithm were improved by 1.4, 1.1, and 1.5%, respectively, compared with YOLOv8n, and the model parameters and computational costs were reduced by 6.6 and 7.3%, respectively. The experimental results demonstrate the effectiveness and superiority of the improved YOLOv8 algorithm, which provides an efficient, accurate, and easy-to-deploy solution for maize leaf pest detection. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

22 pages, 14158 KiB  
Article
Enhanced YOLOv8 for Robust Pig Detection and Counting in Complex Agricultural Environments
by Jian Li, Wenkai Ma, Yanan Wei and Tan Wang
Animals 2025, 15(14), 2149; https://doi.org/10.3390/ani15142149 - 21 Jul 2025
Viewed by 165
Abstract
Accurate pig counting is crucial for precision livestock farming, enabling optimized feeding management and health monitoring. Detection-based counting methods face significant challenges due to mutual occlusion, varying illumination conditions, diverse pen configurations, and substantial variations in pig densities. Previous approaches often struggle with [...] Read more.
Accurate pig counting is crucial for precision livestock farming, enabling optimized feeding management and health monitoring. Detection-based counting methods face significant challenges due to mutual occlusion, varying illumination conditions, diverse pen configurations, and substantial variations in pig densities. Previous approaches often struggle with complex agricultural environments where lighting conditions, pig postures, and crowding levels create challenging detection scenarios. To address these limitations, we propose EAPC-YOLO (enhanced adaptive pig counting YOLO), a robust architecture integrating density-aware processing with advanced detection optimizations. The method consists of (1) an enhanced YOLOv8 network incorporating multiple architectural improvements for better feature extraction and object localization. These improvements include DCNv4 deformable convolutions for irregular pig postures, BiFPN bidirectional feature fusion for multi-scale information integration, EfficientViT linear attention for computational efficiency, and PIoU v2 loss for improved overlap handling. (2) A density-aware post-processing module with intelligent NMS strategies that adapt to different crowding scenarios. Experimental results on a comprehensive dataset spanning diverse agricultural scenarios (nighttime, controlled indoor, and natural daylight environments with density variations from 4 to 30 pigs) demonstrate our method achieves 94.2% mAP@0.5 for detection performance and 96.8% counting accuracy, representing 12.3% and 15.7% improvements compared to the strongest baseline, YOLOv11n. This work enables robust, accurate pig counting across challenging agricultural environments, supporting precision livestock management. Full article
Show Figures

Figure 1

20 pages, 3898 KiB  
Article
Synergistic Multi-Model Approach for GPR Data Interpretation: Forward Modeling and Robust Object Detection
by Hang Zhang, Zhijie Ma, Xinyu Fan and Feifei Hou
Remote Sens. 2025, 17(14), 2521; https://doi.org/10.3390/rs17142521 - 20 Jul 2025
Viewed by 212
Abstract
Ground penetrating radar (GPR) is widely used for subsurface object detection, but manual interpretation of hyperbolic features in B-scan images remains inefficient and error-prone. In addition, traditional forward modeling methods suffer from low computational efficiency and strong dependence on field measurements. To address [...] Read more.
Ground penetrating radar (GPR) is widely used for subsurface object detection, but manual interpretation of hyperbolic features in B-scan images remains inefficient and error-prone. In addition, traditional forward modeling methods suffer from low computational efficiency and strong dependence on field measurements. To address these challenges, we propose an unsupervised data augmentation framework that utilizes CycleGAN-based model to generates diverse synthetic B-scan images by simulating varying geological parameters and scanning configurations. This approach achieves GPR data forward modeling and enhances the scenario coverage of training data. We then apply the EfficientDet architecture, which incorporates a bidirectional feature pyramid network (BiFPN) for multi-scale feature fusion, to enhance the detection capability of hyperbolic signatures in B-scan images under challenging conditions such as partial occlusions and background noise. The proposed method achieves a mean average precision (mAP) of 0.579 on synthetic datasets, outperforming YOLOv3 and RetinaNet by 16.0% and 23.5%, respectively, while maintaining robust multi-object detection in complex field conditions. Full article
(This article belongs to the Special Issue Advanced Ground-Penetrating Radar (GPR) Technologies and Applications)
Show Figures

Figure 1

25 pages, 8583 KiB  
Article
YOLO-MAD: Multi-Scale Geometric Structure Feature Extraction and Fusion for Steel Surface Defect Detection
by Hantao Ding, Junkai Chen, Hairong Ye and Yanbing Chen
Appl. Sci. 2025, 15(14), 7887; https://doi.org/10.3390/app15147887 - 15 Jul 2025
Viewed by 234
Abstract
Lightweight visual models are crucial for industrial defect detection tasks. Traditional methods and even some lightweight detectors often struggle with the trade-off between high computational demands and insufficient accuracy. To overcome these issues, this study introduces YOLO-MAD, an innovative model optimized through a [...] Read more.
Lightweight visual models are crucial for industrial defect detection tasks. Traditional methods and even some lightweight detectors often struggle with the trade-off between high computational demands and insufficient accuracy. To overcome these issues, this study introduces YOLO-MAD, an innovative model optimized through a multi-scale geometric structure feature extraction and fusion scheme. YOLO-MAD integrates three key modules: AKConv for robust geometric feature extraction, BiFPN to facilitate effective multi-scale feature integration, and Detect_DyHead for dynamic optimization of detection capabilities. Empirical evaluations demonstrate significant performance improvements: YOLO-MAD achieves a 5.4% mAP increase on the NEU-DET dataset and a 4.8% mAP increase on the GC10-DET dataset. Crucially, this is achieved under a moderate computational load (9.4 GFLOPs), outperforming several prominent lightweight models in detection accuracy while maintaining comparable efficiency. The model also shows enhanced recognition performance for most defect categories. This work presents a pioneering approach that balances lightweight design with high detection performance by efficiently leveraging multi-scale geometric feature extraction and fusion, offering a new paradigm for industrial defect detection. Full article
Show Figures

Figure 1

26 pages, 6233 KiB  
Article
A Method for Recognizing Dead Sea Bass Based on Improved YOLOv8n
by Lizhen Zhang, Chong Xu, Sai Jiang, Mengxiang Zhu and Di Wu
Sensors 2025, 25(14), 4318; https://doi.org/10.3390/s25144318 - 10 Jul 2025
Viewed by 189
Abstract
Deaths occur during the culture of sea bass, and if timely harvesting is not carried out, it will lead to water pollution and the continued spread of sea bass deaths. Therefore, it is necessary to promptly detect dead fish and take countermeasures. Existing [...] Read more.
Deaths occur during the culture of sea bass, and if timely harvesting is not carried out, it will lead to water pollution and the continued spread of sea bass deaths. Therefore, it is necessary to promptly detect dead fish and take countermeasures. Existing object detection algorithms, when applied to the task of detecting dead sea bass, often suffer from excessive model complexity, high computational cost, and reduced accuracy in the presence of occlusion. To overcome these limitations, this study introduces YOLOv8n-Deadfish, a lightweight and high-precision detection model. First, the homemade sea bass death recognition dataset was expanded to enhance the generalization ability of the neural network. Second, the C2f-faster–EMA (efficient multi-scale attention) convolutional module was designed to replace the C2f module in the backbone network of YOLOv8n, reducing redundant calculations and memory access, thereby more effectively extracting spatial features. Then, a weighted bidirectional feature pyramid network (BiFPN) was introduced to achieve a more thorough integration of deep and shallow features. Finally, in order to compensate for the weak generalization and slow convergence of the CIoU loss function in detection tasks, the Inner-CIoU loss function was used to accelerate bounding box regression and further improve the detection performance of the model. The experimental results show that the YOLOv8n-Deadfish model has an accuracy, recall, and mean precision of 90.0%, 90.4%, and 93.6%, respectively, which is an improvement of 2.0, 1.4, and 1.3 percentage points, respectively, over the original base network YOLOv8n. The number of model parameters and GFLOPs were reduced by 23.3% and 18.5%, respectively, and the detection speed was improved from the original 304.5 FPS to 424.6 FPS. This method can provide a technical basis for the identification of dead sea bass in the process of intelligent aquaculture. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

22 pages, 3494 KiB  
Article
Parcel Segmentation Method Combined YOLOV5s and Segment Anything Model Using Remote Sensing Image
by Xiaoqin Wu, Dacheng Wang, Caihong Ma, Yi Zeng, Yongze Lv, Xianmiao Huang and Jiandong Wang
Land 2025, 14(7), 1429; https://doi.org/10.3390/land14071429 - 8 Jul 2025
Viewed by 364
Abstract
Accurate land parcel segmentation in remote sensing imagery is critical for applications such as land use analysis, agricultural monitoring, and urban planning. However, existing methods often underperform in complex scenes due to small-object segmentation challenges, blurred boundaries, and background interference, often influenced by [...] Read more.
Accurate land parcel segmentation in remote sensing imagery is critical for applications such as land use analysis, agricultural monitoring, and urban planning. However, existing methods often underperform in complex scenes due to small-object segmentation challenges, blurred boundaries, and background interference, often influenced by sensor resolution and atmospheric variation. To address these limitations, we propose a dual-stage framework that combines an enhanced YOLOv5s detector with the Segment Anything Model (SAM) to improve segmentation accuracy and robustness. The improved YOLOv5s module integrates Efficient Channel Attention (ECA) and BiFPN to boost feature extraction and small-object recognition, while Soft-NMS is used to reduce missed detections. The SAM module receives bounding-box prompts from YOLOv5s and incorporates morphological refinement and mask stability scoring for improved boundary continuity and mask quality. A composite Focal-Dice loss is applied to mitigate class imbalance. In addition to the publicly available CCF BDCI dataset, we constructed a new WuJiang dataset to evaluate cross-domain performance. Experimental results demonstrate that our method achieves an IoU of 89.8% and a precision of 90.2%, outperforming baseline models and showing strong generalizability across diverse remote sensing conditions. Full article
Show Figures

Figure 1

18 pages, 4447 KiB  
Article
Ripe-Detection: A Lightweight Method for Strawberry Ripeness Detection
by Helong Yu, Cheng Qian, Zhenyang Chen, Jing Chen and Yuxin Zhao
Agronomy 2025, 15(7), 1645; https://doi.org/10.3390/agronomy15071645 - 6 Jul 2025
Viewed by 319
Abstract
Strawberry (Fragaria × ananassa), a nutrient-dense fruit with significant economic value in commercial cultivation, faces critical detection challenges in automated harvesting due to complex growth conditions such as foliage occlusion and variable illumination. To address these limitations, this study proposes Ripe-Detection, [...] Read more.
Strawberry (Fragaria × ananassa), a nutrient-dense fruit with significant economic value in commercial cultivation, faces critical detection challenges in automated harvesting due to complex growth conditions such as foliage occlusion and variable illumination. To address these limitations, this study proposes Ripe-Detection, a novel lightweight object detection framework integrating three key innovations: a PEDblock detection head architecture with depth-adaptive feature learning capability, an ADown downsampling method for enhanced detail perception with reduced computational overhead, and BiFPN-based hierarchical feature fusion with learnable weighting mechanisms. Developed using a purpose-built dataset of 1021 annotated strawberry images (Fragaria × ananassa ‘Red Face’ and ‘Sachinoka’ varieties) from Changchun Xiaohongmao Plantation and augmented through targeted strategies to enhance model robustness, the framework demonstrates superior performance over existing lightweight detectors, achieving mAP50 improvements of 13.0%, 9.2%, and 3.9% against YOLOv7-tiny, YOLOv10n, and YOLOv11n, respectively. Remarkably, the architecture attains 96.4% mAP50 with only 1.3M parameters (57% reduction from baseline) and 4.4 GFLOPs (46% lower computation), simultaneously enhancing accuracy while significantly reducing resource requirements, thereby providing a robust technical foundation for automated ripeness assessment and precision harvesting in agricultural robotics. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

17 pages, 3477 KiB  
Article
Rapid Identification of Mangrove Leaves Based on Improved YOLOv10 Model
by Haitao Sang, Ziming Li, Xiaoxue Shen, Shuwen Wang and Ying Zhang
Forests 2025, 16(7), 1068; https://doi.org/10.3390/f16071068 - 26 Jun 2025
Viewed by 228
Abstract
To address the issue of low recognition accuracy caused by the high morphological similarity of mangrove plant leaves, this study proposes a rapid identification method for mangrove leaves based on the YOLOv10 model, with corresponding improvements made to the baseline model. First, the [...] Read more.
To address the issue of low recognition accuracy caused by the high morphological similarity of mangrove plant leaves, this study proposes a rapid identification method for mangrove leaves based on the YOLOv10 model, with corresponding improvements made to the baseline model. First, the open-source tool LabelImg was employed to annotate leaf images and construct a mangrove leaf species dataset. Subsequently, a PSA attention mechanism was introduced to enhance the extraction of leaf detail features, while the SCDown downsampling method was adopted to preserve key characteristics. Furthermore, a BiFPN architecture incorporating SE modules was implemented to dynamically adjust channel weights for multi-scale feature fusion. Finally, the classification and regression tasks are decoupled by separating the detection head, and the final model is named YOLOv10-MSDet. Experimental results demonstrate that the improved model achieves rapid and accurate identification of various mangrove leaf species, with an average recognition accuracy of 92.4%—a 2.84 percentage point improvement over the baseline model, significantly enhancing the precision of mangrove leaf detection. Full article
Show Figures

Figure 1

18 pages, 3936 KiB  
Article
BSE-YOLO: An Enhanced Lightweight Multi-Scale Underwater Object Detection Model
by Yuhang Wang, Hua Ye and Xin Shu
Sensors 2025, 25(13), 3890; https://doi.org/10.3390/s25133890 - 22 Jun 2025
Viewed by 590
Abstract
Underwater images often exhibit characteristics such as low contrast, blurred and small targets, object clustering, and considerable variations in object morphology. Traditional detection methods tend to be susceptible to omission and false positives under these circumstances. Furthermore, owing to the constrained memory and [...] Read more.
Underwater images often exhibit characteristics such as low contrast, blurred and small targets, object clustering, and considerable variations in object morphology. Traditional detection methods tend to be susceptible to omission and false positives under these circumstances. Furthermore, owing to the constrained memory and limited computing power of underwater robots, there is a significant demand for lightweight models in underwater object detection tasks. Therefore, we propose an enhanced lightweight YOLOv10n-based model, BSE-YOLO. Firstly, we replace the original neck with an improved Bidirectional Feature Pyramid Network (Bi-FPN) to reduce parameters. Secondly, we propose a Multi-Scale Attention Synergy Module (MASM) to enhance the model’s perception of difficult features and make it focus on the important regions. Finally, we integrate Efficient Multi-Scale Attention (EMA) into the backbone and neck to improve feature extraction and fusion. The experiment results demonstrate that the proposed BSE-YOLO reaches 83.7% mAP@0.5 on URPC2020 and 83.9% mAP@0.5 on DUO, with the parameters reducing 2.47 M. Compared to the baseline model YOLOv10n, our BSE-YOLO improves mAP@0.5 by 2.2% and 3.0%, respectively, while reducing the number of parameters by approximately 0.2 M. The BSE-YOLO achieves a good balance between accuracy and lightweight, providing an effective solution for underwater object detection. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Back to TopTop