Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (151)

Search Parameters:
Keywords = UAV target recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 1949 KB  
Article
A Novel Object Detection-Based Air-to-Ground Target Search and Localization Strategy
by Haoran Li, Qinling Zhang and Mi Zhen
Drones 2026, 10(5), 375; https://doi.org/10.3390/drones10050375 - 13 May 2026
Abstract
The ability of uncrewed aerial vehicles (UAVs) to hover, recognize, and localize ground targets is crucial for efficient and accurate intelligent low-altitude operations, such as material delivery, emergency rescue, and firefighting. This paper presents a technical solution for low-altitude UAV target recognition and [...] Read more.
The ability of uncrewed aerial vehicles (UAVs) to hover, recognize, and localize ground targets is crucial for efficient and accurate intelligent low-altitude operations, such as material delivery, emergency rescue, and firefighting. This paper presents a technical solution for low-altitude UAV target recognition and search localization. The core algorithm is a RepViT-enhanced detection model, which integrates the Re-Parameterization Vision Transformer (RepViT) lightweight neural network with an efficient object detection framework, further augmented by the Convolutional Block Attention Module (CBAM) to improve detection accuracy. The search localization strategy implements a tiered approach for exploring nearby areas from the current position, assigning targets to priority tiers and visiting them in order of priority. Experimental results demonstrate that the RepViT-enhanced model achieves a mean average precision (mAP) of 98.58% on a custom emergency rescue dataset, improving real-time detection speed by two frames per second (18.70 FPS vs. 16.70 FPS for the standard YOLOv4 baseline). Thus, the proposed method effectively enhances both detection accuracy and speed, enabling better target search and localization in complex environments. The search strategy was validated through simulations, confirming its feasibility. Full article
25 pages, 9539 KB  
Article
ACF-YOLO: Feature Enhancement and Multi-Scale Alignment for Sustainable Crop Small Object Detection
by Chuanxiang Li, Yihang Li, Wenzhong Yang and Danny Chen
Sustainability 2026, 18(9), 4168; https://doi.org/10.3390/su18094168 - 22 Apr 2026
Viewed by 230
Abstract
Sustainable precision agriculture is crucial for optimizing resource utilization, reducing chemical inputs, and ensuring global food security. High-precision automatic recognition and monitoring of key crop organs (e.g., wheat heads and flower clusters) serve as the technological foundation for sustainable agricultural management decisions. However, [...] Read more.
Sustainable precision agriculture is crucial for optimizing resource utilization, reducing chemical inputs, and ensuring global food security. High-precision automatic recognition and monitoring of key crop organs (e.g., wheat heads and flower clusters) serve as the technological foundation for sustainable agricultural management decisions. However, visual perception in natural field environments is highly susceptible to external conditions. To address the challenges of severe background interference and feature dilution in crop small object detection within complex agricultural scenarios, this paper proposes an enhanced detection network, ACF-YOLO, based on YOLO11. First, an Aggregated Multi-scale Local-Global Attention (AMLGA) module is designed to enhance the feature representation of weak targets by fusing local details with global semantics. Second, a Context-Guided Fusion Module (CGFM) and a Soft-Neighbor Interpolation (SNI) strategy are introduced. Their synergy alleviates feature aliasing effects and ensures the precise alignment of deep semantic information with shallow spatial details. Furthermore, the Inner-MPDIoU loss function is employed to optimize the bounding box regression accuracy for non-rigid targets by incorporating geometric constraints and auxiliary scale factors. To verify the detection capability of the proposed method, we constructed a UAV Wheat Head Dataset (UWHD) and conducted extensive experiments on the UWHD, GWHD2021, and RFRB datasets. The experimental results demonstrate that ACF-YOLO outperforms other comparative methods, confirming its stable detection performance and contributing to the sustainable development of agriculture. Full article
(This article belongs to the Section Sustainable Agriculture)
Show Figures

Figure 1

20 pages, 2593 KB  
Article
Radar UAV/Bird Trajectory Feature Classification Based on TCN-Transformer and the PC-TimeGAN Data Augmentation Framework
by Fei Tong, Kun Zhang, Guisheng Liao, Lin Li, Jingwei Xu and Keting Jiang
Sensors 2026, 26(8), 2528; https://doi.org/10.3390/s26082528 - 20 Apr 2026
Viewed by 471
Abstract
To address the challenges of scarce unmanned aerial vehicle (UAV) track samples, severe class imbalance, and high motion similarity between UAVs and birds in low-altitude radar recognition, this paper proposes a trajectory classification method integrating a TCN-Transformer model with a physics-constrained TimeGAN (PC-TimeGAN) [...] Read more.
To address the challenges of scarce unmanned aerial vehicle (UAV) track samples, severe class imbalance, and high motion similarity between UAVs and birds in low-altitude radar recognition, this paper proposes a trajectory classification method integrating a TCN-Transformer model with a physics-constrained TimeGAN (PC-TimeGAN) data augmentation framework. Specifically, the PC-TimeGAN generates high-quality, kinematically compliant UAV trajectories to alleviate data scarcity and class imbalance. A multi-scale TCN-Transformer is then constructed to comprehensively extract features, utilizing multi-kernel dilated convolutions for local temporal correlations and self-attention mechanisms for global temporal dependencies, thereby improving the discrimination between UAV and bird trajectories with similar motion patterns. Furthermore, a joint loss function combining Focal Loss and Triplet Loss is employed to optimize the decision boundaries and feature space, enhancing model robustness and generalization. Experiments on a measured dataset demonstrate that, under the 15-dimensional input setting, the proposed method achieves a UAV recall of 80.00%, an FAR of 3.15%, a precision of 64.00%, and an F1-score of 0.7111. Compared to baseline methods (e.g., SVM, LSTM, GRU, Transformer, and 1D-CNN), the proposed approach significantly improves UAV recall under limited trajectory information while keeping the false-alarm rate of misclassifying birds as UAVs low. Ultimately, this method markedly enhances the comprehensive performance of rapid track-level target classification for low-altitude surveillance radars. Full article
Show Figures

Figure 1

18 pages, 22316 KB  
Article
Optimization of Multi-Scale Feature Extraction and Loss Functions in YOLOv8 for Insulator Defect Detection
by Meng Su, Shuailun Geng, Hong Yu, Shuai Zhou, Lihua Zhou and Jiao Luo
Mathematics 2026, 14(8), 1376; https://doi.org/10.3390/math14081376 - 19 Apr 2026
Viewed by 401
Abstract
To address the challenges of high miss detection rates and accuracy degradation in UAV-based insulator defect detection—primarily stemming from complex background interference and the loss of fine-grained features—this paper presents an optimized lightweight detection framework based on an improved YOLOv8 model. The integration [...] Read more.
To address the challenges of high miss detection rates and accuracy degradation in UAV-based insulator defect detection—primarily stemming from complex background interference and the loss of fine-grained features—this paper presents an optimized lightweight detection framework based on an improved YOLOv8 model. The integration of a Spatial-to-Depth Convolution (SPDConv) module strengthens the extraction of fine-grained features for microscopic defects, while the incorporation of an SCConv module suppresses computational redundancy, leading to a 2.80% accuracy improvement. This architecture is further enhanced by a Channel and Spatial Reconstruction Attention Module (CSRAM), which dynamically prioritizes target-related regions and mitigates noise from vegetation and infrastructure. To improve regression robustness against low-quality annotations and blurred boundaries, a Focal-WIoU loss function utilizing a dynamic non-monotonic focusing mechanism is introduced. Experimental results on complex insulator datasets demonstrate that the proposed model achieves an mAP@0.5 of 91.75% and an mAP@0.5:0.95 of 59.86%, representing a 4.40% and 5.04% increase over the YOLOv8 baseline, respectively. Notably, while maintaining a lightweight profile with only 11.14 M parameters and 28.66 G FLOPs, the model achieves a high inference speed of 376.56 FPS, effectively enabling precise multi-scale defect recognition under extreme operational conditions. Full article
(This article belongs to the Special Issue Optimization Models and Algorithms in Data Science, 2nd Edition)
Show Figures

Figure 1

18 pages, 5307 KB  
Article
MSA-DETR: A Multi-Scale Attention Augmented Model for Small Object Detection in UAV Imagery
by Zhihao Li and Liang Qi
Remote Sens. 2026, 18(8), 1179; https://doi.org/10.3390/rs18081179 - 15 Apr 2026
Viewed by 500
Abstract
Small object detection in UAV imagery presents challenges due to factors such as minute scale, indistinct features, and severe background clutter, which constrain the recognition performance of end-to-end models like RT-DETR. To enhance detection accuracy for small-scale objects, this paper proposes MSA-DETR, a [...] Read more.
Small object detection in UAV imagery presents challenges due to factors such as minute scale, indistinct features, and severe background clutter, which constrain the recognition performance of end-to-end models like RT-DETR. To enhance detection accuracy for small-scale objects, this paper proposes MSA-DETR, a Multi-scale Spatial Attention-enhanced detection model based on RT-DETR (Res18). Three specific structural improvements are introduced. First, a PercepConv module is designed to capture comprehensive multi-scale information through 1 × 1, 3 × 3, and 5 × 5 convolutions, as well as dilated convolutions. This module integrates a lightweight channel attention mechanism to adaptively emphasize regions containing small objects. Second, the SODAttention module is introduced to jointly model local spatial details and global contextual information, thereby enhancing the discriminative capability in key regions and significantly suppressing interference from complex backgrounds. Finally, a dedicated small object detection layer is added to the detection head, incorporating shallow fine-grained features to compensate for the semantic limitations of deep layers concerning small targets. Experimental results demonstrate that the proposed MSA-DETR achieves significant performance gains on the VisDrone2019 dataset, increasing mAP@50 from 47.5% to 52.2% and mAP@50–95 from 29.3% to 33.2%. Moreover, the proposed model outperforms the baseline by an absolute margin of 1.9% on the small-object-specific metric APs, achieving 20.3%. These results validate the effectiveness of the proposed method for small object detection in UAV scenarios. Full article
Show Figures

Figure 1

20 pages, 5966 KB  
Article
Target Recognition Model for Seedling Sugar Beets from UAV Aerial Imagery
by Meijuan Cheng, Yuankai Chen, Yu Deng, Zhixiong Zeng, Jiahui Song, Xiao Wu, Jie Liu, Zhen Yin and Zhigang Zhang
Agriculture 2026, 16(7), 737; https://doi.org/10.3390/agriculture16070737 - 26 Mar 2026
Viewed by 455
Abstract
The extensive cultivation scale of sugar beet seedlings has resulted in the necessity for accurate identification and monitoring of the seedling count, a task which has become crucial and highly challenging in the sugar industry. However, sugar beet seedlings in UAV aerial photography [...] Read more.
The extensive cultivation scale of sugar beet seedlings has resulted in the necessity for accurate identification and monitoring of the seedling count, a task which has become crucial and highly challenging in the sugar industry. However, sugar beet seedlings in UAV aerial photography scenarios are mostly small targets with complex backgrounds. Existing general detection models not only have insufficient detection accuracy, but also struggle to balance computational efficiency and resource consumption. To meet the practical needs of field monitoring, this paper proposes the LDH-RTDETR, a sugar beet seedling detection model that balances high accuracy and light weight. This model uses LSNet for feature extraction to reduce size, adds a deformable attention (DAttention) module to capture fine-grained seedling features, and adopts HS-FPN to improve multi-scale feature fusion in the neck network. Experimental results show that the improved model significantly outperforms the original RT-DETR model, with a 3.6% increase in accuracy, a 2.1% increase in mAP50, a recall rate of 86.0%, and a final model size of only 43.3 MB, thus achieving an effective balance between accuracy and model size. This study’s improved model offers an efficient solution for large-area identification and counting of sugar beet seedlings, and is highly significant for advancing the automation of sugar crop field management and agricultural digital transformation. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

32 pages, 3916 KB  
Article
An Automated Detection Method for Motor Vehicles Encroaching on Non-Motorized Lanes Based on Unmanned Aerial Vehicle Imagery and Civilized Behavior Monitoring
by Zichan Tan, Yin Tan, Peijing Lin, Wenjie Su, Tian He and Weishen Wu
Sensors 2026, 26(7), 2027; https://doi.org/10.3390/s26072027 - 24 Mar 2026
Viewed by 409
Abstract
Motor vehicle encroachment into non-motorized lanes is a common but hard-to-verify violation in urban intersections, especially when monitored from unmanned aerial vehicles (UAVs) or high-mounted overhead views. Existing rule-based solutions built on horizontal bounding boxes and center-point/line-crossing criteria are sensitive to perspective distortion, [...] Read more.
Motor vehicle encroachment into non-motorized lanes is a common but hard-to-verify violation in urban intersections, especially when monitored from unmanned aerial vehicles (UAVs) or high-mounted overhead views. Existing rule-based solutions built on horizontal bounding boxes and center-point/line-crossing criteria are sensitive to perspective distortion, occlusion, and frame-to-frame jitter, resulting in unstable decisions and low evidential value. This paper presents a cascaded UAV-view system that closes the loop from perception to evidence output through detection–segmentation–recognition–decision. First, we adopt a two-stage detection cascade: a lightweight vehicle detector localizes vehicles using axis-aligned bounding boxes, and a dedicated YOLOv5n-based oriented bounding box (OBB) license plate detector, constructed via architecture grafting and weight transfer, is then applied within each vehicle region of interest (ROI) to localize rotated license plates under large pose variation and small-target conditions. Second, a U-Net lane region segmentation module provides pixel-level spatial constraints to define an enforceable lane occupancy region. Third, a perspective rectification step is integrated with the PP-OCRv4 optical character recognition (OCR) framework to improve license plate recognition reliability for tilted plates. Finally, an area ratio criterion and an N-frame temporal counter are used to suppress transient misdetections and stabilize alarms. On a representative 100-sample controlled encroachment benchmark, the proposed system improves detection accuracy from 67.0% to 92.0% and reduces the false positive rate from 32.35% to 5.88% compared with a baseline horizontal bounding box (HBB)-based rule. The system outputs both violation alarms and license plate evidence, supporting practical deployment for multi-view traffic governance. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

32 pages, 7914 KB  
Article
UAV Target Detection and Tracking Integrating a Dynamic Brain–Computer Interface
by Jun Wang, Zanyang Li, Lirong Yan, Muhammad Imtiaz, Hang Li, Muhammad Usman Shoukat, Jianatihan Jinsihan, Benjun Feng, Yi Yang, Fuwu Yan, Shumo He and Yibo Wu
Drones 2026, 10(3), 222; https://doi.org/10.3390/drones10030222 - 21 Mar 2026
Viewed by 1200
Abstract
To address the inherent limitations in the robustness of fully autonomous unmanned aerial vehicle (UAV) visual perception and the high cognitive workload associated with manual control, this paper proposes a human-in-the-loop brain–computer interface (BCI) control framework. The system integrates steady-state visual evoked potential [...] Read more.
To address the inherent limitations in the robustness of fully autonomous unmanned aerial vehicle (UAV) visual perception and the high cognitive workload associated with manual control, this paper proposes a human-in-the-loop brain–computer interface (BCI) control framework. The system integrates steady-state visual evoked potential (SSVEP) with deep learning techniques to create a spatio-temporally dynamic interaction paradigm, enabling real-time alignment between visual targets and frequency stimuli. At the perception level, an enhanced YOLOv11 network incorporating partial convolution (PConv) and shape intersection over union (Shape-IoU) loss is developed and coupled with the DeepSort multi-object tracking algorithm. This configuration ensures high-speed execution on edge computing platforms while maintaining stable stimulus coverage over dynamic targets, thus providing a robust visual induction environment for EEG decoding. At the neural decoding level, an enhanced task-discriminant component analysis (TDCA-V) algorithm is introduced to improve signal detection stability within non-stationary flight conditions. Experimental results demonstrate that within the predefined fixation task window, the system achieves 100% success in maintaining target identity (ID). The BCI system achieved an average command recognition accuracy of 91.48% within a 1.0 s time window, with the TDCA-V algorithm significantly outperforming traditional spatial filtering methods in dynamic scenarios. These findings demonstrate the system’s effectiveness in decoupling human cognitive intent from machine execution, providing a robust solution for human–machine collaborative control. Full article
(This article belongs to the Section Artificial Intelligence in Drones (AID))
Show Figures

Figure 1

21 pages, 15774 KB  
Article
Two-Phase Forest Damage Assessment with Sentinel-2 NDVI Double Differencing and UAV-Based Segmentation in the Sopron Mountains
by Norbert Ács, Bálint Heil, Botond Szász, Ádám Folcz, Márk Preisinger, Gyula Sándor and Kornél Czimber
Remote Sens. 2026, 18(5), 803; https://doi.org/10.3390/rs18050803 - 6 Mar 2026
Viewed by 658
Abstract
Due to climate change, drought periods are becoming more frequent and more intense, posing substantial stress to Central European forest stands, especially climatically sensitive conifer forests. The early detection and accurate spatial delineation of forest damage are essential for supporting adaptive forest management [...] Read more.
Due to climate change, drought periods are becoming more frequent and more intense, posing substantial stress to Central European forest stands, especially climatically sensitive conifer forests. The early detection and accurate spatial delineation of forest damage are essential for supporting adaptive forest management decisions. This study presents a two-tier, multi-step forest damage assessment approach that combines Sentinel-2 satellite-based NDVI double-difference analysis with UAV-based high-resolution photogrammetric evaluation. In the first phase, potential damaged forest patches were identified in two sample areas of the Sopron Mountains using double-difference maps derived from monthly window NDVI maxima calculated from Sentinel-2 data. In the second phase, UAV surveys were carried out over the selected forest compartments, resulting in individual-tree-level canopy segmentation and object-based NDVI analysis. The photogrammetric point clouds were combined with ground points derived from airborne laser scanning to enable the accurate generation of canopy height models. The results confirmed that NDVI double-difference analysis is suitable for the spatial detection of both gradual drought-related damage and sudden disturbances—such as forest fire—even under sequences of drought and moderate years occurring in a sporadic pattern. The UAV-based analysis corroborated the satellite observations in detail and enabled an accurate inventory of damaged trees as well as the exploration of their spatial distribution. The proposed methodology provides an efficient, cost-effective, and operational tool for multi-scale monitoring of forest damage, contributing to the timely recognition of climate-change impacts and to the substantiation of targeted forest management interventions. Full article
(This article belongs to the Section Biogeosciences Remote Sensing)
Show Figures

Figure 1

29 pages, 7593 KB  
Article
UAV-Based Visual Detection and Tracking of Drowning Victims in Maritime Rescue Operations
by Thanh Binh Ngo, Long Ngo, Danh Thanh Nguyen, Anh Vu Phi, Asanka Perera and Andy Nguyen
Drones 2026, 10(2), 146; https://doi.org/10.3390/drones10020146 - 19 Feb 2026
Cited by 1 | Viewed by 1321
Abstract
Maritime search and rescue (SAR) operations are challenged by vast search areas, poor visibility, and the time-critical nature of victim survival, particularly in dynamic coastal areas. This study presents an intelligent unmanned aerial vehicle (UAV) framework for real-time detection, tracking, and prioritization of [...] Read more.
Maritime search and rescue (SAR) operations are challenged by vast search areas, poor visibility, and the time-critical nature of victim survival, particularly in dynamic coastal areas. This study presents an intelligent unmanned aerial vehicle (UAV) framework for real-time detection, tracking, and prioritization of people in distress at sea. Unlike existing UAV-based SAR systems that rely on visual sensing or offline human intervention, the proposed framework integrates RGB-thermal multimodal sensing and posture recognition to enhance victim prioritization and survivability estimation. Visual-thermal data support human posture detection, inference of physiological indicators, and autonomous UAV navigation. Metadata are transmitted to a ground control station to enable adaptive altitude control, trajectory rejoining, and multi-target prioritization. Field-inspired experiments in Quang Ninh Province, Vietnam demonstrated robust real-time performance, achieving 23 FPS with detection accuracy up to 84% for swimming subjects and over 50% for drowning postures. These findings demonstrate that Edge-AI-enabled UAVs can serve as a practical and efficient solution for maritime SAR, reducing response times and improving mission outcomes. Full article
Show Figures

Figure 1

28 pages, 6880 KB  
Article
Non-Appearance-Based Discrimination of UAVs and Birds in Optical Remote Sensing: Using Kinematic and Time–Frequency Features
by Yifei Yao, Jiazhou Geng, Guiting Chen, Tao Lei, Lvjiyuan Jiang and Yi Cui
Drones 2026, 10(2), 98; https://doi.org/10.3390/drones10020098 - 29 Jan 2026
Viewed by 670
Abstract
Unmanned aerial vehicles (UAVs) and birds are typical low-altitude small targets in optical remote sensing, often occupying only a few pixels and exhibiting highly similar appearances, which limits the effectiveness of appearance-based discrimination at long distances and low resolutions. To overcome this, we [...] Read more.
Unmanned aerial vehicles (UAVs) and birds are typical low-altitude small targets in optical remote sensing, often occupying only a few pixels and exhibiting highly similar appearances, which limits the effectiveness of appearance-based discrimination at long distances and low resolutions. To overcome this, we propose a non-appearance-based classification framework using kinematic and time–frequency features. At the trajectory level, kinematic features—including the coefficient of variation of velocity and acceleration, the Spatiotemporal Box-counting Fractal Dimension (SBFD), and the Local Higuchi Fractal Dimension (LHFD)—quantify multi-scale trajectory complexity. At the scale-variation level, time–frequency features, specifically the Time-Frequency Aware Singular Value Entropy (TF-SVE) derived from bounding-box area sequences, capture non-stationary oscillations from bird wing flapping, reflecting behavioral differences from rigid UAV motion. Experiments on a complex real-world dataset show that stacking these features achieves 99.47% classification accuracy, demonstrating a robust, resolution-invariant, and practically effective approach for non-appearance-based recognition of low-altitude targets. Full article
Show Figures

Figure 1

19 pages, 42892 KB  
Article
DMR-YOLO: An Improved Wind Turbine Blade Surface Damage Detection Method Based on YOLOv8
by Lijuan Shi, Sifan Wang, Jian Zhao, Zhejun Kuang, Liu Wang, Lintao Ma, Han Yang and Haiyan Wang
Appl. Sci. 2026, 16(3), 1333; https://doi.org/10.3390/app16031333 - 28 Jan 2026
Cited by 2 | Viewed by 642
Abstract
Wind turbine blades (WTBs) are inevitably exposed to harsh environmental conditions, leading to surface damages such as cracks and corrosion that compromise power generation efficiency. While UAV-based inspection offers significant potential, it frequently encounters challenges in handling irregular defect shapes and preserving fine [...] Read more.
Wind turbine blades (WTBs) are inevitably exposed to harsh environmental conditions, leading to surface damages such as cracks and corrosion that compromise power generation efficiency. While UAV-based inspection offers significant potential, it frequently encounters challenges in handling irregular defect shapes and preserving fine edge details. To address these limitations, this paper proposes DMR-YOLO, an Improved Wind Turbine Blade Surface Damage Detection Method Based on YOLOv8. The proposed framework incorporates three key innovations: First, a C2f-DCNv2-MPCA module is designed to dynamically adjust feature weights, enabling the model to more effectively focus on the geometric structural details of irregular defects. Secondly, a Multi-Scale Edge Perception Enhancement (MEPE) module is introduced to extract edge textures directly within the network. This approach prevents the decoupling of edge features from global context information, effectively resolving the issue of edge information loss and enhancing the recognition of small targets. Finally, the detection head is optimized using a Re-parameterized Shared Convolution Detection Head (RSCD) strategy. By employing weight sharing combined with Diverse Branch Blocks (DBB), this design significantly reduces computational redundancy while maintaining high localization accuracy. Experimental results demonstrate that DMR-YOLO outperforms the baseline YOLOv8n, achieving a 1.8% increase in mAP@0.5 to 82.2%, with a notable 3.2% improvement in the “damage” category. Furthermore, the computational load is reduced by 9.9% to 7.3 GFLOPs, while maintaining an inference speed of 92.6 FPS, providing an effective solution for real-time wind farm defect detection. Full article
Show Figures

Figure 1

24 pages, 15789 KB  
Data Descriptor
Multi-Background UAV Spraying Behavior Recognition Dataset for Precision Agriculture
by Chang Meng, Lei Shu and Leijing Bai
J. Sens. Actuator Netw. 2026, 15(1), 14; https://doi.org/10.3390/jsan15010014 - 26 Jan 2026
Viewed by 1409
Abstract
The rapid growth of precision agriculture has accelerated the deployment of plant protection unmanned aerial vehicles (UAVs). However, reliable data resources for vision-based intelligent supervision of operational states, such as whether a UAV is currently spraying, remain limited. Most publicly available UAV detection [...] Read more.
The rapid growth of precision agriculture has accelerated the deployment of plant protection unmanned aerial vehicles (UAVs). However, reliable data resources for vision-based intelligent supervision of operational states, such as whether a UAV is currently spraying, remain limited. Most publicly available UAV detection datasets target urban security and surveillance scenarios, where annotations emphasize object localization rather than agricultural operation state recognition, making them insufficient for farmland spraying supervision. Therefore, agricultural-oriented data resources are needed to cover diverse backgrounds and include operation state labels, thereby supporting both academic research and practical deployment. In this study, we construct and release the first multi-background dataset dedicated to agricultural UAV spraying behavior recognition. The dataset contains 9548 high-quality annotated images spanning the following six typical backgrounds: green cropland, bare farmland, orchard, woodland, mountainous terrain, and sky. For each UAV instance, we provide both a bounding box and a binary operation state label, namely spraying and flying without spraying. We further conduct systematic benchmark evaluations of mainstream object detection algorithms on this dataset. The dataset captures agriculture-specific challenges, including a high proportion of small objects, substantial scale variation, motion blur, and complex dynamic backgrounds, and can be used to assess algorithm robustness in real-world agricultural settings. Benchmark results show that YOLOv5n achieves the best overall performance, with an accuracy of 97.86% and an mAP@50 of 98.30%. This dataset provides critical data support for automated supervision of plant protection UAV spraying operations and precision agriculture monitoring platforms. Full article
(This article belongs to the Special Issue AI-Assisted Machine-Environment Interaction)
Show Figures

Figure 1

36 pages, 4183 KB  
Article
Distinguishing a Drone from Birds Based on Trajectory Movement and Deep Learning
by Andrii Nesteruk, Valerii Nikitin, Yosyp Albrekht, Łukasz Ścisło, Damian Grela and Paweł Król
Sensors 2026, 26(3), 755; https://doi.org/10.3390/s26030755 - 23 Jan 2026
Viewed by 1127
Abstract
Unmanned aerial vehicles (UAVs) increasingly share low-altitude airspace with birds, making early distinguishing between drones and biological targets critical for safety and security. This work addresses long-range scenarios where objects occupy only a few pixels and appearance-based recognition becomes unreliable. We develop a [...] Read more.
Unmanned aerial vehicles (UAVs) increasingly share low-altitude airspace with birds, making early distinguishing between drones and biological targets critical for safety and security. This work addresses long-range scenarios where objects occupy only a few pixels and appearance-based recognition becomes unreliable. We develop a model-driven simulation pipeline that generates synthetic data with a controlled camera model, atmospheric background and realistic motion of three aerial target types: multicopter, fixed-wing UAV and bird. From these sequences, each track is encoded as a time series of image-plane coordinates and apparent size, and a bidirectional long short-term memory (LSTM) network is trained to classify trajectories as drone-like or bird-like. The model learns characteristic differences in smoothness, turning behavior and velocity fluctuations, and to achieve reliable separation between drone and bird motion patterns on synthetic test data. Motion-trajectory cues alone can support early distinguishing of drones from birds when visual details are scarce, providing a complementary signal to conventional image-based detection. The proposed synthetic data and sequence classification pipeline forms a reproducible testbed that can be extended with real trajectories from radar or video tracking systems and used to prototype and benchmark trajectory-based recognizers for integrated surveillance solutions. The proposed method is designed to generalize naturally to real surveillance systems, as it relies on trajectory-level motion patterns rather than appearance-based features that are sensitive to sensor quality, illumination, or weather conditions. Full article
(This article belongs to the Section Industrial Sensors)
Show Figures

Graphical abstract

17 pages, 4414 KB  
Article
Fast Helmet Detection in Low-Resolution Surveillance via Super-Resolution and ROI-Guided Inference
by Taiming He, Ziyue Wang and Lu Yang
Appl. Sci. 2026, 16(2), 967; https://doi.org/10.3390/app16020967 - 17 Jan 2026
Viewed by 548
Abstract
Reliable detection of safety helmets is essential for ensuring personnel protection in large-scale outdoor operations. However, recognition becomes difficult when monitoring relies on low-resolution or compressed video streams captured by fixed or mobile platforms such as UAVs—conditions commonly encountered in intelligent transportation and [...] Read more.
Reliable detection of safety helmets is essential for ensuring personnel protection in large-scale outdoor operations. However, recognition becomes difficult when monitoring relies on low-resolution or compressed video streams captured by fixed or mobile platforms such as UAVs—conditions commonly encountered in intelligent transportation and urban surveillance. This study proposes a super-resolution-enhanced detection framework that integrates video super-resolution with ROI-guided inference to improve the visibility of small targets while reducing computational cost. Focusing on a single, carefully selected VSR module (BasicVSR++), the framework achieves an F1-score of 0.904 in helmet detection across multiple low-quality surveillance scenarios. This demonstrates the framework’s effectiveness for robust helmet monitoring in low-resolution and compressed surveillance scenarios. Full article
(This article belongs to the Section Transportation and Future Mobility)
Show Figures

Figure 1

Back to TopTop