Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (64)

Search Parameters:
Keywords = light and small drones

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 3823 KiB  
Article
Lightweight UAV-Based System for Early Fire-Risk Identification in Wild Forests
by Akmalbek Abdusalomov, Sabina Umirzakova, Alpamis Kutlimuratov, Dilshod Mirzaev, Adilbek Dauletov, Tulkin Botirov, Madina Zakirova, Mukhriddin Mukhiddinov and Young Im Cho
Fire 2025, 8(8), 288; https://doi.org/10.3390/fire8080288 - 23 Jul 2025
Viewed by 322
Abstract
The escalating impacts and occurrence of wildfires threaten the public, economies, and global ecosystems. Physiologically declining or dead trees are a great portion of the fires because these trees are prone to higher ignition and have lower moisture content. To prevent wildfires, hazardous [...] Read more.
The escalating impacts and occurrence of wildfires threaten the public, economies, and global ecosystems. Physiologically declining or dead trees are a great portion of the fires because these trees are prone to higher ignition and have lower moisture content. To prevent wildfires, hazardous vegetation needs to be removed, and the vegetation should be identified early on. This work proposes a real-time fire risk tree detection framework using UAV images, which is based on lightweight object detection. The model uses the MobileNetV3-Small spine, which is optimized for edge deployment, combined with an SSD head. This configuration results in a highly optimized and fast UAV-based inference pipeline. The dataset used in this study comprises over 3000 annotated RGB UAV images of trees in healthy, partially dead, and fully dead conditions, collected from mixed real-world forest scenes and public drone imagery repositories. Thorough evaluation shows that the proposed model outperforms conventional SSD and recent YOLOs on Precision (94.1%), Recall (93.7%), mAP (90.7%), F1 (91.0%) while being light-weight (8.7 MB) and fast (62.5 FPS on Jetson Xavier NX). These findings strongly support the model’s effectiveness for large-scale continuous forest monitoring to detect health degradations and mitigate wildfire risks proactively. The framework UAV-based environmental monitoring systems differentiates itself by incorporating a balance between detection accuracy, speed, and resource efficiency as fundamental principles. Full article
Show Figures

Figure 1

21 pages, 12122 KiB  
Article
RA3T: An Innovative Region-Aligned 3D Transformer for Self-Supervised Sim-to-Real Adaptation in Low-Altitude UAV Vision
by Xingrao Ma, Jie Xie, Di Shao, Aiting Yao and Chengzu Dong
Electronics 2025, 14(14), 2797; https://doi.org/10.3390/electronics14142797 - 11 Jul 2025
Viewed by 268
Abstract
Low-altitude unmanned aerial vehicle (UAV) vision is critically hindered by the Sim-to-Real Gap, where models trained exclusively on simulation data degrade under real-world variations in lighting, texture, and weather. To address this problem, we propose RA3T (Region-Aligned 3D Transformer), a novel self-supervised framework [...] Read more.
Low-altitude unmanned aerial vehicle (UAV) vision is critically hindered by the Sim-to-Real Gap, where models trained exclusively on simulation data degrade under real-world variations in lighting, texture, and weather. To address this problem, we propose RA3T (Region-Aligned 3D Transformer), a novel self-supervised framework that enables robust Sim-to-Real adaptation. Specifically, we first develop a dual-branch strategy for self-supervised feature learning, integrating Masked Autoencoders and contrastive learning. This approach extracts domain-invariant representations from unlabeled simulated imagery to enhance robustness against occlusion while reducing annotation dependency. Leveraging these learned features, we then introduce a 3D Transformer fusion module that unifies multi-view RGB and LiDAR point clouds through cross-modal attention. By explicitly modeling spatial layouts and height differentials, this component significantly improves recognition of small and occluded targets in complex low-altitude environments. To address persistent fine-grained domain shifts, we finally design region-level adversarial calibration that deploys local discriminators on partitioned feature maps. This mechanism directly aligns texture, shadow, and illumination discrepancies which challenge conventional global alignment methods. Extensive experiments on UAV benchmarks VisDrone and DOTA demonstrate the effectiveness of RA3T. The framework achieves +5.1% mAP on VisDrone and +7.4% mAP on DOTA over the 2D adversarial baseline, particularly on small objects and sparse occlusions, while maintaining real-time performance of 17 FPS at 1024 × 1024 resolution on an RTX 4080 GPU. Visual analysis confirms that the synergistic integration of 3D geometric encoding and local adversarial alignment effectively mitigates domain gaps caused by uneven illumination and perspective variations, establishing an efficient pathway for simulation-to-reality UAV perception. Full article
(This article belongs to the Special Issue Innovative Technologies and Services for Unmanned Aerial Vehicles)
Show Figures

Figure 1

18 pages, 4939 KiB  
Article
LiDAR-Based Detection of Field Hamster (Cricetus cricetus) Burrows in Agricultural Fields
by Florian Thürkow, Milena Mohri, Jonas Ramstetter and Philipp Alb
Sustainability 2025, 17(14), 6366; https://doi.org/10.3390/su17146366 - 11 Jul 2025
Viewed by 271
Abstract
Farmers face increasing pressure to maintain vital populations of the critically endangered field hamster (Cricetus cricetus) while managing crop damage caused by field mice. This challenge is linked to the UN Sustainable Development Goals (SDGs) 2 and 15, addressing food security [...] Read more.
Farmers face increasing pressure to maintain vital populations of the critically endangered field hamster (Cricetus cricetus) while managing crop damage caused by field mice. This challenge is linked to the UN Sustainable Development Goals (SDGs) 2 and 15, addressing food security and biodiversity. Consequently, the reliable detection of hamster activity in agricultural fields is essential. While remote sensing offers potential for wildlife monitoring, commonly used RGB imagery has limitations in detecting small burrow entrances in vegetated areas. This study investigates the potential of drone-based Light Detection and Ranging (LiDAR) data for identifying field hamster burrow entrances in agricultural landscapes. A geostatistical method was developed to detect local elevation minima as indicators of burrow openings. The analysis used four datasets captured at varying flight altitudes and spatial resolutions. The method successfully detected up to 20 out of 23 known burrow entrances and achieved an F1-score of 0.83 for the best-performing dataset. Detection was most accurate at flight altitudes of 30 m or lower, with performance decreasing at higher altitudes due to reduced point density. These findings demonstrate the potential of UAV-based LiDAR to support non-invasive species monitoring and habitat management in agricultural systems, contributing to sustainable conservation practices in line with the SDGs. Full article
(This article belongs to the Special Issue Ecology, Biodiversity and Sustainable Conservation)
Show Figures

Figure 1

29 pages, 44456 KiB  
Article
AUHF-DETR: A Lightweight Transformer with Spatial Attention and Wavelet Convolution for Embedded UAV Small Object Detection
by Hengyu Guo, Qunyong Wu and Yuhang Wang
Remote Sens. 2025, 17(11), 1920; https://doi.org/10.3390/rs17111920 - 31 May 2025
Cited by 1 | Viewed by 849
Abstract
Real-time object detection on embedded unmanned aerial vehicles (UAVs) is crucial for emergency rescue, autonomous driving, and target tracking applications. However, UAVs’ hardware limitations create conflicts between model size and detection accuracy. Moreover, challenges such as complex backgrounds from the UAV’s perspective, severe [...] Read more.
Real-time object detection on embedded unmanned aerial vehicles (UAVs) is crucial for emergency rescue, autonomous driving, and target tracking applications. However, UAVs’ hardware limitations create conflicts between model size and detection accuracy. Moreover, challenges such as complex backgrounds from the UAV’s perspective, severe occlusion, densely packed small targets, and uneven lighting conditions complicate real-time detection for embedded UAVs. To tackle these challenges, we propose AUHF-DETR, an embedded detection model derived from RT-DETR. In the backbone, we introduce a novel WTC-AdaResNet paradigm that utilizes reversible connections to decouple small-object features. We further replace the original global attention mechanism with the PSA module to strengthen inter-feature relationships within each ROI, thereby resolving the embedded challenges posed by RT-DETR’s complex token computations. In the encoder, we introduce a BDFPN for multi-scale feature fusion, effectively mitigating the small-object detection difficulties caused by the baseline’s Hungarian assignment. Extensive experiments on the public VisDrone2019, HIT-UAV, and CARPK datasets demonstrate that compared with RT-DETR-r18, AUHF-DETR achieves a 2.1% increase in APs on VisDrone2019, reduces the parameter count by 49.0%, and attains 68 FPS (AGX Xavier), thus satisfying the real-time requirements for small-object detection in embedded UAVs. Full article
Show Figures

Figure 1

23 pages, 7532 KiB  
Article
Real-Time Aerial Multispectral Object Detection with Dynamic Modality-Balanced Pixel-Level Fusion
by Zhe Wang and Qingling Zhang
Sensors 2025, 25(10), 3039; https://doi.org/10.3390/s25103039 - 12 May 2025
Viewed by 712
Abstract
Aerial object detection plays a critical role in numerous fields, utilizing the flexibility of airborne platforms to achieve real-time tasks. Combining visible and infrared sensors can overcome limitations under low-light conditions, enabling full-time tasks. While feature-level fusion methods exhibit comparable performances in visible–infrared [...] Read more.
Aerial object detection plays a critical role in numerous fields, utilizing the flexibility of airborne platforms to achieve real-time tasks. Combining visible and infrared sensors can overcome limitations under low-light conditions, enabling full-time tasks. While feature-level fusion methods exhibit comparable performances in visible–infrared multispectral object detection, they suffer from heavy model size, inadequate inference speed and visible light preferences caused by inherent modality imbalance, limiting their applications in airborne platform deployment. To address these challenges, this paper proposes a YOLO-based real-time multispectral fusion framework combining pixel-level fusion with dynamic modality-balanced augmentation called Full-time Multispectral Pixel-wise Fusion Network (FMPFNet). Firstly, we introduce the Multispectral Luminance Weighted Fusion (MLWF) module consisting of attention-based modality reconstruction and feature fusion. By leveraging YUV color space transformation, this module efficiently fuses RGB and IR modalities while minimizing computational overhead. We also propose the Dynamic Modality Dropout and Threshold Masking (DMDTM) strategy, which balances modality attention and improves detection performance in low-light scenarios. Additionally, we refine our model to enhance the detection of small rotated objects, a requirement commonly encountered in aerial detection applications. Experimental results on the DroneVehicle dataset demonstrate that our FMPFNet achieves 76.80% mAP50 and 132 FPS, outperforming state-of-the-art feature-level fusion methods in both accuracy and inference speed. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

24 pages, 11423 KiB  
Article
YOLO-UFS: A Novel Detection Model for UAVs to Detect Early Forest Fires
by Zitong Luo, Haining Xu, Yanqiu Xing, Chuanhao Zhu, Zhupeng Jiao and Chengguo Cui
Forests 2025, 16(5), 743; https://doi.org/10.3390/f16050743 - 26 Apr 2025
Cited by 2 | Viewed by 815
Abstract
Forest fires endanger ecosystems and human life, making early detection crucial for effective prevention. Traditional detection methods are often inadequate due to large coverage areas and inherent limitations. However, drone technology combined with deep learning holds promise. This study investigates using small drones [...] Read more.
Forest fires endanger ecosystems and human life, making early detection crucial for effective prevention. Traditional detection methods are often inadequate due to large coverage areas and inherent limitations. However, drone technology combined with deep learning holds promise. This study investigates using small drones equipped with lightweight deep learning models to detect forest fires early. A high-quality dataset constructed through aerial image analysis supports robust model training. The proposed YOLO-UFS network, based on YOLOv5s, integrates enhancements such as the C3-MNV4 module, BiFPN, AF-IoU loss function, and NAM attention mechanism. These modifications achieve a 91.3% mAP on the self-built early forest fire dataset. Compared to the original model, YOLO-UFS improves accuracy by 3.8%, recall by 4.1%, and average accuracy by 3.2%, while reducing computational parameters by 74.7% and 78.3%. It outperforms other mainstream YOLO algorithms on drone platforms, balancing accuracy and real-time performance. In generalization experiments using public datasets, the model’s mAP0.5 increased from 85.2% to 86.3%, and mAP0.5:0.95 from 56.7% to 57.9%, with an overall mAP gain of 3.3%. The optimized model runs efficiently on the Jetson Nano platform with 258 GB of RAM, 7.4 MB of storage memory, and an average frame rate of 30 FPS. In this study, airborne visible light images are used to provide a low-cost and high-precision solution for the early detection of forest fires, so that low-computing UAVs can achieve the requirements of early detection, early mobilization, and early extinguishment. Future work will focus on multi-sensor data fusion and human–robot collaboration to further improve the accuracy and reliability of detection. Full article
(This article belongs to the Section Natural Hazards and Risk Management)
Show Figures

Figure 1

30 pages, 225854 KiB  
Article
LGWheatNet: A Lightweight Wheat Spike Detection Model Based on Multi-Scale Information Fusion
by Zhaomei Qiu, Fei Wang, Tingting Li, Chongjun Liu, Xin Jin, Shunhao Qing, Yi Shi, Yuntao Wu and Congbin Liu
Plants 2025, 14(7), 1098; https://doi.org/10.3390/plants14071098 - 2 Apr 2025
Cited by 1 | Viewed by 893
Abstract
Wheat spike detection holds significant importance for agricultural production as it enhances the efficiency of crop management and the precision of operations. This study aims to improve the accuracy and efficiency of wheat spike detection, enabling efficient crop monitoring under resource-constrained conditions. To [...] Read more.
Wheat spike detection holds significant importance for agricultural production as it enhances the efficiency of crop management and the precision of operations. This study aims to improve the accuracy and efficiency of wheat spike detection, enabling efficient crop monitoring under resource-constrained conditions. To this end, a wheat spike dataset encompassing multiple growth stages was constructed, leveraging the advantages of MobileNet and ShuffleNet to design a novel network module, SeCUIB. Building on this foundation, a new wheat spike detection network, LGWheatNet, was proposed by integrating a lightweight downsampling module (DWDown), spatial pyramid pooling (SPPF), and a lightweight detection head (LightDetect). The experimental results demonstrate that LGWheatNet excels in key performance metrics, including Precision, Recall, and Mean Average Precision (mAP50 and mAP50-95). Specifically, the model achieved a Precision of 0.956, a Recall of 0.921, an mAP50 of 0.967, and an mAP50-95 of 0.747, surpassing several YOLO models as well as EfficientDet and RetinaNet. Furthermore, LGWheatNet demonstrated superior resource efficiency with a parameter count of only 1,698,529 and GFLOPs of 5.0, significantly lower than those of competing models. Additionally, when combined with the Slicing Aided Hyper Inference strategy, LGWheatNet further improved the detection accuracy of wheat spikes, especially for small-scale targets and edge regions, when processing large-scale high-resolution images. This strategy significantly enhanced both inference efficiency and accuracy, making it particularly suitable for image analysis from drone-captured data. In wheat spike counting experiments, LGWheatNet also delivered exceptional performance, particularly in predictions during the filling and maturity stages, outperforming other models by a substantial margin. This study not only provides an efficient and reliable solution for wheat spike detection but also introduces innovative methods for lightweight object detection tasks in resource-constrained environments. Full article
Show Figures

Figure 1

25 pages, 14872 KiB  
Article
RFAG-YOLO: A Receptive Field Attention-Guided YOLO Network for Small-Object Detection in UAV Images
by Chengmeng Wei and Wenhong Wang
Sensors 2025, 25(7), 2193; https://doi.org/10.3390/s25072193 - 30 Mar 2025
Cited by 5 | Viewed by 1698
Abstract
The YOLO series of object detection methods have achieved significant success in a wide range of computer vision tasks due to their efficiency and accuracy. However, detecting small objects in UAV images remains a formidable challenge due to factors such as a low [...] Read more.
The YOLO series of object detection methods have achieved significant success in a wide range of computer vision tasks due to their efficiency and accuracy. However, detecting small objects in UAV images remains a formidable challenge due to factors such as a low resolution, complex background interference, and significant scale variations, which collectively degrade the quality of feature extraction and limit detection performance. To address these challenges, we propose the receptive field attention-guided YOLO (RFAG-YOLO) method, an advanced adaptation of YOLOv8 tailored for small-object detection in UAV imagery, with a focus on improving feature representation and detection robustness. To this end, we introduce a novel network building block, termed the receptive field network block (RFN block), which leverages dynamic kernel parameter adjustments to enhance the model’s ability to capture fine-grained local details. To effectively harness multi-scale features, we designed an enhanced FasterNet module based on RFN blocks as the core component of the backbone network in RFAG-YOLO, enabling robust feature extraction across varying resolutions. This approach achieves a balance of semantic information by employing staged downsampling and a hierarchical arrangement of RFN blocks, ensuring optimal feature representation across different resolutions. Additionally, we introduced a Scale-Aware Feature Amalgamation (SAF) component prior to the detection head of RFAG-YOLO. This component employs a scale attention mechanism to dynamically weight features from both higher and lower layers, facilitating richer information flow and significantly improving the model’s robustness to complex backgrounds and scale variations. Experimental results on the VisDrone2019 dataset demonstrated that RFAG-YOLO outperformed state-of-the-art models, including YOLOv7, YOLOv8, YOLOv10, and YOLOv11, in terms of detection accuracy and efficiency. In particular, RFAG-YOLO achieved an mAP50 of 38.9%, representing substantial improvements over multiple baseline models: a 12.43% increase over YOLOv7, a 5.99% improvement over YOLOv10, and significant gains of 16.12% compared to YOLOv8n and YOLOv11. Moreover, compared to the larger YOLOv8s model, RFAG-YOLO achieved 97.98% of its mAP50 performance while utilizing only 53.51% of the parameters, highlighting its exceptional efficiency in terms of the performance-to-parameter ratio and making it highly suitable for resource-constrained UAV applications. These results underscore the substantial potential of RFAG-YOLO for real-world UAV applications, particularly in scenarios demanding accurate detection of small objects under challenging conditions such as varying lighting, complex backgrounds, and diverse scales. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

23 pages, 28505 KiB  
Article
Drone-Based Detection and Classification of Greater Caribbean Manatees in the Panama Canal Basin
by Javier E. Sanchez-Galan, Kenji Contreras, Allan Denoce, Héctor Poveda, Fernando Merchan and Hector M. Guzmán
Drones 2025, 9(4), 230; https://doi.org/10.3390/drones9040230 - 21 Mar 2025
Viewed by 844
Abstract
This study introduces a novel, drone-based approach for the detection and classification of Greater Caribbean Manatees (Trichechus manatus manatus) in the Panama Canal Basin by integrating advanced deep learning techniques. Leveraging the high-performance YOLOv8 model augmented with Sliced Aided Hyper Inferencing (SAHI) for [...] Read more.
This study introduces a novel, drone-based approach for the detection and classification of Greater Caribbean Manatees (Trichechus manatus manatus) in the Panama Canal Basin by integrating advanced deep learning techniques. Leveraging the high-performance YOLOv8 model augmented with Sliced Aided Hyper Inferencing (SAHI) for improved small-object detection, our system accurately identifies individual manatees, mother–calf pairs, and group formations across a challenging aquatic environment. Additionally, the use of AltCLIP for zero-shot classification enables robust demographic analysis without extensive labeled data, enhancing model adaptability in data-scarce scenarios. For this study, more than 57,000 UAV images were acquired from multiple drone flights covering diverse regions of Gatun Lake and its surroundings. In cross-validation experiments, the detection model achieved precision levels as high as 93% and mean average precision (mAP) values exceeding 90% under ideal conditions. However, testing on unseen data revealed a lower recall, highlighting challenges in detecting manatees under variable altitudes and adverse lighting conditions. Furthermore, the integrated zero-shot classification approach demonstrated a robust top-2 accuracy close to 90%, effectively categorizing manatee demographic groupings despite overlapping visual features. This work presents a deep learning framework integrated with UAV technology, offering a scalable, non-invasive solution for real-time wildlife monitoring. By enabling precise detection and classification, it lays the foundation for enhanced habitat assessments and more effective conservation planning in similar tropical wetland ecosystems. Full article
Show Figures

Figure 1

26 pages, 29021 KiB  
Article
Efficient Coastal Mangrove Species Recognition Using Multi-Scale Features Enhanced by Multi-Head Attention
by Shaolin Guo, Yixuan Wang, Yin Tan, Tonglai Liu and Qin Qin
Symmetry 2025, 17(3), 461; https://doi.org/10.3390/sym17030461 - 19 Mar 2025
Cited by 1 | Viewed by 464
Abstract
Recognizing mangrove species is a challenging task in coastal wetland ecological monitoring due to the complex environment, high species similarity, and the inherent symmetry within the structural features of mangrove species. Many species coexist, exhibiting only subtle differences in leaf shape and color, [...] Read more.
Recognizing mangrove species is a challenging task in coastal wetland ecological monitoring due to the complex environment, high species similarity, and the inherent symmetry within the structural features of mangrove species. Many species coexist, exhibiting only subtle differences in leaf shape and color, which increases the risk of misclassification. Additionally, mangroves grow in intertidal environments with varying light conditions and surface reflections, further complicating feature extraction. Small species are particularly hard to distinguish in dense vegetation due to their symmetrical features that are difficult to differentiate at the pixel level. While hyperspectral imaging offers some advantages in species recognition, its high equipment costs and data acquisition complexity limit its practical application. To address these challenges, we propose MHAGFNet, a segmentation-based mangrove species recognition network. The network utilizes easily accessible RGB remote sensing images captured by drones, ensuring efficient data collection. MHAGFNet integrates a Multi-Scale Feature Fusion Module (MSFFM) and a Multi-Head Attention Guide Module (MHAGM), which enhance species recognition by improving feature capture across scales and integrating both global and local details. In this study, we also introduce MSIDBG, a dataset created using high-resolution UAV images from the Shankou Mangrove National Nature Reserve in Beihai, China. Extensive experiments demonstrate that MHAGFNet significantly improves accuracy and robustness in mangrove species recognition. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

22 pages, 52708 KiB  
Article
CSMR: A Multi-Modal Registered Dataset for Complex Scenarios
by Chenrui Li, Kun Gao, Zibo Hu, Zhijia Yang, Mingfeng Cai, Haobo Cheng and Zhenyu Zhu
Remote Sens. 2025, 17(5), 844; https://doi.org/10.3390/rs17050844 - 27 Feb 2025
Viewed by 974
Abstract
Complex scenarios pose challenges to tasks in computer vision, including image fusion, object detection, and image-to-image translation. On the one hand, complex scenarios involve fluctuating weather or lighting conditions, where even images of the same scenarios appear to be different. On the other [...] Read more.
Complex scenarios pose challenges to tasks in computer vision, including image fusion, object detection, and image-to-image translation. On the one hand, complex scenarios involve fluctuating weather or lighting conditions, where even images of the same scenarios appear to be different. On the other hand, the large amount of textural detail in the given images introduces considerable interference that can conceal the useful information contained in them. An effective solution to these problems is to use the complementary details present in multi-modal images, such as visible-light and infrared images. Visible-light images contain rich textural information while infrared images contain information about the temperature. In this study, we propose a multi-modal registered dataset for complex scenarios under various environmental conditions, targeting security surveillance and the monitoring of low-slow-small targets. Our dataset contains 30,819 images, where the targets are labeled as three classes of “person”, “car”, and “drone” using Yolo format bounding boxes. We compared our dataset with those used in the literature for computer vision-related tasks, including image fusion, object detection, and image-to-image translation. The results showed that introducing complementary information through image fusion can compensate for missing details in the original images, and we also revealed the limitations of visual tasks in single-modal images with complex scenarios. Full article
(This article belongs to the Special Issue Recent Advances in Infrared Target Detection)
Show Figures

Figure 1

22 pages, 25824 KiB  
Article
NoctuDroneNet: Real-Time Semantic Segmentation of Nighttime UAV Imagery in Complex Environments
by Ruokun Qu, Jintao Tan, Yelu Liu, Chenglong Li and Hui Jiang
Drones 2025, 9(2), 97; https://doi.org/10.3390/drones9020097 - 27 Jan 2025
Viewed by 1110
Abstract
Nighttime semantic segmentation represents a challenging frontier in computer vision, made particularly difficult by severe low-light conditions, pronounced noise, and complex illumination patterns. These challenges intensify when dealing with Unmanned Aerial Vehicle (UAV) imagery, where varying camera angles and altitudes compound the difficulty. [...] Read more.
Nighttime semantic segmentation represents a challenging frontier in computer vision, made particularly difficult by severe low-light conditions, pronounced noise, and complex illumination patterns. These challenges intensify when dealing with Unmanned Aerial Vehicle (UAV) imagery, where varying camera angles and altitudes compound the difficulty. In this paper, we introduce NoctuDroneNet (Nocturnal UAV Drone Network, hereinafter referred to as NoctuDroneNet), a real-time segmentation model tailored specifically for nighttime UAV scenarios. Our approach integrates convolution-based global reasoning with training-only semantic alignment modules to effectively handle diverse and extreme nighttime conditions. We construct a new dataset, NUI-Night, focusing on low-illumination UAV scenes to rigorously evaluate performance under conditions rarely represented in standard benchmarks. Beyond NUI-Night, we assess NoctuDroneNet on the Varied Drone Dataset (VDD), a normal-illumination UAV dataset, demonstrating the model’s robustness and adaptability to varying flight domains despite the lack of large-scale low-light UAV benchmarks. Furthermore, evaluations on the Night-City dataset confirm its scalability and applicability to complex nighttime urban environments. NoctuDroneNet achieves state-of-the-art performance on NUI-Night, surpassing strong real-time baselines in both segmentation accuracy and speed. Qualitative analyses highlight its resilience to under-/over-exposure and small-object detection, underscoring its potential for real-world applications like UAV emergency landings under minimal illumination. Full article
Show Figures

Figure 1

23 pages, 5215 KiB  
Article
A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism
by Zhe Quan and Jun Sun
Sensors 2025, 25(2), 589; https://doi.org/10.3390/s25020589 - 20 Jan 2025
Viewed by 2480
Abstract
With the rapid development of AI algorithms and computational power, object recognition based on deep learning frameworks has become a major research direction in computer vision. UAVs equipped with object detection systems are increasingly used in fields like smart transportation, disaster warning, and [...] Read more.
With the rapid development of AI algorithms and computational power, object recognition based on deep learning frameworks has become a major research direction in computer vision. UAVs equipped with object detection systems are increasingly used in fields like smart transportation, disaster warning, and emergency rescue. However, due to factors such as the environment, lighting, altitude, and angle, UAV images face challenges like small object sizes, high object density, and significant background interference, making object detection tasks difficult. To address these issues, we use YOLOv8s as the basic framework and introduce a multi-level feature fusion algorithm. Additionally, we design an attention mechanism that links distant pixels to improve small object feature extraction. To address missed detections and inaccurate localization, we replace the detection head with a dynamic head, allowing the model to route objects to the appropriate head for final output. We also introduce Slideloss to improve the model’s learning of difficult samples and ShapeIoU to better account for the shape and scale of bounding boxes. Experiments on datasets like VisDrone2019 show that our method improves accuracy by nearly 10% and recall by about 11% compared to the baseline. Additionally, on the AI-TODv1.5 dataset, our method improves the mAP50 from 38.8 to 45.2. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

26 pages, 394 KiB  
Review
Monitoring Yield and Quality of Forages and Grassland in the View of Precision Agriculture Applications—A Review
by Abid Ali and Hans-Peter Kaul
Remote Sens. 2025, 17(2), 279; https://doi.org/10.3390/rs17020279 - 15 Jan 2025
Cited by 6 | Viewed by 3014
Abstract
The potential of precision agriculture (PA) in forage and grassland management should be more extensively exploited to meet the increasing global food demand on a sustainable basis. Monitoring biomass yield and quality traits directly impacts the fertilization and irrigation practises and frequency of [...] Read more.
The potential of precision agriculture (PA) in forage and grassland management should be more extensively exploited to meet the increasing global food demand on a sustainable basis. Monitoring biomass yield and quality traits directly impacts the fertilization and irrigation practises and frequency of utilization (cuts) in grasslands. Therefore, the main goal of the review is to examine the techniques for using PA applications to monitor productivity and quality in forage and grasslands. To achieve this, the authors discuss several monitoring technologies for biomass and plant stand characteristics (including quality) that make it possible to adopt digital farming in forages and grassland management. The review provides an overview about mass flow and impact sensors, moisture sensors, remote sensing-based approaches, near-infrared (NIR) spectroscopy, and mapping field heterogeneity and promotes decision support systems (DSSs) in this field. At a small scale, advanced sensors such as optical, thermal, and radar sensors mountable on drones; LiDAR (Light Detection and Ranging); and hyperspectral imaging techniques can be used for assessing plant and soil characteristics. At a larger scale, we discuss coupling of remote sensing with weather data (synergistic grassland yield modelling), Sentinel-2 data with radiative transfer modelling (RTM), Sentinel-1 backscatter, and Catboost–machine learning methods for digital mapping in terms of precision harvesting and site-specific farming decisions. It is known that the delineation of sward heterogeneity is more difficult in mixed grasslands due to spectral similarity among species. Thanks to Diversity-Interactions models, jointly assessing various species interactions under mixed grasslands is allowed. Further, understanding such complex sward heterogeneity might be feasible by integrating spectral un-mixing techniques such as the super-pixel segmentation technique, multi-level fusion procedure, and combined NIR spectroscopy with neural network models. This review offers a digital option for enhancing yield monitoring systems and implementing PA applications in forages and grassland management. The authors recommend a future research direction for the inclusion of costs and economic returns of digital technologies for precision grasslands and fodder production. Full article
Show Figures

Graphical abstract

22 pages, 6188 KiB  
Article
Quadrotor Unmanned Aerial Vehicle-Mounted Launch Device Precision Analysis for Countering Intruding Drones
by Yanan He, Lingsong Di, Huiqi Xu, Weigui Zeng, Wei Li and Silei Cao
Drones 2025, 9(1), 42; https://doi.org/10.3390/drones9010042 - 9 Jan 2025
Viewed by 1441
Abstract
In light of the current issue of commercial and modified drones frequently intruding into sensitive locations such as airports, this paper proposes the use of quadrotor drones equipped with a launch device for interception. For an “X”-shaped quadrotor drone equipped with a two-degree-of-freedom [...] Read more.
In light of the current issue of commercial and modified drones frequently intruding into sensitive locations such as airports, this paper proposes the use of quadrotor drones equipped with a launch device for interception. For an “X”-shaped quadrotor drone equipped with a two-degree-of-freedom gimbal launch device, dynamic, control, and ballistic models are constructed to analyze the impact of single-shot firing on the drone’s attitude and the factors affecting accuracy in rapid-fire scenarios. Simulation results indicate that downward firing has the highest shooting accuracy and is suitable for counter-drone missions. For single-shot firing, lateral downward firing compared to frontal firing can effectively reduce attitude and position changes, which is beneficial for engaging stationary targets. In the case of rapid fire, control of the firing interval is crucial for accuracy; a larger firing interval can significantly enhance shooting precision. When firing small payloads in rapid succession, vertical downward firing has the highest accuracy, while lateral firing results in a larger distribution radius of the impact points. To improve shooting accuracy, frontal firing is recommended. Future research will further explore the dynamic response of drones under different firing conditions and develop more advanced control strategies to enhance their practical performance and reliability. Full article
(This article belongs to the Section Drone Design and Development)
Show Figures

Figure 1

Back to TopTop