Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,843)

Search Parameters:
Keywords = UAV datasets

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 2314 KB  
Article
EF-YOLO: Detecting Small Targets in Early-Stage Agricultural Fires via UAV-Based Remote Sensing
by Jun Tao, Zhihan Wang, Jianqiu Wu, Yunqin Li, Tomohiro Fukuda and Jiaxin Zhang
Remote Sens. 2026, 18(8), 1119; https://doi.org/10.3390/rs18081119 - 9 Apr 2026
Abstract
Early detection of agricultural fires with Unmanned Aerial Vehicles (UAVs) is important for environmental safety, yet it remains difficult because ignition cues are extremely small, smoke patterns vary widely, and farmland scenes often contain strong background interference such as specular reflections. Model development [...] Read more.
Early detection of agricultural fires with Unmanned Aerial Vehicles (UAVs) is important for environmental safety, yet it remains difficult because ignition cues are extremely small, smoke patterns vary widely, and farmland scenes often contain strong background interference such as specular reflections. Model development is further constrained by the scarcity of data from the early ignition stage. To address these challenges, we propose a joint data and model optimization framework. We first build a hybrid dataset through an ROI-guided synthesis pipeline, in which latent diffusion models are used to insert high-fidelity, carefully screened fire samples into real farmland backgrounds. We then introduce EF-YOLO, a detector designed for high sensitivity to small targets. The network uses SPD-Conv to reduce feature loss during spatial downsampling and includes a high-resolution P2 head to improve the detection of minute objects. To reduce background clutter, a Dual-Path Frequency–Spatial Enhancement (DP-FSE) module serves as a lightweight statistical surrogate that extracts global contextual cues and local salient features in parallel, thereby suppressing high-frequency noise. Experimental results show that EF-YOLO achieves an APs of 40.2% on sub-pixel targets, exceeding the YOLOv8s baseline by 15.4 percentage points. With a recall of 88.7% and a real-time inference speed of 78 FPS, the proposed framework offers a strong balance between detection performance and efficiency, making it well suited for edge-deployed agricultural fire early-warning systems. Full article
19 pages, 4855 KB  
Article
Development of a Thermal Helipad for UAVs and Detection with Deep Learning
by Ersin Demiray, Mehmet Konar and Seda Arık Hatipoğlu
Drones 2026, 10(4), 266; https://doi.org/10.3390/drones10040266 - 7 Apr 2026
Abstract
For Unmanned Aerial Vehicles (UAVs), optical sensing for reliable landing and the detection of the landing area is a crucial element. In low-light conditions, at night, and in foggy weather, where optical sensing is not feasible, thermal imaging can be utilised. Although this [...] Read more.
For Unmanned Aerial Vehicles (UAVs), optical sensing for reliable landing and the detection of the landing area is a crucial element. In low-light conditions, at night, and in foggy weather, where optical sensing is not feasible, thermal imaging can be utilised. Although this situation has been widely researched, most UAV landing approaches rely on GNSS assistance or single-mode detection, which limits their robustness and scalability in real-world operations. This study proposes an actively heated thermal helicopter landing pad designed using electrically powered resistive heating elements and a high-emissivity surface coating. Furthermore, optical and thermal images collected during actual UAV flight experiments under daytime and night-time conditions were processed using image fusion techniques with AVGF, DWTF, GPF, LPF, MPF, and HWTF fusions, and their performance in deep learning models was compared. The obtained optical, thermal, and fused datasets are used to train and evaluate deep learning-based helicopter landing pad detection models based on the YOLOv8 architecture. Experimental results show that models trained with single-mode data exhibit limited cross-domain generalisation, while fusion-based learning significantly improves detection robustness in optical and thermal domains. Among the evaluated methods, LPF, MPF and HWTF provide the most consistent performance improvements. The findings indicate that electrically heated thermal helicopter landing pads, when combined with image fusion and deep learning-based detection, can increase the landing detectability of UAVs at night and in low-visibility conditions. This detection-focused approach contributes to UAV flight safety by enhancing the visibility of the landing area without relying on active infrared markers or additional navigation infrastructure. Full article
Show Figures

Graphical abstract

29 pages, 11454 KB  
Article
CASGNet: A Lightweight Content-Aware Spatial Gating Network for Cross-Regional Wheat Lodging Mapping from UAV Imagery
by Yueying Zhang, Zhuangzhi Nie, Chaowei Hu, Shouguan Xiao, Yuxi Wang, Shuqing Yang and Fanggang Wang
Electronics 2026, 15(7), 1530; https://doi.org/10.3390/electronics15071530 - 6 Apr 2026
Viewed by 208
Abstract
We investigate wheat lodging segmentation from UAV RGB imagery acquired over real production fields rather than controlled experimental sites. Besides pixel-level accuracy, our evaluation also emphasizes robustness under heterogeneous farmland conditions and deployment-oriented efficiency. We propose CASGNet, an edge-oriented segmentation network with a [...] Read more.
We investigate wheat lodging segmentation from UAV RGB imagery acquired over real production fields rather than controlled experimental sites. Besides pixel-level accuracy, our evaluation also emphasizes robustness under heterogeneous farmland conditions and deployment-oriented efficiency. We propose CASGNet, an edge-oriented segmentation network with a content-aware spatial gating mechanism that reweights intermediate features according to local structural variation. Instead of uniformly aggregating features, the module suppresses responses in homogeneous regions while preserving activation in structurally complex areas. In practice, this improves the continuity of irregular lodging shapes and reduces spurious responses in relatively homogeneous backgrounds. The dataset spans 46 farms across Jiaozuo, Jiyuan, and Luoyang, covering progressively fragmented farmland. Under a stricter mission-level data-isolation protocol, CASGNet achieves 94.4% mIoU and 90.38% IoU for the lodging class on the combined dataset. Under sequential regional adaptation, performance remains relatively stable in continuous parcels, and degradation is less severe than most compact baselines in highly fragmented landscapes. On Jetson Nano, CASGNet achieves 1.94 FPS embedded inference under the 5 W mode. Smaller networks achieve higher speed but show reduced structural continuity in complex scenes. The results indicate that CASGNet provides a favorable balance between structural fidelity and computational cost, while its robustness remains constrained by scene complexity. Full article
(This article belongs to the Collection Electronics for Agriculture)
Show Figures

Figure 1

21 pages, 5751 KB  
Article
A Hybrid VMD-Transformer-BiLSTM Framework with Cross-Attention Fusion for Aileron Fault Diagnosis in UAVs
by Yang Song, Weihang Zheng, Xiaoyu Zhang and Rong Guo
Sensors 2026, 26(7), 2256; https://doi.org/10.3390/s26072256 - 6 Apr 2026
Viewed by 178
Abstract
Aileron fault diagnosis in fixed-wing unmanned aerial vehicles (UAVs) faces significant challenges due to strong noise, multi-modal coupling, and limited fault samples. This paper presents a hybrid fault diagnosis framework that integrates variational mode decomposition (VMD) with a cross-attention-based feature fusion mechanism. First, [...] Read more.
Aileron fault diagnosis in fixed-wing unmanned aerial vehicles (UAVs) faces significant challenges due to strong noise, multi-modal coupling, and limited fault samples. This paper presents a hybrid fault diagnosis framework that integrates variational mode decomposition (VMD) with a cross-attention-based feature fusion mechanism. First, residual signals are generated from UAV kinematic models and decomposed into multi-scale intrinsic mode functions (IMFs) using VMD to extract multiscale frequency-localized features. An integrated framework is then constructed, where Transformer encoders capture the global features and bidirectional long short-term memory (BiLSTM) networks extract local temporal dynamics. To effectively combine these complementary features, a cross-attention fusion module is designed to focus on the discriminative time-frequency features. Furthermore, a hybrid pooling strategy integrating max pooling and attention pooling is introduced to enhance classification robustness. Experiments on the AirLab failure and anomaly (ALFA) dataset demonstrate that the proposed method achieves 95.12% accuracy with improved fault separability, outperforming VMD + BiLSTM (87.66%), VMD + Transformer (86.89%), Transformer + BiLSTM (84.83%), Transformer (72.24%), CNN + LSTM (94.05%), and HDMTL (94.86%). Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

24 pages, 766 KB  
Article
Systematic Evaluation of YOLOv8 Variants for UAV-Based Object Detection
by Chieh-Min Liu and Jyh-Ching Juang
Appl. Sci. 2026, 16(7), 3559; https://doi.org/10.3390/app16073559 - 6 Apr 2026
Viewed by 225
Abstract
Detecting small objects in drone imagery remains challenging because of extreme object scale variations, dense scenes, and limited pixel information. Although recent YOLOv8 variants provide multiple model scales and architectural options, systematic guidance on their practical use in UAV-based detection remains limited. Rather [...] Read more.
Detecting small objects in drone imagery remains challenging because of extreme object scale variations, dense scenes, and limited pixel information. Although recent YOLOv8 variants provide multiple model scales and architectural options, systematic guidance on their practical use in UAV-based detection remains limited. Rather than proposing novel network architectures, this study provides a quantitative cost–benefit analysis and empirical deployment guidelines by comprehensively evaluating the complete YOLOv8 family on the VisDrone dataset to assess the effects of the model capacity, input resolution, and architectural modifications on the small-object detection performance. The results showed that increasing the model capacity exhibited diminishing returns: YOLOv8l achieved the best overall accuracy (15.9% mAP50), while the larger YOLOv8x model exhibited a substantial performance degradation (7.32% mAP50) owing to training instability under data-constrained conditions. Scaling the input resolution from 640 to 1280 yielded a 25% improvement in detection performance, substantially exceeding the gains obtained through architectural modifications, such as adding a P2 detection layer (+6%). The optimal configuration (YOLOv8l @ 1280) achieved a 488% improvement compared to the YOLOv5 baseline. These findings demonstrate that, for UAV-based small-object detection, prioritizing an appropriate model capacity and input resolution is more effective than increasing the architectural complexity. Full article
Show Figures

Figure 1

32 pages, 43664 KB  
Article
MVFF: Multi-View Feature Fusion Network for Small UAV Detection
by Kunlin Zou, Haitao Zhao, Xingwei Yan, Wei Wang, Yan Zhang and Yaxiu Zhang
Drones 2026, 10(4), 264; https://doi.org/10.3390/drones10040264 - 4 Apr 2026
Viewed by 319
Abstract
With the widespread adoption of various types of Unmanned Aerial Vehicles (UAVs), their non-compliant operations pose a severe challenge to public safety, necessitating the urgent identification and detection of UAV targets. However, in complex backgrounds, UAV targets exhibit small-scale dimensions and low contrast, [...] Read more.
With the widespread adoption of various types of Unmanned Aerial Vehicles (UAVs), their non-compliant operations pose a severe challenge to public safety, necessitating the urgent identification and detection of UAV targets. However, in complex backgrounds, UAV targets exhibit small-scale dimensions and low contrast, coupled with extremely low signal-to-noise ratios. This forces conventional target detection methods to confront issues such as feature convergence, missed detections, and false alarms. To address these challenges, we propose a Multi-View Feature Fusion Network (MVFF) that achieves precise identification of small, low-contrast UAV targets by leveraging complementary multi-view information. First, we design a collaborative view alignment fusion module. This module employs a cross-map feature fusion attention mechanism to establish pixel-level mapping relationships and perform deep fusion, effectively resolving geometric distortion and semantic overlap caused by imaging angle differences. Furthermore, we introduce a view feature smoothing module that employs displacement operators to construct a lightweight long-range modeling mechanism. This overcomes the limitations of traditional convolutional local receptive fields, effectively eliminating ghosting artifacts and response discontinuities arising from multi-view fusion. Additionally, we developed a small object binary cross-entropy loss function. By incorporating scale-adaptive gain factors and confidence-aware weights, this function enhances the learning capability of edge features in small objects, significantly reducing prediction uncertainty caused by background noise. Comparative experiments conducted on a multi-perspective UAV dataset demonstrate that our approach consistently outperforms existing state-of-the-art methods across multiple performance metrics. Specifically, it achieves a Structure-measure of 91.50% and an F-measure of 85.14%, validating the effectiveness and superiority of the proposed method. Full article
Show Figures

Figure 1

24 pages, 4159 KB  
Article
A UAV–Satellite Hybrid Pipeline for Wildfire Detection and Dynamic Perimeter Prediction
by Hossein Keshmiri and Khan A. Wahid
Drones 2026, 10(4), 263; https://doi.org/10.3390/drones10040263 - 4 Apr 2026
Viewed by 296
Abstract
Effective wildfire management demands seamless integration of real-time detection and long-term spread forecasting. This paper proposes a novel power-efficient UAV–satellite hybrid pipeline that synergizes the agility of UAVs with the scale of satellite intelligence. The system begins with a dashboard-guided, multi-UAV detection module [...] Read more.
Effective wildfire management demands seamless integration of real-time detection and long-term spread forecasting. This paper proposes a novel power-efficient UAV–satellite hybrid pipeline that synergizes the agility of UAVs with the scale of satellite intelligence. The system begins with a dashboard-guided, multi-UAV detection module that scores fire likelihood from historical satellite data and enables scalable, energy-efficient deployment with low-latency onboard processing. This aerial component ensures persistent surveillance and reliable ignition detection, supported by a Dual LoRa (Long Range) communication scheme for robust and low-power connectivity. It achieves an F1-score of 97.4% while minimizing power consumption to extend operational flight times. Following detection, the pipeline transitions to a dynamic perimeter-prediction phase utilizing a custom Canadian boreal dataset. We employ a Squeeze-and-Excitation Residual U-Net (SE-ResUNet) to model spatiotemporal fire propagation based on static terrain and dynamic environmental features. The model was validated using a dynamic simulation framework that evaluates temporal consistency and convergence behavior against final cumulative burned-area masks, effectively addressing the absence of daily ground truth. Under these conditions, the model achieves a recall of 84% and an AUC of 0.97, demonstrating a strong capability to delineate active fire fronts. By coupling dashboard-driven UAV sensing with satellite-based predictive modeling, this work establishes a modular, foundational framework to support data-scarce forecasting in modern wildfire management. Full article
(This article belongs to the Special Issue UAVs and UGVs Robotics for Emergency Response in a Changing Climate)
Show Figures

Figure 1

22 pages, 2592 KB  
Article
Predicting Rice Quality in Indica Rice Using Multidimensional Data and Machine Learning Strategies
by Xiang Zhang, Yongqiang Liu, Junming Yu, Ni Cao, Wei Zhou, Jiaming Wu, Rumeng Zhao, Shaoqing Tang, Song Chen, Ying Chen, Fengli Zhao, Jiwai He and Gaoneng Shao
Agriculture 2026, 16(7), 807; https://doi.org/10.3390/agriculture16070807 - 4 Apr 2026
Viewed by 244
Abstract
Integrating agricultural remote sensing and phenomics for full-growth-period rice quality prediction is vital for early non-destructive screening and breeding; however, studies integrating genomic and multi-source phenotypic data across multiple environments remain limited. This study addressed this gap by integrating genomic SNP data, UAV-based [...] Read more.
Integrating agricultural remote sensing and phenomics for full-growth-period rice quality prediction is vital for early non-destructive screening and breeding; however, studies integrating genomic and multi-source phenotypic data across multiple environments remain limited. This study addressed this gap by integrating genomic SNP data, UAV-based spectral data, and individual multidimensional phenotypic data of 61 indica rice varieties (field and greenhouse environments). As a proof-of-concept study, feature selection methods (LASSO, MI, RFE, SPA) were used to mitigate overfitting and the “p >> n” problem, with further validation needed in larger populations. The results showed that amylose content is genetically dominated, protein content is genetically determined and influenced by gene-environment interactions, and chalkiness traits are determined by three combined factors. For amylose content, SNP data under the Random Forest model at the population level (phenomics data from field UAV remote sensing of variety populations) achieved optimal performance (R2 = 0.92; MAE = 1.1; RMSE = 1.5), while the Stacking Ensemble method enhanced accuracy at the individual level (phenomics data from greenhouse single-plant phenotyping per variety). Chalky grain rate and chalkiness degree showed SNP-comparable prediction accuracy, with Stacking significantly improving performance at the population level (R2 = 0.89 and 0.85, respectively). Protein content prediction remained relatively low (optimal R2 = 0.56) due to strong environmental sensitivity and complex interactions. This framework extends traditional single-environment/single-data-source approaches, providing an effective strategy for early, high-throughput, non-destructive rice quality screening. Further validation with larger datasets, more growing seasons, or independent populations is required for reliable application in breeding-related practices. Full article
(This article belongs to the Section Agricultural Product Quality and Safety)
Show Figures

Figure 1

22 pages, 12678 KB  
Article
A UAV Localization Method Based on Unique Semantic Instances
by Yineng Li, Qinghua Zeng, Ziqi Jin, Junjie Wu, Rongbing Li and Junwei Wan
Remote Sens. 2026, 18(7), 1084; https://doi.org/10.3390/rs18071084 - 3 Apr 2026
Viewed by 185
Abstract
The unmanned aerial vehicle (UAV) localization method based on global features is a fast and efficient approach for satellite-denied environments. Such methods typically extract global features from aerial images and retrieve matches from a constructed feature database to locate UAVs. However, constructing the [...] Read more.
The unmanned aerial vehicle (UAV) localization method based on global features is a fast and efficient approach for satellite-denied environments. Such methods typically extract global features from aerial images and retrieve matches from a constructed feature database to locate UAVs. However, constructing the feature database requires traversing the entire map, leading to storage redundancy. Moreover, the reference images in the database often have fixed fields of view and orientations, making it difficult to adapt to the changes in aerial images caused by the altitude and attitude changes of the UAV. To address these challenges, this paper explores the uniqueness of semantic instances within the mission region and proposes a UAV localization method based on unique semantic instances. The proposed method first extracts the labels of unique semantic instances from aerial images. These labels are then used to retrieve and match the corresponding feature vectors stored in the database. The location is determined based on the centroid positions of the matched unique semantic instances stored in the feature vectors. Experimental results on both simulation and flight datasets show that the proposed method achieves a localization success rate exceeding 95% in the mission region and remains robust to changes in the attitude and field of view of aerial images. The proposed method requires storing only the categories and locations of the instances, significantly reducing data storage requirements. Full article
Show Figures

Figure 1

33 pages, 2275 KB  
Article
SymbioMamba: An Efficient Dual-Stream State-Space Framework for Real-Time Maize Disease and Yield Analysis on UAV Platforms
by Zihuan Wang, Yuru Wang, Bocheng Zhou, Xu Yan, Peijiang Guo, Hanyu Yang and Yihong Song
Agriculture 2026, 16(7), 801; https://doi.org/10.3390/agriculture16070801 - 3 Apr 2026
Viewed by 140
Abstract
In UAV (unmanned aerial vehicle)-enabled precision agriculture, achieving high-accuracy disease diagnosis and yield estimation simultaneously on resource-constrained edge devices remains a significant challenge. Existing solutions are commonly hindered by conflicts in visual feature scales, the absence of explicit agronomic causal logic, and the [...] Read more.
In UAV (unmanned aerial vehicle)-enabled precision agriculture, achieving high-accuracy disease diagnosis and yield estimation simultaneously on resource-constrained edge devices remains a significant challenge. Existing solutions are commonly hindered by conflicts in visual feature scales, the absence of explicit agronomic causal logic, and the trade-off between lightweight design and global modeling capability. To address these challenges, a heterogeneous dual-stream state-space framework termed SymbioMamba is proposed. The proposed framework incorporates three key innovations: first, a heterogeneous dual-stream encoder is constructed, in which a micro-texture stream captures high-frequency disease details while a macro-context-scan stream models field-scale biomass continuity; second, a pathology–biomass collaborative interaction (PBCI) module is designed to explicitly inject the biological prior—disease stress leading to yield reduction—into the feature space. Third, a topology-aligning cross-architecture distillation (TACAD) paradigm is introduced to transfer global knowledge from a heavyweight teacher to a lightweight student. Experimental results from a maize UAV dataset comprising 12,074 annotated image patches demonstrate that SymbioMamba achieves 89.4% mAP@0.5 and an R2 of 0.915. Compared to the industry-standard YOLOv11, the framework improves mAP@0.5:0.95 by 2.4% while reducing the parameter count to 6.2 M—a 50% decrease relative to monolithic state-space baselines. Furthermore, yield prediction error is significantly reduced to an RMSE of 485.6 kg/ha. With a compact model size of 6.2 M parameters and 2.4 G FLOPs, SymbioMamba attains an inference speed of 38.2 FPS on the NVIDIA Jetson AGX Orin platform, providing a high-performance, real-time solution for intelligent agricultural phenotypic analysis. Full article
(This article belongs to the Special Issue Smart Sensor-Based Systems for Crop Monitoring)
Show Figures

Figure 1

26 pages, 12902 KB  
Article
Soft Threshold Denoising-Based Environmental Adaptive UAV Signal Modulation Recognition for Small-Sample Scenarios
by Fang Jin, Yang Shao, Yunhong He, Zhihao Ye, Fangmin He, Zhipeng Lin and Han Xiao
Drones 2026, 10(4), 257; https://doi.org/10.3390/drones10040257 - 3 Apr 2026
Viewed by 200
Abstract
As a key technology for wireless signal identification, modulation recognition plays an important role in the fields of unmanned aerial vehicle (UAV) communications, low-altitude spectrum management, etc. However, the accuracy of modulation recognition often cannot be guaranteed in scenarios with serious noise interference [...] Read more.
As a key technology for wireless signal identification, modulation recognition plays an important role in the fields of unmanned aerial vehicle (UAV) communications, low-altitude spectrum management, etc. However, the accuracy of modulation recognition often cannot be guaranteed in scenarios with serious noise interference when a few samples are available. In this paper, we propose an intelligent modulation recognition method for UAV signals based on small-sample augmentation and soft threshold denoising. We first propose a new dual-driven dataset expansion method by combining the UAV air–ground channel propagation model with the received data samples. Then, we construct a background learning-based long short-term memory (BL-LSTM) model to extract the environmental background features embedded in the UAV signal, including Line-of-Sight (LoS) state, multi-scale fading parameters and Doppler shift characteristics. We integrate environmental background information into the data training model and optimize the authenticity of data distribution. As a result, the model adaptability can be enhanced. Finally, we construct a deep residual shrinkage network based on the soft threshold function (STF-DRSN). By leveraging the capability of the soft threshold that resists noise interference, we integrate it into each residual block of the deep residual shrinkage network. Simulation results show that compared with the state of the art, our method can improve the modulation recognition accuracy of UAV signals in small-sample scenarios. Full article
Show Figures

Figure 1

23 pages, 4047 KB  
Article
UAV-Based Estimation of Tea Leaf Area Index in Mountainous Terrain: Integrating Topographic Correction and Interpretable Machine Learning
by Na Lin, Jian Zhao, Huxiang Shao, Miaomiao Wang and Hong Chen
Sensors 2026, 26(7), 2218; https://doi.org/10.3390/s26072218 - 3 Apr 2026
Viewed by 216
Abstract
Leaf Area Index (LAI) is a fundamental parameter for characterizing the growth of tea (Camellia sinensis L.). However, in rugged mountainous regions, the combined effects of topographic relief and canopy structural heterogeneity severely constrain the accuracy of UAV-based multispectral LAI retrieval. This [...] Read more.
Leaf Area Index (LAI) is a fundamental parameter for characterizing the growth of tea (Camellia sinensis L.). However, in rugged mountainous regions, the combined effects of topographic relief and canopy structural heterogeneity severely constrain the accuracy of UAV-based multispectral LAI retrieval. This study develops an integrated framework combining topographic correction with interpretable machine learning to improve LAI estimation. We utilized a UAV multispectral dataset collected during the peak growing season from a typical tea-growing region in Fujian Province, China (altitude range: 58–186 m), comprising a total of 90 samples. Three topographic correction methods, including Sun–Canopy–Sensor (SCS), SCS with C correction (SCS+C), and Minnaert+SCS, were evaluated in combination with Linear Regression (LR), Decision Tree (DT), Random Forest (RF), and Extreme Gradient Boosting (XGBoost) models. Results indicated that the SCS+C algorithm outperformed other methods by effectively accounting for direct and diffuse radiation components, thereby reducing topographic dependence while maintaining radiometric consistency across heterogeneous surfaces. The XGBoost model combined with SCS+C correction achieved the highest performance (R2 = 0.8930, RMSE = 0.6676, nRMSE = 7.93%, MAE = 0.4936, Bias = −0.0836). SHapley Additive exPlanations (SHAP) analysis revealed a structure-dominated retrieval mechanism, in which red-band textural features (Correlation_R) exhibited higher importance than conventional vegetation indices. Compared with previous studies that primarily focus on either topographic correction or model development, this study provides quantitative insights into the underlying retrieval mechanisms. This framework improves the precision of tea LAI retrieval in complex terrains and provides a robust methodological basis for digital management in mountainous agriculture. Full article
(This article belongs to the Special Issue AI UAV-Based Systems for Agricultural Monitoring)
Show Figures

Figure 1

14 pages, 1601 KB  
Article
Real-Time UAV-Based Oil Pipeline and Visual Anomaly Detection Using YOLOv26n: A Dataset and Edge-Deployment Study
by Hatem Keshk and Ayman Abdallah
Drones 2026, 10(4), 255; https://doi.org/10.3390/drones10040255 - 3 Apr 2026
Viewed by 249
Abstract
Ensuring the structural integrity and operational safety of oil and gas pipelines is a critical challenge due to their extensive geographical coverage and exposure to environmental and anthropogenic risks. Traditional inspection approaches including ground patrols and manned aerial surveys are labor-intensive, costly, and [...] Read more.
Ensuring the structural integrity and operational safety of oil and gas pipelines is a critical challenge due to their extensive geographical coverage and exposure to environmental and anthropogenic risks. Traditional inspection approaches including ground patrols and manned aerial surveys are labor-intensive, costly, and often lack real-time responsiveness. While unmanned aerial vehicles (UAVs) enable flexible and high-resolution monitoring, their practical deployment requires lightweight, robust detection models capable of real-time inference on embedded edge hardware under heterogeneous environmental conditions. This paper presents an end-to-end, edge-deployable UAV inspection framework for simultaneous detection of above-ground pipelines and visually observable anomaly/leak indicators using the official Ultralytics YOLOv26n object detector. A curated dataset of 6127 UAV images acquired across desert, semi-urban, and industrial environments was annotated with two classes (Pipeline and Anomaly/Leak) and partitioned into training 87.5%, validation 8.3%, and testing 4.2% subsets. The detector was fine-tuned from COCO-pretrained weights for 300 epochs at 600 × 600 resolution and evaluated using COCO-style metrics. On the held-out test set, the proposed model achieved 92.4% mAP@0.5 and 75.0% mAP@0.5:0.95, with 89.7% precision, 90.2% recall, and 89.9% F1-score at the selected operating threshold. Optimized TensorRT deployment on an NVIDIA Jetson Xavier NX sustained real-time inference at 18 FPS, demonstrating suitability for onboard UAV processing. Rather than proposing a new detector architecture, the study contributes a domain-specific annotated UAV dataset, deployment-oriented benchmarking, and an end-to-end edge inference workflow for corridor-scale monitoring. The proposed framework can help reduce environmental contamination risk and improve personnel safety during pipeline inspection. Full article
(This article belongs to the Special Issue Autonomy Challenges in Unmanned Aviation)
Show Figures

Figure 1

19 pages, 8523 KB  
Article
DAMFusion: Multi-Spectral Image Segmentation via Competitive Query and Boundary Region Attention
by Miao Yu, Xing Lu, Ziyao Yang, Daoxing Gao and Guoqiang Zhong
Remote Sens. 2026, 18(7), 1064; https://doi.org/10.3390/rs18071064 - 2 Apr 2026
Viewed by 244
Abstract
To address the challenges of modal differences in multimodal farmland images and insufficient segmentation accuracy for small targets, this paper proposes a multi-source image fusion branch (DAMFusion) based on modal competitive selection. The branch dynamically selects infrared and visible light features through the [...] Read more.
To address the challenges of modal differences in multimodal farmland images and insufficient segmentation accuracy for small targets, this paper proposes a multi-source image fusion branch (DAMFusion) based on modal competitive selection. The branch dynamically selects infrared and visible light features through the Competitive Query Module (CQM) using Top-K screening, combined with IOU-aware loss optimization to avoid cross-modal interference. The multimodal fusion module (MMFormer) employs cross-modal attention and symmetric mechanisms, enhancing single-modal features through a self-enhancement module and unifying multimodal distributions via linear projection. The Boundary Region Attention Multi-level Fusion Module (BRM) extracts boundary information through feature differencing, strengthens it with spatial attention, and fuses it with shallow features to achieve cross-layer detail recovery. Through the collaborative design of dynamic modal feature selection, cross-modal distribution unification, and boundary region enhancement, DAMFusion effectively solves the problems of multimodal differences and small target segmentation in multispectral images, providing precise feature representation for fine farmland segmentation. Experiments on the OUC-UAV-MSEG dataset show that DAMFusion achieves 93.25% OA, 91.71% F1, and 89.70% mIoU, demonstrating clear advantages over representative comparison methods. In addition, ablation results verify the effectiveness of the proposed modules, where CQM improves OA from 91.00% to 93.25%, confirming the importance of discriminative modality selection before fusion. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

31 pages, 28128 KB  
Article
HMF-DEIM: High-Fidelity Multi-Domain Fusion Transformer for UAV Small Object Detection
by Lan Ma, Yun Luo and Jiajun Xu
Sensors 2026, 26(7), 2187; https://doi.org/10.3390/s26072187 - 1 Apr 2026
Viewed by 304
Abstract
Unmanned aerial vehicle (UAV) small object detection faces critical challenges including irreversible geometric detail loss during multi-level downsampling, cross-scale feature distortion from interpolation blur and aliasing, and limited long-range dependency modeling due to constrained receptive fields. To address these limitations, we propose HMF-DEIM [...] Read more.
Unmanned aerial vehicle (UAV) small object detection faces critical challenges including irreversible geometric detail loss during multi-level downsampling, cross-scale feature distortion from interpolation blur and aliasing, and limited long-range dependency modeling due to constrained receptive fields. To address these limitations, we propose HMF-DEIM (High-Fidelity Multi-Domain Fusion Transformer for UAV Small Object Detection), an end-to-end architecture tailored for UAV small object detection. First, we design a lightweight hierarchical differentiation backbone that removes redundant deepest-layer features (P5) to prevent tiny object information loss, employing Multi-Domain Feature Blending (MDFB) in shallow layers for geometric detail preservation and a Hierarchical Attention-guided Feature Modulation Block (HAFMB) in deep layers for global semantic modeling. Second, we develop a full-chain high-fidelity feature transformation framework comprising Channel-Adaptive Shift Upsampling (CASU) for interpolation-free resolution recovery, Multi-scale Context Alignment Fusion (MCAF) for bridging deep–shallow semantic gaps via bidirectional gating, and Diversified Residual Frequency-aware Downsampling (DRFD) for aliasing suppression through a three-branch parallel architecture. Finally, we devise the FocusFeature module that aligns multi-scale features to a unified scale and employs parallel multi-scale large-kernel depthwise convolutions to capture cross-scale long-range dependencies, generating dual-scale (P3/P4) features balancing details and semantics. Experiments demonstrate that HMF-DEIM outperforms DEIM on VisDrone2019 test by 0.405 mAP50 (+2.1%) and 0.235 mAP50–95 (+1.6%), with a remarkable 21.3% relative improvement in APs for tiny objects, while maintaining real-time inference (465 FPS with TensorRT FP16) on an NVIDIA A100 GPU with only 11.87M parameters and 34.1 GFLOPs. Further validation on AI-TOD v2 and DOTA v1.5 datasets confirms robust generalization across diverse aerial scenarios, making it a practical solution for resource-constrained UAV applications. Full article
(This article belongs to the Special Issue Communications and Networking Based on Artificial Intelligence)
Show Figures

Figure 1

Back to TopTop