Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (79)

Search Parameters:
Keywords = UAV geolocalization

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 19359 KB  
Article
VLM-Guided and Spatially Consistent Cross-View Matching and Localization of UGVs in UAV-Stitched Map
by Yusheng Yang, Xinxu Ma, Ziluan Jiang, Pengfei Sun, Xun Zhao, Yangmin Xie and Wei Qian
Drones 2025, 9(12), 873; https://doi.org/10.3390/drones9120873 - 17 Dec 2025
Abstract
In Global Navigation Satellite System (GNSS)-denied urban environments, unmanned ground vehicles (UGVs) face significant difficulties in maintaining reliable localization due to occlusion and structural complexity. Unmanned aerial vehicles (UAVs), with their global perspective, provide complementary information for cross-view matching and localization of UGVs. [...] Read more.
In Global Navigation Satellite System (GNSS)-denied urban environments, unmanned ground vehicles (UGVs) face significant difficulties in maintaining reliable localization due to occlusion and structural complexity. Unmanned aerial vehicles (UAVs), with their global perspective, provide complementary information for cross-view matching and localization of UGVs. However, robust cross-view matching and localization are hindered by geometric distortions, semantic inconsistencies, and the lack of stable spatial anchors, limiting the effectiveness of conventional methods. To overcome these challenges, we proposed a cross-view matching and localization (CVML) framework that contains two components. The first component is the Vision-Language Model (VLM)-guided and spatially consistent cross-view matching network (VSCM-Net), which integrates two novel attention modules. One is the VLM-guided positional correction module that leverages semantic cues to refine the projected UGV image within the UAV map, and the other is the shape-aware attention module that enforces topological consistency across ground and aerial views. The second component is a ground-to-aerial mapping module that projects cross-view correspondences from the UGV image onto the UAV-stitched map, thereby localizing the capture position of the UGV image and enabling accurate trajectory-level localization and navigation. Extensive experiments on public and self-collected datasets demonstrate that the proposed method achieves superior accuracy, robustness, and real-world applicability compared with state-of-the-art methods in both cross-view image matching and localization. Full article
(This article belongs to the Special Issue UAV Detection, Classification, and Tracking)
Show Figures

Figure 1

31 pages, 37241 KB  
Article
DEM-Based UAV Geolocation of Thermal Hotspots on Complex Terrain
by Lucile Rossi, Frédéric Morandini, Antoine Burglin, Jean Bertrand, Clément Wandon, Aurélien Tollard and Antoine Pieri
Remote Sens. 2025, 17(23), 3911; https://doi.org/10.3390/rs17233911 - 2 Dec 2025
Viewed by 406
Abstract
Reliable geolocation of thermal hotspots, such as smoldering embers that can reignite after vegetation fire suppression, deep-seated peat fires, or underground coal seam fires, is critical to prevent fire resurgence, limit prolonged greenhouse gas emissions, and mitigate environmental and health impacts. This study [...] Read more.
Reliable geolocation of thermal hotspots, such as smoldering embers that can reignite after vegetation fire suppression, deep-seated peat fires, or underground coal seam fires, is critical to prevent fire resurgence, limit prolonged greenhouse gas emissions, and mitigate environmental and health impacts. This study develops and tests an algorithm to estimate the GPS positions of thermal hotspots detected in infrared images acquired by an unmanned aerial vehicle (UAV), designed to operate over flat and mountainous terrain. Its originality lies in a reformulated Bresenham traversal of the digital elevation model (DEM), combined with a lightweight, ray-tracing-inspired strategy that efficiently detects the intersection of the optical ray with the terrain by approximating the ray altitude at the cell level. UAV flight experiments in complex terrain were conducted, with thermal image acquisitions performed at 60 m and 120 m above ground level and simulated hotspots generated using controlled heat sources. The tests were carried out with two thermal cameras: a Zenmuse H20T mounted on a Matrice 300 UAV flown both with and without Real-Time Kinematic (RTK) positioning, and a Matrice 30T UAV without RTK. The implementation supports both real-time and post-processed operation modes. The results demonstrated robust and reliable geolocation performance, with mean positional errors consistently below 4.2 m for all the terrain configurations tested. A successful real-time operation in the test confirmed the suitability of the algorithm for time-critical intervention scenarios. Since July 2024, the post-processed version of the method has been in operational use by the Corsica fire services. Full article
Show Figures

Graphical abstract

29 pages, 6001 KB  
Article
Vision-Based Geolocation of Moving Ground Targets Using Kalman Filtering with a Gimbal Camera on Board a UAV
by Jaemin Kim, Youngrun Kim, SuHyeon Kim, Hyeongjun Cho and Dongwon Jung
Aerospace 2025, 12(12), 1065; https://doi.org/10.3390/aerospace12121065 - 30 Nov 2025
Viewed by 348
Abstract
Unmanned aerial vehicles (UAVs) are vital for surveillance missions requiring the geolocation of moving ground targets, yet small, resource-constrained platforms often lack integrated, robust systems that can handle disturbances such as wind, occlusions, and noise. This paper presents an integrated, end-to-end vision-based geolocation [...] Read more.
Unmanned aerial vehicles (UAVs) are vital for surveillance missions requiring the geolocation of moving ground targets, yet small, resource-constrained platforms often lack integrated, robust systems that can handle disturbances such as wind, occlusions, and noise. This paper presents an integrated, end-to-end vision-based geolocation pipeline specifically designed for embedded deployment on resource-constrained UAVs with gimbal cameras. Starting from a rough initial position estimate, pan/tilt angles are computed to orient the gimbal, and then a visual tracking module combining object detection (via Tiny-YOLO) and feedback control (using CSRT) centers the target in the frame. The target’s absolute position is derived from UAV inertial data and gimbal angles. To mitigate noisy or unavailable direct geolocation due to disturbances or visual lock loss, Kalman filtering is integrated with a unicycle-based motion model. Both an extended Kalman filter (EKF) and unscented Kalman filter (UKF) are evaluated and tuned in high-fidelity simulations, with the UKF demonstrating superior performance by reducing the 2D position RMSE by 33% compared to the EKF in occlusion scenarios. The system is implemented on embedded hardware and validated through real flight tests, establishing the operational capability of vision-based surveillance on small UAV platforms. Full article
Show Figures

Figure 1

34 pages, 9458 KB  
Article
Assessing Wildlife Impact on Forest Regeneration Through Drone-Based Thermal Imaging
by Claudia C. Jordan-Fragstein, Michael G. Müller, Niklas Bielefeld, Richard Georgi and Robert Friedrich
Forests 2025, 16(12), 1787; https://doi.org/10.3390/f16121787 - 28 Nov 2025
Viewed by 477
Abstract
Assessing the extent and magnitude of wildlife impact on forest regeneration (e.g., % browsed seedlings or reduction in regeneration density) remains a central challenge. This study explores the potential of unmanned aircraft systems (UAS) to quantify wildlife impact through the integration of drone-based [...] Read more.
Assessing the extent and magnitude of wildlife impact on forest regeneration (e.g., % browsed seedlings or reduction in regeneration density) remains a central challenge. This study explores the potential of unmanned aircraft systems (UAS) to quantify wildlife impact through the integration of drone-based thermal surveys and vegetation assessments. Specifically, it evaluates whether UAS-derived wildlife density estimates can be linked to browsing intensity and regeneration structure, thereby enabling an indirect assessment of silviculturally relevant forest dynamics. By combining remotely sensed wildlife data with field-based vegetation inventories, the study aims to identify measurable relationships between structural forest characteristics and browsing effects. This approach contributes to the development of spatially efficient, objective, and reproducible monitoring methods at the forest–wildlife interface. Ultimately, the study provides a novel framework for integrating modern remote sensing technologies into wildlife–ecological monitoring and for improving adaptive, evidence-based management in forest ecosystems increasingly affected by high ungulate densities and climate-related stressors. Two silviculturally contrasting study areas were selected: a broadleaf-dominated mixed forest in Hesse, where high ungulate densities were expected, and a pine-dominated site in Brandenburg, anticipated to experience lower browsing pressure. Thermal surveys were conducted using a DJI Matrice 30T drone equipped with a high-resolution infrared camera to detect and geolocate wildlife. In parallel, browsing impact was assessed using a modified circular transect method (“Neuzeller method”). Regeneration was recorded by tree species, height class, and browsing intensity. Statistical analyses and GIS-based spatial visualizations were used to examine the relationship between estimated ungulate densities and browsing levels. Results revealed clear differences in wildlife abundance and browsing intensity between the two sites. In the Heppenheim forest, roe deer densities exceeded 40 individuals per 100 ha, correlating with high browsing pressure—particularly on ecologically and silviculturally valuable species such as sycamore maple and sessile oak. In contrast, the Rochauer Heide exhibited lower densities and a comparatively moderate browsing impact, although certain tree species still showed signs of selective pressure. This study demonstrates that drone-based wildlife monitoring offers an innovative, non-invasive means to indirectly evaluate forest structural conditions in regeneration layers. The findings highlight the relevance of UAV-supported methods for evidence-based wildlife management and the adaptive planning of silvicultural measures. The method enhances transparency and spatial resolution in forest–wildlife management and supports evidence-based decision-making in times of ecological and climatic change. Full article
(This article belongs to the Section Forest Ecology and Management)
Show Figures

Graphical abstract

21 pages, 9487 KB  
Article
Low-Cost Real-Time Remote Sensing and Geolocation of Moving Targets via Monocular Bearing-Only Micro UAVs
by Peng Sun, Shiji Tong, Kaiyu Qin, Zhenbing Luo, Boxian Lin and Mengji Shi
Remote Sens. 2025, 17(23), 3836; https://doi.org/10.3390/rs17233836 - 27 Nov 2025
Viewed by 320
Abstract
Low-cost and real-time remote sensing of moving targets is increasingly required in civilian applications. Micro unmanned aerial vehicles (UAVs) provide a promising platform for such missions because of their small size and flexible deployment, but they are constrained by payload capacity and energy [...] Read more.
Low-cost and real-time remote sensing of moving targets is increasingly required in civilian applications. Micro unmanned aerial vehicles (UAVs) provide a promising platform for such missions because of their small size and flexible deployment, but they are constrained by payload capacity and energy budget. Consequently, they typically carry lightweight monocular cameras only. These cameras cannot directly measure distance and suffer from scale ambiguity, which makes accurate geolocation difficult. This paper tackles geolocation and short-term trajectory prediction of moving targets over uneven terrain using bearing-only measurements from a monocular camera. We present a two-stage estimation framework in which a pseudo-linear Kalman filter (PLKF) provides real-time state estimates, while a sliding-window nonlinear least-squares (NLS) back end refines them. Future target positions are obtained by extrapolating the estimated trajectory. To improve localization accuracy, we analyze the relationship between the UAV path and the Cramér–Rao lower bound (CRLB) using the Fisher Information Matrix (FIM) and derive an observability-enhanced trajectory planning method. Real-flight experiments validate the framework, showing that accurate geolocation can be achieved in real time using only low-cost monocular bearing measurements. Full article
Show Figures

Figure 1

24 pages, 248126 KB  
Article
Image Matching for UAV Geolocation: Classical and Deep Learning Approaches
by Fatih Baykal, Mehmet İrfan Gedik, Constantino Carlos Reyes-Aldasoro and Cefa Karabağ
J. Imaging 2025, 11(11), 409; https://doi.org/10.3390/jimaging11110409 - 12 Nov 2025
Viewed by 838
Abstract
Today, unmanned aerial vehicles (UAVs) are heavily dependent on Global Navigation Satellite Systems (GNSSs) for positioning and navigation. However, GNSS signals are vulnerable to jamming and spoofing attacks. This poses serious security risks, especially for military operations and critical civilian missions. In order [...] Read more.
Today, unmanned aerial vehicles (UAVs) are heavily dependent on Global Navigation Satellite Systems (GNSSs) for positioning and navigation. However, GNSS signals are vulnerable to jamming and spoofing attacks. This poses serious security risks, especially for military operations and critical civilian missions. In order to solve this problem, an image-based geolocation system has been developed that eliminates GNSS dependency. The proposed system estimates the geographical location of the UAV by matching the aerial images taken by the UAV with previously georeferenced high-resolution satellite images. For this purpose, common visual features were determined between satellite and UAV images and matching operations were carried out using methods based on the homography matrix. Thanks to image processing, a significant relationship has been established between the area where the UAV is located and the geographical coordinates, and reliable positioning is ensured even in cases where GNSS signals cannot be used. Within the scope of the study, traditional methods such as SIFT, AKAZE, and Multiple Template Matching were compared with learning-based methods including SuperPoint, SuperGlue, and LoFTR. The results showed that deep learning-based approaches can make successful matches, especially at high altitudes. Full article
Show Figures

Figure 1

19 pages, 2598 KB  
Article
DOCB: A Dynamic Online Cross-Batch Hard Exemplar Recall for Cross-View Geo-Localization
by Wenchao Fan, Xuetao Tian, Long Huang, Xiuwei Zhang and Fang Wang
ISPRS Int. J. Geo-Inf. 2025, 14(11), 418; https://doi.org/10.3390/ijgi14110418 - 26 Oct 2025
Viewed by 504
Abstract
Image-based geo-localization is a challenging task that aims to determine the geographic location of a ground-level query image captured by an Unmanned Ground Vehicle (UGV) by matching it to geo-tagged nadir-view (top-down) images from an Unmanned Aerial Vehicle (UAV) stored in a reference [...] Read more.
Image-based geo-localization is a challenging task that aims to determine the geographic location of a ground-level query image captured by an Unmanned Ground Vehicle (UGV) by matching it to geo-tagged nadir-view (top-down) images from an Unmanned Aerial Vehicle (UAV) stored in a reference database. The challenge comes from the perspective inconsistency between matched objects. In this work, we propose a novel metric learning scheme for hard exemplar mining to improve the performance of cross-view geo-localization. Specifically, we introduce a Dynamic Online Cross-Batch (DOCB) hard exemplar mining scheme that solves the problem of the lack of hard exemplars in mini-batches in the middle and late stages of training, which leads to training stagnation. It mines cross-batch hard negative exemplars according to the current network state and reloads them into the network to make the gradient of negative exemplars participating in back-propagation. Since the feature representation of cross-batch negative examples adapts to the current network state, the triplet loss calculation becomes more accurate. Compared with methods only considering the gradient of anchors and positives, adding the gradient of negative exemplars helps us to obtain the correct gradient direction. Therefore, our DOCB scheme can better guide the network to learn valuable metric information. Moreover, we design a simple Siamese-like network called multi-scale feature aggregation (MSFA), which can generate multi-scale feature aggregation by learning and fusing multiple local spatial embeddings. The experimental results demonstrate that our DOCB scheme and MSFA network achieve an accuracy of 95.78% on the CVUSA dataset and 86.34% on the CVACT_val dataset, which outperforms those of other existing methods in the field. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

17 pages, 4493 KB  
Article
VimGeo: An Efficient Visual Model for Cross-View Geo-Localization
by Kaiqian Yang, Yujin Zhang, Li Wang, A. A. M. Muzahid, Ferdous Sohel, Fei Wu and Qiong Wu
Electronics 2025, 14(19), 3906; https://doi.org/10.3390/electronics14193906 - 30 Sep 2025
Viewed by 645
Abstract
Cross-view geo-localization is a challenging task due to the significant changes in the appearance of target scenes from variable perspectives. Most existing methods primarily adopt Transformers or ConvNeXt as backbone models but often face high computational costs and accuracy degradation in complex scenarios. [...] Read more.
Cross-view geo-localization is a challenging task due to the significant changes in the appearance of target scenes from variable perspectives. Most existing methods primarily adopt Transformers or ConvNeXt as backbone models but often face high computational costs and accuracy degradation in complex scenarios. Therefore, this paper proposes a visual Mamba framework based on the state-space model (SSM) for cross-view geo-localization. Compared with the existing methods, Vision Mamba is more efficient in modeling and memory usage and achieves more efficient cross-view matching by combining the twin architecture of shared weights with multiple mixed losses. Additionally, this paper introduces Dice Loss to handle scale differences and imbalance issues in cross-view images. Extensive experiments on the public cross-view dataset University-1652 demonstrate that Vision Mamba not only achieves excellent performance in UAV target localization tasks but also attains the highest efficiency with lower memory consumption. This work provides a novel solution for cross-view geo-localization tasks and shows great potential to become the backbone model for the next generation of cross-view geo-localization. Full article
Show Figures

Figure 1

25 pages, 29114 KB  
Article
Towards UAV Localization in GNSS-Denied Environments: The SatLoc Dataset and a Hierarchical Adaptive Fusion Framework
by Xiang Zhou, Xiangkai Zhang, Xu Yang, Jiannan Zhao, Zhiyong Liu and Feng Shuang
Remote Sens. 2025, 17(17), 3048; https://doi.org/10.3390/rs17173048 - 2 Sep 2025
Viewed by 2466
Abstract
Precise and robust localization for micro Unmanned Aerial Vehicles (UAVs) in GNSS-denied environments is hindered by the lack of diverse datasets and the limited real-world performance of existing visual matching methods. To address these gaps, we introduce two contributions: (1) the SatLoc dataset, [...] Read more.
Precise and robust localization for micro Unmanned Aerial Vehicles (UAVs) in GNSS-denied environments is hindered by the lack of diverse datasets and the limited real-world performance of existing visual matching methods. To address these gaps, we introduce two contributions: (1) the SatLoc dataset, a new benchmark featuring synchronized, multi-source data from varied real-world scenarios tailored for UAV-to-satellite image matching, and (2) SatLoc-Fusion, a hierarchical localization framework. Our proposed pipeline integrates three complementary layers: absolute geo-localization via satellite imagery using DinoV2, high-frequency relative motion tracking from visual odometry with XFeat, and velocity estimation using optical flow. An adaptive fusion strategy dynamically weights the output of each layer based on real-time confidence metrics, ensuring an accurate and self-consistent state estimate. Deployed on a 6 TFLOPS edge computer, our system achieves real-time operation at over 2 Hz, with an absolute localization error below 15 m and effective trajectory coverage exceeding 90%, demonstrating state-of-the-art performance. The SatLoc dataset and fusion pipeline provide a robust and comprehensive baseline for advancing UAV navigation in challenging environments. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

25 pages, 24334 KB  
Article
Unsupervised Knowledge Extraction of Distinctive Landmarks from Earth Imagery Using Deep Feature Outliers for Robust UAV Geo-Localization
by Zakhar Ostrovskyi, Oleksander Barmak, Pavlo Radiuk and Iurii Krak
Mach. Learn. Knowl. Extr. 2025, 7(3), 81; https://doi.org/10.3390/make7030081 - 13 Aug 2025
Viewed by 994
Abstract
Vision-based navigation is a common solution for the critical challenge of GPS-denied Unmanned Aerial Vehicle (UAV) operation, but a research gap remains in the autonomous discovery of robust landmarks from aerial survey imagery needed for such systems. In this work, we propose a [...] Read more.
Vision-based navigation is a common solution for the critical challenge of GPS-denied Unmanned Aerial Vehicle (UAV) operation, but a research gap remains in the autonomous discovery of robust landmarks from aerial survey imagery needed for such systems. In this work, we propose a framework to fill this gap by identifying visually distinctive urban buildings from aerial survey imagery and curating them into a landmark database for GPS-free UAV localization. The proposed framework constructs semantically rich embeddings using intermediate layers from a pre-trained YOLOv11n-seg segmentation network. This novel technique requires no additional training. An unsupervised landmark selection strategy, based on the Isolation Forest algorithm, then identifies objects with statistically unique embeddings. Experimental validation on the VPAIR aerial-to-aerial benchmark shows that the proposed max-pooled embeddings, assembled from selected layers, significantly improve retrieval performance. The top-1 retrieval accuracy for landmarks more than doubled compared to typical buildings (0.53 vs. 0.31), and a Recall@5 of 0.70 is achieved for landmarks. Overall, this study demonstrates that unsupervised outlier selection in a carefully constructed embedding space yields a highly discriminative, computation-friendly set of landmarks suitable for real-time, robust UAV navigation. Full article
(This article belongs to the Special Issue Deep Learning in Image Analysis and Pattern Recognition, 2nd Edition)
Show Figures

Graphical abstract

23 pages, 3858 KB  
Article
MCFA: Multi-Scale Cascade and Feature Adaptive Alignment Network for Cross-View Geo-Localization
by Kaiji Hou, Qiang Tong, Na Yan, Xiulei Liu and Shoulu Hou
Sensors 2025, 25(14), 4519; https://doi.org/10.3390/s25144519 - 21 Jul 2025
Viewed by 1220
Abstract
Cross-view geo-localization (CVGL) presents significant challenges due to the drastic variations in perspective and scene layout between unmanned aerial vehicle (UAV) and satellite images. Existing methods have made certain advancements in extracting local features from images. However, they exhibit limitations in modeling the [...] Read more.
Cross-view geo-localization (CVGL) presents significant challenges due to the drastic variations in perspective and scene layout between unmanned aerial vehicle (UAV) and satellite images. Existing methods have made certain advancements in extracting local features from images. However, they exhibit limitations in modeling the interactions among local features and fall short in aligning cross-view representations accurately. To address these issues, we propose a Multi-Scale Cascade and Feature Adaptive Alignment (MCFA) network, which consists of a Multi-Scale Cascade Module (MSCM) and a Feature Adaptive Alignment Module (FAAM). The MSCM captures the features of the target’s adjacent regions and enhances the model’s robustness by learning key region information through association and fusion. The FAAM, with its dynamically weighted feature alignment module, adaptively adjusts feature differences across different viewpoints, achieving feature alignment between drone and satellite images. Our method achieves state-of-the-art (SOTA) performance on two public datasets, University-1652 and SUES-200. In generalization experiments, our model outperforms existing SOTA methods, with an average improvement of 1.52% in R@1 and 2.09% in AP, demonstrating its effectiveness and strong generalization in cross-view geo-localization tasks. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

23 pages, 29759 KB  
Article
UAV-Satellite Cross-View Image Matching Based on Adaptive Threshold-Guided Ring Partitioning Framework
by Yushi Liao, Juan Su, Decao Ma and Chao Niu
Remote Sens. 2025, 17(14), 2448; https://doi.org/10.3390/rs17142448 - 15 Jul 2025
Cited by 1 | Viewed by 3458
Abstract
Cross-view image matching between UAV and satellite platforms is critical for geographic localization but remains challenging due to domain gaps caused by disparities in imaging sensors, viewpoints, and illumination conditions. To address these challenges, this paper proposes an Adaptive Threshold-guided Ring Partitioning Framework [...] Read more.
Cross-view image matching between UAV and satellite platforms is critical for geographic localization but remains challenging due to domain gaps caused by disparities in imaging sensors, viewpoints, and illumination conditions. To address these challenges, this paper proposes an Adaptive Threshold-guided Ring Partitioning Framework (ATRPF) for UAV–satellite cross-view image matching. Unlike conventional ring-based methods with fixed partitioning rules, ATRPF innovatively incorporates heatmap-guided adaptive thresholds and learnable hyperparameters to dynamically adjust ring-wise feature extraction regions, significantly enhancing cross-domain representation learning through context-aware adaptability. The framework synergizes three core components: brightness-aligned preprocessing to reduce illumination-induced domain shifts, hybrid loss functions to improve feature discriminability across domains, and keypoint-aware re-ranking to refine retrieval results by compensating for neural networks’ localization uncertainty. Comprehensive evaluations on the University-1652 benchmark demonstrate the framework’s superiority; it achieves 82.50% Recall@1 and 84.28% AP for UAV→Satellite geo-localization, along with 90.87% Recall@1 and 80.25% AP for Satellite→UAV navigation. These results validate the framework’s capability to bridge UAV–satellite domain gaps while maintaining robust matching precision under heterogeneous imaging conditions, providing a viable solution for practical applications such as UAV navigation in GNSS-denied environments. Full article
(This article belongs to the Special Issue Temporal and Spatial Analysis of Multi-Source Remote Sensing Images)
Show Figures

Figure 1

22 pages, 23971 KB  
Article
Remote Target High-Precision Global Geolocalization of UAV Based on Multimodal Visual Servo
by Xuyang Zhou, Ruofei He, Wei Jia, Hongjuan Liu, Yuanchao Ma and Wei Sun
Remote Sens. 2025, 17(14), 2426; https://doi.org/10.3390/rs17142426 - 12 Jul 2025
Cited by 1 | Viewed by 836
Abstract
In this work, we propose a geolocation framework for distant ground targets integrating laser rangefinder sensors with multimodal visual servo control. By simulating binocular visual servo measurements through monocular visual servo tracking at fixed time intervals, our approach requires only single-session sensor attitude [...] Read more.
In this work, we propose a geolocation framework for distant ground targets integrating laser rangefinder sensors with multimodal visual servo control. By simulating binocular visual servo measurements through monocular visual servo tracking at fixed time intervals, our approach requires only single-session sensor attitude correction calibration to accurately geolocalize multiple targets during a single flight, which significantly enhances operational efficiency in multi-target geolocation scenarios. We design a step-convergent target geolocation optimization algorithm. By adjusting the step size and the scale factor of the cost function, we achieve fast accuracy convergence for different UAV reconnaissance modes, while maintaining the geolocation accuracy without divergence even when the laser ranging sensor is turned off for a short period. The experimental results show that through the UAV’s continuous reconnaissance measurements, the geolocalization error of remote ground targets based on our algorithm is less than 7 m for 3000 m, and less than 3.5 m for 1500 m. We have realized the fast and high-precision geolocalization of remote targets on the ground under the high-altitude reconnaissance of UAVs. Full article
Show Figures

Figure 1

31 pages, 31711 KB  
Article
On the Usage of Deep Learning Techniques for Unmanned Aerial Vehicle-Based Citrus Crop Health Assessment
by Ana I. Gálvez-Gutiérrez, Frederico Afonso and Juana M. Martínez-Heredia
Remote Sens. 2025, 17(13), 2253; https://doi.org/10.3390/rs17132253 - 30 Jun 2025
Cited by 3 | Viewed by 1355
Abstract
This work proposes an end-to-end solution for leaf segmentation, disease detection, and damage quantification, specifically focusing on citrus crops. The primary motivation behind this research is to enable the early detection of phytosanitary problems, which directly impact the productivity and profitability of Spanish [...] Read more.
This work proposes an end-to-end solution for leaf segmentation, disease detection, and damage quantification, specifically focusing on citrus crops. The primary motivation behind this research is to enable the early detection of phytosanitary problems, which directly impact the productivity and profitability of Spanish and Portuguese agricultural developments, while ensuring environmentally safe management practices. It integrates an onboard computing module for Unmanned Aerial Vehicles (UAVs) using a Raspberry Pi 4 with Global Positioning System (GPS) and camera modules, allowing the real-time geolocation of images in citrus croplands. To address the lack of public data, a comprehensive database was created and manually labelled at the pixel level to provide accurate training data for a deep learning approach. To reduce annotation effort, we developed a custom automation algorithm for pixel-wise labelling in complex natural backgrounds. A SegNet architecture with a Visual Geometry Group 16 (VGG16) backbone was trained for the semantic, pixel-wise segmentation of citrus foliage. The model was successfully integrated as a modular component within a broader system architecture and was tested with UAV-acquired images, demonstrating accurate disease detection and quantification, even under varied conditions. The developed system provides a robust tool for the efficient monitoring of citrus crops in precision agriculture. Full article
(This article belongs to the Special Issue Application of Satellite and UAV Data in Precision Agriculture)
Show Figures

Figure 1

19 pages, 6772 KB  
Article
A Cross-Mamba Interaction Network for UAV-to-Satallite Geolocalization
by Lingyun Tian, Qiang Shen, Yang Gao, Simiao Wang, Yunan Liu and Zilong Deng
Drones 2025, 9(6), 427; https://doi.org/10.3390/drones9060427 - 12 Jun 2025
Cited by 1 | Viewed by 1545
Abstract
The geolocalization of unmanned aerial vehicles (UAVs) in satellite-denied environments has emerged as a key research focus. Recent advancements in this area have been largely driven by learning-based frameworks that utilize convolutional neural networks (CNNs) and Transformers. However, both CNNs and Transformers face [...] Read more.
The geolocalization of unmanned aerial vehicles (UAVs) in satellite-denied environments has emerged as a key research focus. Recent advancements in this area have been largely driven by learning-based frameworks that utilize convolutional neural networks (CNNs) and Transformers. However, both CNNs and Transformers face challenges in capturing global feature dependencies due to their restricted receptive fields. Inspired by state-space models (SSMs), which have demonstrated efficacy in modeling long sequences, we propose a pure Mamba-based method called the Cross-Mamba Interaction Network (CMIN) for UAV geolocalization. CMIN consists of three key components: feature extraction, information interaction, and feature fusion. It leverages Mamba’s strengths in global information modeling to effectively capture feature correlations between UAV and satellite images over a larger receptive field. For feature extraction, we design a Siamese Feature Extraction Module (SFEM) based on two basic vision Mamba blocks, enabling the model to capture the correlation between UAV and satellite image features. In terms of information interaction, we introduce a Local Cross-Attention Module (LCAM) to fuse cross-Mamba features, providing a solution for feature matching via deep learning. By aggregating features from various layers of SFEMs, we generate heatmaps for the satellite image that help determine the UAV’s geographical coordinates. Additionally, we propose a Center Masking strategy for data augmentation, which promotes the model’s ability to learn richer contextual information from UAV images. Experimental results on benchmark datasets show that our method achieves state-of-the-art performance. Ablation studies further validate the effectiveness of each component of CMIN. Full article
Show Figures

Figure 1

Back to TopTop