Next Article in Journal
Research on Lightweight Dynamic Security Protocol for Intelligent In-Vehicle CAN Bus
Previous Article in Journal
Short-Term Photovoltaic Power Forecasting Based on an Improved Zebra Optimization Algorithm—Stochastic Configuration Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Integrated Navigation Method Based on the Strapdown Inertial Navigation System/Scene-Matching Navigation System for UAVs

1
School of Mechatronical Engineering, Beijing Institute of Technology, Beijing 100081, China
2
The Key Laboratory of Intelligent Perception and Image Understanding of Ministry of Education, School of Artificial Intelligence, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3379; https://doi.org/10.3390/s25113379
Submission received: 24 April 2025 / Revised: 20 May 2025 / Accepted: 26 May 2025 / Published: 27 May 2025
(This article belongs to the Section Navigation and Positioning)

Abstract

To address the challenges of discontinuous heterogeneous image matching, significant matching errors in specific regions, and poor real-time performance in GNSS-denied environments for unmanned aerial vehicles (UAVs), we propose an integrated navigation method based on the strapdown inertial navigation system (SINS)/scene-matching navigation system (SMNS). First, we designed a heterogeneous image-matching and positioning approach using infrared images to obtain an estimation of the UAV’s position. Then, we established a mathematical model for the integrated SINS/SMNS navigation system. Finally, a Kalman filter (KF) was employed to fuse the inertial navigation data with absolute position data from scene matching, achieving high-precision and highly reliable navigation positioning. We constructed a navigation data acquisition platform and conducted simulation studies using flight data collected from this platform. The results demonstrate that the integrated SINS/SMNS navigation method significantly outperforms standalone scene-matching navigation in horizontal positioning accuracy, improving latitude accuracy by 52.34% and longitude accuracy by 45.54%.

1. Introduction

Unmanned aerial vehicles (UAVs) have rapidly emerged as a significant technology, attracting considerable attention from various countries. They have been extensively applied across diverse fields, including agriculture, military, and meteorology, significantly influencing the development of these sectors [1]. In the military domain, UAV applications primarily encompass aerial reconnaissance, terminal guidance for cruise navigation, and more [2]. With the rise of the low-altitude economy, the demand for UAVs in the civilian sector is increasing sharply. Beyond surveillance, management, emergency rescue, and climate regulation, emerging application scenarios for UAVs are being continuously developed. In UAV systems, the navigation and positioning subsystem is one of the most critical components, as its performance directly impacts flight safety and mission success.
At present, the most well-developed and effective UAV navigation and positioning system relies on the integration of satellite navigation and inertial navigation technologies [3]. However, satellite navigation can become unreliable in challenging terrains such as mountains or valleys. Moreover, frequent interference with satellite signals during wartime or conflicts significantly degrades the performance of satellite navigation [4]. Inertial navigation systems, on the other hand, suffer from error accumulation over time, which limits their ability to support independent long-term navigation for UAVs. F. D’Ippolito et al. [5] propose a hybrid observer that fuses inertial data with sporadic position updates, achieving fast and robust estimation with formal ISS guarantees. In [6], a vision-based localization method is introduced, where UAV geolocalization is achieved through image matching between onboard camera frames and orthophotos, without requiring GNSS or external infrastructure. Previous research [7] categorizes visual positioning methodologies for UAV navigation in GNSS-deprived scenarios into two principal classifications—Relative Vision Localization (RVL) and Absolute Vision Localization (AVL)—which adopt varied approaches to exploit imaging data for positional estimation. Recent analyses highlight sensor fusion as critical for UAV navigation in GNSS-denied settings. Tong et al. [8] systematically evaluate SLAM architectures and swarm visual localization, demonstrating how the synergistic fusion of visual–inertial, LiDAR–inertial, and LiDAR–visual configurations overcomes single-modality limitations. The SMNS, characterized by its simple structure, strong environmental adaptability, and robust implementation capability, has proven to be an excellent navigation solution in GNSS-denied environments [9].
To successfully implement an SMNS, the following sequential steps need to be undertaken: (1) acquiring and preprocessing infrared images and reference satellite maps to enable image registration; (2) selecting the matching region of interest (ROI) areas based on the pre-planned trajectory of the UAV or the real-time inertial position; (3) performing image matching, through which the geographic coordinate information of key points within the matching ROI areas is obtained; (4) estimating the UAV attitude, including visual position calculation, attitude determination, and correcting inertial navigation deviations through inertial data fusion [10]. Among these steps, image matching is the critical step that directly influences the accuracy of UAV attitude estimation.
Multi-modal image matching is a crucial technology that addresses image matching challenges arising from various image acquisition sources in the SMNS [11]. Generally, there are two paradigms for multi-modal image registration: the first paradigm is the feature-based matching method, which involves feature extraction, feature description, and feature matching; the second paradigm is image matching via template matching [12]. Based on the types of features extracted, feature matching techniques can be categorized into methods based on point features and methods based on structural features. Among these, point feature-based matching methods require less computational effort, making them easier and more efficient to implement. As a result, they are particularly suitable for hardware environments with limited resources. Consequently, numerous studies have focused on such methods under conditions of constrained hardware computing performance in the past, including SIFT (Scale-Invariant Feature Transform), SURF (Speeded-Up Robust Features) [13], FAST (Features from Accelerated Segment Test) [14], BRIEF (Binary Robust Independent Elementary Features) [15], and ORB (Oriented FAST and Rotated BRIEF) [16].
The classical features mentioned above are primarily extracted at the pixel gray level, making them highly susceptible to illumination variations. However, they are well suited for image matching in scenarios with minimal gray-level changes between successive frames. Due to their robustness, these classical features are widely utilized in visual odometry, SLAM, UAV optical flow, and other related applications. In recent years, many researchers have focused on improving these methods based on classical features to achieve multi-modal image matching. To mitigate the impact of nonlinear radiation distortion (NRD), the nonlinear diffusion scale space was introduced in PSO-SIFT [17]. This approach employs a nonlinear diffusion filter and multi-scale parameters to extract uniform gradient information between Synthetic Aperture Radar (SAR) and visible images. Deng et al. [18] proposed a two-step matching approach combining global and local matching to obtain more control points for SIFT-like very-high-resolution SAR image registration, significantly improving the number and quality of matched control points. Zhang et al. [19] combined SIFT with Canny edge detection to remove unstable points and smooth SAR images, significantly enhancing registration accuracy and speed using FLANN and PROSAC. Wu et al. [20] proposed an improved ORB algorithm using affine transformation to extract stable descriptor bits, improving matching accuracy and employing an enhanced F-SORT for refined matching. Zhang et al. [21] proposed an OS-SIFT-based method with a cascaded sample consensus approach for robust optical and SAR image registration, improving gradient consistency and increasing correct correspondences to enhance registration accuracy and robustness. Nehme et al. [22] proposed a deep learning-based wavefront-shaping method that optimizes the optical parameters of a multi-channel imaging system, thereby enhancing image resolution and feature extraction capabilities. Jhan et al. [23] introduced a normalized SURF (N-SURF) method for multi-spectral image matching, significantly increasing the number of correct matching points. The effectiveness of N-SURF was remarkable, with the number of correct matches being several times or even an order of magnitude higher than that of the traditional SURF method. In addition to the classical feature-based methods mentioned above, methods based on structural feature extraction exhibit stronger robustness in image matching. Li et al. [24] proposed a Radiation-Variation-Insensitive Feature Transform (RIFT) which utilizes phase congruency for matching and demonstrates greater robustness compared to traditional methods. Yu et al. [25] introduced a novel consistent feature transform (NCFT), designed to address the significant NRD problem between multi-modal images, aiming to extract rich and robust features.
Yao et al. [26] proposed a space-matching method based on a co-occurrence filter, which can effectively weaken or eliminate NRD while extracting richer structural features in the co-occurrence space. Rouse et al. [27] introduced a structural similarity (SSIM) metric for image quality assessment. SSIM compares the luminance, contrast, and structural information of images, enabling a more accurate reflection of human visual system perception of image quality. Additionally, Yao et al. [28] proposed the Multi-Orientation Tensor Index Feature (MOTIF) for the registration of Synthetic Aperture Radar (SAR) images and optical images. By extracting structural information from multiple orientations, MOTIF significantly reduces the impact of inherent speckle noise in SAR images on the registration process. This method demonstrates strong robustness in the task of SAR and optical image registration, effectively improving registration accuracy. Furthermore, research based on the histogram of absolute phase consistency gradients (HAPCG) [29,30] introduced an anisotropic weighted moment diagram. This diagram effectively extracted edge extreme points from the images and enhanced the number of matching pairs by utilizing the absolute phase direction to determine the main direction and generate characteristic descriptors. These algorithms have significantly advanced research on multi-modal remote sensing image matching. However, their applicability is limited by various factors, including geographical location, scale, rotation, and computational complexity.
The second method to solve the matching problem is template matching based on grayscale correlation or structural similarity. Early image registration methods primarily focused on pixel-gray-level similarity [31]. These methods assess image similarity based on pixel gray levels, making them unsuitable for multi-modal image registration. Currently, robust template matching techniques primarily utilize the structural features of images.
These methods better reflect the common properties of multi-modal images, such as gradients and tensors [32]. Therefore, Ye et al. [33] proposed a matching method based on structural features by introducing the concept of the Histogram of Oriented Phase Congruency (HOPC). An improved HOPC algorithm, called Channel Features of Oriented Gradients (CFOG) [34], was proposed to enhance efficiency by characterizing image structure features pixel by pixel, and template matching was accelerated using the Fourier transform. Ruslan et al. [35] introduced a structure tensor-based multi-modal image-matching method for unmanned aerial vehicle (UAV) scene-matching systems. By extracting the structural tensor features of images, this method can effectively handle illumination variations and demonstrates high robustness and accuracy in multi-modal image matching.
Parallax is calculated based on the positional differences in matching points across different images. Since parallax is related to the distance between the object and the camera, as well as the camera’s attitude, it can be used to derive information about scene depth and changes in the drone’s relative position. By integrating the strapdown inertial navigation system (SINS) [36] with the scene-matching navigation system (SMNS), the UAV can ensure the completion of regular flight tasks even in the absence of satellite positioning signals.
To address the positioning challenges of UAVs in GNSS-denied environments, we propose an integrated navigation framework based on SINS/SMNS fusion. A schematic diagram of the proposed method is illustrated in Figure 1. The key methodological advancements of this work are threefold:
  • A real-time infrared image orthorectification technique based on SINS data is introduced to reduce the impact of UAV attitude on image matching. This method features simple operation and high computational efficiency, delivering excellent real-time performance while achieving optimal processing results.
  • A Log-Gabor [37] filter-based feature extraction method is proposed to extract the structural features of images, addressing the critical challenges of multi-modal image matching and achieving highly robust matching results between real-time images and reference images.
  • A cascaded Kalman filtering [38] mechanism is designed to integrate high-frequency SINS measurements with SMNS positional updates. This fusion strategy reduces cumulative errors by 52.34% in latitude and 45.54% in longitude compared to standalone SMNS implementations.
Figure 1. Schematic diagram of SINS/SMNS integrated navigation.
Figure 1. Schematic diagram of SINS/SMNS integrated navigation.
Sensors 25 03379 g001
Through the above procedures, the UAV can achieve long-term stable navigation performance. To validate the real-world performance of the proposed method, extensive simulation experiments were conducted using real flight data. The experimental results further confirm the effectiveness and practicality of the proposed method.
The remainder of this paper is organized as follows: Section 2, Section 3 and Section 4 present the three main procedures of the proposed method, while Section 5 provides the experimental results and analysis. Finally, Section 6 concludes the paper with a summary of the work.

2. Orthorectification of Oblique Images Based on Inertial Attitude

In general, scene-matching navigation systems utilize satellite images derived from orthophotos as the reference map. However, real-time images captured by the UAV at different attitude angles during flight exhibit significant geometric distortion. This distortion substantially impacts the matching accuracy between the real-time image and the reference satellite image. Therefore, it is essential to correct the real-time image into an orthophoto using the camera installation error, UAV attitude, and camera-intrinsic parameters prior to real-time scene matching.
The coordinate frames used in this paper are defined as follows:
  • OXbYbZb: UAV body frame. The x-axis points to the right wing, the y-axis points to the front of the UAV, and the z-axis points vertically upward.
  • OXcYcZc: camera frame. O represents the optical center of the camera, which may not coincide with the body-frame center O. The x-axis points to the right wing, the y-axis points to the tail of the UAV, and the z-axis points vertically downward.
  • OXgYgZg: local gravity frame. The origin is centered at the UAV’s center of mass. The x-axis points geographically east, the y-axis points geographically north, and the z-axis points vertically upward, perpendicular to the local reference ellipsoid surface, and is almost opposite to the direction of gravity.
The camera is rigidly mounted on the UAV, with OZc aligned parallel to the normal axis of the UAV and oriented perpendicular to the UAV body, pointing downward. This configuration ensures that the real-time ground scene can be continuously captured during the UAV’s movement.
In this study, we utilize attitude information from a high-precision laser inertial navigation system to assist in the orthorectification of images, with the specific operations illustrated in Figure 2. Firstly, we identify the control points within the UAV-captured real-time images, and their positions in the camera frame are computed based on the intrinsic parameters of the camera. Given that the camera is co-aligned with the UAV, the attitude matrix enables the projection of these control points into the local gravity frame. Subsequently, within the local gravity frame, the positions of the control points relative to the UAV can be determined by using the triangle similarity principle, enabling the calculation of their geographical positions.
The coordinates of the control points are illustrated in Figure 3. The image resolution is w × h , where w represents the pixel width of the image, and h represents the pixel height of the image. The four corner points and the center point are selected as the control points, denoted as P i ( x i , y i ) (where i = 1, 2, …, 5). Their coordinates in the pixel frame are as follows: P 1 ( 0 ,   0 ) , P 2 w ,   0 , P 3 ( 0 ,   h ) , P 4 ( w ,   h ) , and P 5 ( w / 2 ,   h / 2 ) .
The coordinates of the control points in the camera frame can be expressed as follows:
P c i = M p c P i 1 = M p c x i y i 1
M p c = P x y 0 w 2 P x y 0 P x y h 2 P x y 0 0 f
In these equations, P c i represents the coordinates of the control points in the camera frame, M p c is the transformation matrix mapping the control point positions to the camera frame, P x y denotes the camera sensor’s pixel size (unit: mm/pixel), and f represents the focal length of the camera.
Since the distortion of industrial cameras is typically very small and usually unknown, the impact of camera distortion on orthographic images can be essentially neglected. Therefore, the positions of the control points in the camera frame are determined using the intrinsic parameters of the camera.
Denoting the pitch angle of the UAV by θ, the roll angle by ϕ , and the yaw angle by ψ , the direction cosine matrix R of the UAV can be expressed as Equation (3):
R = cos ϕ cos ψ + sin ϕ sin θ sin ψ cos θ sin ψ sin ϕ cos ψ cos ϕ sin θ sin ψ cos ϕ sin ψ + sin ϕ sin θ cos ψ cos θ cos ψ sin ϕ sin ψ cos ϕ sin θ cos ψ sin ϕ cos θ sin θ cos ϕ cos θ
The camera and UAV are rigidly mounted, and the control points P g i in the local gravity frame can be calculated using the matrix R . It can be expressed as follows:
P g i = x g i y g i z g i = R 1 P c i = R 1 x c i y c i f
According to the definition of the local gravity frame, the z-axis is perpendicular to the ground and opposite to the direction of gravity. The intersection point of the z-axis with the ground represents the relative flight height of the UAV, denoted as h r . Assuming that the flight altitude of the UAV is h b , the elevation of the ground point directly below the UAV can be obtained from the DEM, denoted as h g i . The relative height of the control point is denoted as z d i , and z d i = h r = h b h g i . Therefore, the position of the control point in the local gravity frame can be calculated based on the triangle similarity principle, i.e.,
P d i = z d i z g i P g i = h b h g i z g i P g i = x g i z g i ( h b h g i ) y g i z g i ( h b h g i ) h b h g i
Since R is a unit orthogonal matrix, we can assume that
R 1 = r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33
Based on Equations (1), (2) and (4), it can be inferred that
z g i = x i w / 2 P x y r 31 + y i h / 2 P x y r 32 + r 33 f
Based on Equations (1)–(6), Equation (8) can be inferred as follows:
P d i = z d i z g i R 1 M p c P i 1 = h b h g i z g i r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 P x y 0 ( w / 2 ) P x y 0 P x y ( h / 2 ) P x y 0 0 f x i y i 1 = ( h b h g i ) ( x i w / 2 ) P x y r 11 + ( y i h / 2 ) P x y r 12 + r 13 f ( x i w / 2 ) P x y r 31 + ( y i h / 2 ) P x y r 32 + r 33 f ( x i w / 2 ) P x y r 21 + ( y i h / 2 ) P x y r 22 + r 23 f ( x i w / 2 ) P x y r 31 + ( y i h / 2 ) P x y r 32 + r 33 f 1
To facilitate image registration, it is essential to ensure that the scale of the orthophoto should match that of the reference satellite image. And the resolution of the orthograph should be the same as the reference satellite image. The resolution accuracy of the reference map can be expressed as m r e s , and the scaling and translation operations on the control points can be computed as shown in Equation (9).
P n i = P d i P d 5 m r e s , i = 1 ,   2 ,   3 ,   4
The affine matrix can be calculated using the Gaussian elimination method by designating the four corners of the image as control points. Assuming that the matrix is denoted as M A , it can be formulated as follows:
M A = a b c d e f 0 0 1
P n i = M A P i 1 = M A x i y i 1
Through the substitution of P n i (where i = 1, 2, 3, 4) and P i (where i = 1, 2, 3, 4) into the above equations, the solution can be obtained using the Gaussian elimination method. Consequently, the orthographic image can be generated by applying the affine matrix M A .

3. Image Registration Method Based on Maximum Index Map

In image matching, structural features are more suitable for multi-modal image matching. The template-matching method exhibits greater robustness compared to feature point-based methods, particularly when the scale of the template image aligns with that of the reference image. Additionally, the template-matching approach eliminates the need for complex steps, such as RANSAC, significantly enhancing the overall robustness of the system. Therefore, in this work, we adopt the template-matching [39] approach to achieve image matching. The image-matching process comprises two main steps:
  • Calculate the structural feature maps of the infrared image and the satellite map using the Log-Gabor filter;
  • Utilize template matching to achieve automatic image matching.

3.1. Log-Gabor Filter

Compared with the Gabor filter, the Log-Gabor filter can effectively capture the local frequency-domain information of an image. Moreover, its bandwidth and center frequency can be manually configured. The expression is as follows:
G ( ω ) = e l o g ( ω / ω 0 ) 2 2 log ( ( k / ω 0 ) 2 )
In this equation, ω 0 represents the center frequency of the filter. To obtain constant shape ratio filters, the term k / ω 0 must remain constant for varying ω 0 .
A Two-Dimensional Log-Gabor filter (2D-LGF) consists of two components—a radial filter G r ( r ) and an angle filter G θ ( θ ) —both of which are expressed in Gaussian form:
G r ( r ) = exp ( l o g ( r / f 0 ) 2 2 σ r 2 ) G θ ( θ ) = exp ( ( θ θ 0 ) 2 2 σ θ 2 )
G ( r , θ ) = G r ( r ) G θ ( θ )
In this equation, r and θ represent the radial and angular components, respectively; σ r denotes the radial bandwidth and σ θ represents the angular bandwidth of the filter; and f 0 and θ 0 denote the center frequency and angular direction, respectively. The expression for f 0 is as follows:
f 0 = 1 ω k m t s 1
In this equation, ω represents the frequency parameter, k is a constant typically associated with the filter design, m t denotes the scale parameter, and s is the scale index.
The Log-Gabor filter can only be expressed in the frequency domain, while its spatial-domain representation is expressed in complex form. Through the inverse Fourier transform, the 2D-LGF can be expressed as follows:
G s o ( x , y ) = G s o e ( x , y ) + i G s o o ( x , y )
In this equation, G s o e and G s o o represent the even-symmetric and odd-symmetric filters in terms of scale and direction, respectively. Additionally, G s o e and G s o o are a pair of orthogonal filters.

3.2. Maximum Index Map (MIM)

The MIM [40] is constructed using the Log-Gabor filter. For an image M , M ( x , y ) represents the pixel value at coordinates ( x , y ) (Figure 4). The convolution components of the even part E s o ( x , y ) and the odd part O s o ( x , y ) can be obtained by applying a 2D-LGF:
E s o ( x , y ) , O s o ( x , y ) = M ( x , y ) G s o E , M ( x , y ) G s o O
For a given orientation o and scale s , G s o O denotes the odd-symmetric filter, and G s o E denotes the even-symmetric filter. The phase ϕ s o ( x , y ) and amplitude A s o ( x , y ) of M ( x , y ) are computed as follows:
ϕ s o ( x , y ) = atan 2 ( E s o ( x , y ) , O s o ( x , y ) )
A s o ( x , y ) = E s o ( x , y ) 2 + O s o ( x , y ) 2
For a given orientation o , the amplitudes across all scales are summed to obtain a Log-Gabor layer, expressed as follows:
A o ( x , y ) = A s o ( x , y )
Next, to determine the maximum value A k ( x , y ) and the corresponding orientation index k , we extract the maximum index map value I M I M ( x , y ) for the maximum index k , i.e.,
I M I M ( x , y ) = k ,   where A k ( x , y ) = max { A o ( x , y ) }
The maximum index maps for both the infrared image and the satellite map can be independently computed, effectively mitigating the impact of multi-modal imaging variations caused by differences in illumination and radiation. Subsequently, registration between the infrared image and the satellite map can be achieved using template-matching methods.

3.3. Template Matching

Common template-matching methods include the Sum of Squared Differences (SSD) [41,42], the Normalized Correlation Coefficient (NCC) [41,43], and Mutual Information (MI) [44]. The SSD is one of the simplest similarity metrics, as it identifies control points by directly computing intensity differences between two images. However, the SSD is highly sensitive to radiometric changes, despite its computational efficiency. The NCC is a widely used similarity metric, particularly in remote sensing image registration, due to its invariance to linear intensity variations.
In the template-matching process (Figure 5), the MIM of the infrared image serves as the template, while the MIM of the satellite reference map acts as the reference map. The template is systematically moved from left to right and top to bottom across the reference map. Following the correlation operation, a correlation image is generated, where each pixel value represents the similarity between the infrared image and the satellite reference map.
R ( x , y ) = x , y ( T ( x , y ) I ( x + x , y + y ) ) x , y T ( x , y ) 2 x , y I ( x + x , y + y ) 2
In this equation, T represents the template and I denotes the reference map, and ( x ,   y ) and x , y are the horizontal and vertical coordinates of the pixels in the template and reference image, respectively.
According to Equation (22), the position where R ( x ,   y ) reaches its maximum value corresponds to the matching point between the infrared image and the reference image. Since the position of each pixel in the reference map is known, the absolute horizontal positions of the four corner points of the real-time image within the reference map can be determined. Additionally, altitude information can be obtained from the drone’s barometer.
In template matching, the confidence level of registration results is inversely proportional to the spread of correlation peaks. This study establishes reliability criteria based on peak dominance analysis: when the secondary maximum correlation coefficient R ( x , y ) beyond a 15-pixel radius from the primary peak exhibits > 30% attenuation relative to the principal response, the matching result is deemed credible. This threshold-based peak discrimination enables a computationally efficient confidence assessment.
The spatial extent of satellite reference imagery is dynamically adjusted based on temporal proximity to the last successful georegistration event. Frequent successful matches within brief intervals signify sustained positional consistency between the estimated and actual coordinates, enabling constrained search areas. Conversely, prolonged intervals since the last validated match necessitate the progressive expansion of the reference coverage to accommodate potential inertial navigation drift, thereby maintaining reliable alignment between real-time aerial imagery and geospatial reference databases.
So far, we have established four sets of 3D-2D corresponding points, which constitute a Perspective-n-Point (PnP) problem [45]. Based on the PnP problem-solving methodology, the four sets of correspondences can be utilized to derive a unique solution, namely, the UAV’s pose. Consequently, by solving the PnP problem, we ultimately determine the position of the UAV based on the scene-matching approach.

4. Integrated Navigation Method Based on SINS/SMNS

Due to the low frequency of image registration, the speed of scene-matching position calculation is significantly reduced. Poor ground texture clarity further exacerbates this issue, leading to suboptimal scene-matching results and substantial errors in position estimation. To address these challenges and improve the positioning accuracy of UAVs in satellite-denied environments, we propose a sensor data fusion method that utilizes a KF to establish a mathematical model for an integrated INS/SMNS navigation system.
Generally, the KF consists of two main steps: state prediction and measurement update. The state prediction is based on the dynamic model of the Inertial Measurement Unit (IMU) and is propagated at a frequency of 200 Hz. When the navigation computer completes the motion estimation, the measurement update is triggered and executed. However, scene matching and pose estimation have been proven to be highly time-consuming processes. Therefore, compensating for the time-delay error in the SINS/SMNS system is essential to ensure the system’s real-time performance.

4.1. State Equation Design

The state equation of the integrated navigation system is formulated as follows:
X ˙ ( t ) 15 × 1 = F ( t ) 15 × 15 X ( t ) 15 × 1 + G ( t ) 15 × 6 W ( t ) 6 × 1
where X ( t ) 15 × 1 is the system state vector, F ( t ) 15 × 15 is the system transition matrix, G ( t ) 15 × 6 is the system noise driving matrix, and W ( t ) 6 × 1 is the system zero-mean white noise vector. The system state vector can be expressed as
X = ϕ e ϕ n ϕ u δ v e δ v n δ v u δ L δ λ δ h ε x ε y ε z x y z T        
where ϕ e , ϕ n , and ϕ u represent the misalignment angles; δ v e , δ v n , and δ v u represent velocity errors; δ L , δ λ , and δ h represent position errors; ε x , ε y , and ε z represent constant gyro drifts; and x , y , and z are constant accelerometer offsets.

4.2. Measurement Equation Design

The longitude, latitude, and altitude of the SMNS are represented by λ s m , L s m , and h s m , respectively. To convert these coordinates from the navigation coordinate system to Earth-Centered, Earth-Fixed (ECEF) coordinates ( X s m , Y s m , Z s m ), the following transformation formulas can be applied:
X s m = ( R n + h s m ) cos L s m cos λ s m Y s m = ( R n + h s m ) cos L s m sin λ s m Z s m = [ R n ( 1 e 2 ) + h s m ] sin L s m
where R n represents the radius of curvature in the prime vertical, and e denotes the eccentricity of the Earth’s ellipsoid.
Let λ i n s , L i n s , and h i n s represent the longitude, latitude, and altitude of the INS, respectively. The transformation from the navigation coordinate system to Earth-Centered, Earth-Fixed (ECEF) coordinates ( X i n s , Y i n s , Z i n s ) can be performed using the following formulas:
X i n s = ( R n + h i n s ) cos L i n s cos λ i n s Y i n s = ( R n + h i n s ) cos L i n s sin λ i n s Z i n s = [ R n ( 1 e 2 ) + h i n s ] sin L i n s
where δ λ = λ i n s λ s m , δ L = L i n s L s m , and δ h = h i n s h s m represent the longitude error, latitude error, and altitude error in the navigation coordinate system, respectively. δ X = X i n s X s m , δ Y = Y i n s Y s m , and δ Z = Z i n s Z s m denote the corresponding errors in the ECEF coordinate system.
Based on Equations (25) and (26), the values of δ X , δ Y , and δ Z can be derived as follows:
δ X = X i n s X s m = ( R n + h i n s ) cos L i n s cos λ i n s ( R n + h s m ) cos L s m cos λ s m = ( R n + h i n s ) cos L i n s cos λ i n s ( R n + h i n s δ h ) cos L i n s δ L cos λ i n s δ λ = ( R n + h i n s ) cos L i n s cos λ i n s cos L i n s δ L cos λ i n s δ λ + δ h cos L i n s δ L cos λ i n s δ λ = ( R n + h i n s ) cos L i n s sin λ i n s δ λ + sin L i n s cos λ i n s δ L + δ h cos L i n s cos λ i n s + η X
δ Y = Y i n s Y s m = ( R n + h i n s ) cos L i n s sin λ i n s ( R n + h s m ) cos L s m sin λ s m = ( R n + h i n s ) cos L i n s sin λ i n s ( R n + h i n s δ h ) cos L i n s δ L sin λ i n s δ λ = ( R n + h i n s ) cos L i n s sin λ i n s cos L i n s δ L sin λ i n s δ λ + δ h cos L i n s δ L sin λ i n s δ λ = ( R n + h i n s ) cos L i n s cos λ i n s δ λ sin L i n s sin λ i n s δ L + δ h cos L i n s sin λ i n s + η Y
δ Z = Z i n s Z s m = [ R n ( 1 e 2 ) + h i n s ] sin L i n s [ R n ( 1 e 2 ) + h s m ] sin L s m = [ R n ( 1 e 2 ) + h i n s ] sin L i n s [ R n ( 1 e 2 ) + h i n s δ h ] sin L i n s δ L = [ R n ( 1 e 2 ) + h i n s ] sin L i n s sin L i n s δ L + δ h sin L i n s δ L = [ R n ( 1 e 2 ) + h i n s ] cos L i n s δ L + sin L i n s δ h + η Z
where η X , η Y , and η Z represent second-order small error quantities in different directions.
Based on Equations (27)–(29), the measurement equation can be derived as follows:
Z 3 × 1 ( t ) = X i n s X s m Y i n s Y s m Z i n s Z s m = H 3 × 15 ( t ) X 15 × 1 ( t ) + N 3 × 1 ( t )
And H 3 × 15 can be expressed as follows:
H 3 × 15 = 0 3 × 3 0 3 × 3 M 3 × 3 0 3 × 6
where Z 3 × 1 ( t ) is the measurement vector, H 3 × 15 ( t ) is the observation matrix, and   N 3 × 1 ( t ) is the measurement noise driving matrix.
Additionally,   M 3 × 3 can be expressed as follows:
M 3 × 3 = ( R n + h i n s ) sin L i n s cos λ i n s ( R n + h i n s ) cos L i n s sin λ i n s cos L i n s cos λ i n s ( R n + h i n s ) sin L i n s sin λ i n s ( R n + h i n s ) cos L i n s cos λ i n s cos L i n s sin λ i n s [ R n ( 1 e 2 ) + h i n s ] cos L i n s 0 sin L i n s

4.3. Scene-Matching Position Error Compensation

The location for scene matching is obtained through a series of steps, including image acquisition, image processing, image matching, and position calculation. Notably, the time consumed by image processing and image matching often exceeds 500 ms. To ensure the real-time performance of navigation parameters, it is essential to compensate for the time delay in the location results derived from scene matching.
As shown in Figure 6, r I , k and r I , j   are the UAV positions estimated by inertial navigation at times t k and t j , respectively. The result of the scene-matching position calculation at time t k is r G , k . In reality, the position r G , j at time t j can be approximated as
r G , j = r G , k + ( r I , j r I , k )

4.4. Design and Implementation of KF

Through the discretization of the state equation and measurement equation of the integrated navigation system, the discrete system dynamic equation can be derived as follows:
X k = Φ k , k 1 X k 1 + Γ k , k 1 W k 1 Z k = H k X k + N k
In this equation, Φ k | k 1 is the state transition matrix, and τ is the sampling time: Φ k | k 1 = n = 0 F ( t k ) τ n n ! , Γ k , k 1 = n = 1 F ( t k ) τ n 1 n ! G ( t k ) τ .
The KF filtering process is described as follows:
X ^ k | k 1 = Φ k | k 1 X ^ k 1 P k | k 1 = Φ k | k 1 P k 1 Φ k , k 1 T + Γ k | k 1 Q k 1 Γ k | k 1 T K k = P k | k 1 H k T ( H k P k | k 1 H k T + R k ) 1 X ^ k = X ^ k | k 1 + K k ( Z k H k X ^ k | k 1 ) P k = ( I K k H k ) P k | k 1 ( I K k H k ) T + K k R k K k T
In this equation, P k | k 1 is the one-step prediction covariance matrix, Q k 1 is the system noise covariance matrix, K k is the filtering gain matrix, and R k is the covariance matrix of the observed value.

5. Experimental Results and Analysis

5.1. Integrated Navigation System Verification Platform

We have developed an integrated navigation system verification platform. This platform includes a day–night electro-optical reconnaissance payload, a laser inertial navigation system, an atmospheric data sensor system, a Beidou navigation receiver, a data logger, and a high-performance processor. The purpose of this platform is to collect flight data from various sensors to facilitate research on navigation algorithms.
The architecture of the UAV integrated navigation verification platform is illustrated in Figure 7. The day–night electro-optical reconnaissance payload is a device integrated with multiple electro-optical sensors, including an infrared camera, a visible light camera, and a laser rangefinder. The visible light camera is used for high-resolution image acquisition during the day, capable of capturing detailed color images. The infrared camera is used for thermal imaging at night or in low-light environments, capable of detecting and recording infrared radiation emitted by objects. The laser rangefinder is used to measure target distances. The laser inertial navigation system provides angular velocity and acceleration information. The Beidou navigation receiver provides velocity and position information. The atmospheric data sensor system provides airspeed and barometric altitude information. The high-performance processor module is used to collect data from various sensors and perform integrated navigation algorithm computations. The flight control system receives data from the integrated navigation system to participate in flight control. Component specifications of the integrated navigation system verification platform are detailed in Table 1 and Figure 8.

5.2. Experiments on Orthorectification

To verify the orthorectification results, real-world images of the UAV in turning, level, and climbing flight states were collected, as illustrated in Figure 9, Figure 10 and Figure 11. During a right turn, the UAV’s roll angle exceeded 24°, whereas the pitch angle exceeded 9° during climbing. A comparison was conducted between the corrected images and satellite reference images. The results indicate that the orthorectified UAV infrared images exhibit high consistency with the orthographic reference, thereby validating the effectiveness of the proposed method.

5.3. Experiments on Image Matching

Figure 12, Figure 13 and Figure 14 depict the matching results of the UAV in roll, level flight, and climb states, respectively. The results demonstrate that the UAV is capable of achieving robust matching across various fixed flight conditions. Nevertheless, when the UAV experiences substantial attitude changes, image registration, while feasible, exhibits increased mosaic errors at the image edges, which consequently elevate positioning errors. Conversely, under conditions of minimal attitude variations, the accuracy of image matching and mosaicking is significantly improved.
When the image texture is weak or fails to provide extractable features, the system automatically expands the ROI search range to ensure that UAV-captured images remain within the ROI boundaries. Table 2 presents the processing time statistics for template matching, Log-Gabor filtering, orthorectification, and KF algorithms under varying ROI sizes. As shown in Table 2, with the expansion of the ROI search range, the maximum latency of the template-matching process reaches 472.39 ms, while the Log-Gabor filter exhibits a maximum latency of 207.88 ms. In contrast, the processing times of both the orthorectification and KF algorithms remain virtually unaffected by ROI size variations. Specifically, the orthorectification process requires approximately 4 ms, and the KF algorithm completes its process in less than 0.03 ms.
Since the KF algorithm operates in a separate thread from template matching, Log-Gabor filtering, and orthorectification, and the laser inertial navigation system provides real-time positioning information, image processing delays can be effectively compensated for through a time-delay compensation algorithm. Consequently, the real-time performance of the SINS/SMNS integrated navigation system depends solely on the KF processing time, which is maintained well below 1 ms, fully meeting the operational requirements of unmanned aerial vehicles.

5.4. Experiments on Integrated Navigation

The location of the UAV is determined using the scene-matching positioning method outlined in Section 3. Figure 15, Figure 16 and Figure 17 illustrate the scene-matching position errors for the three flight paths. The mean error e m e a n is calculated as follows:
e m e a n = 1 n i = 1 n ( a i b i ) , i = 1,2 , . . . , n
where a i represents the sample value and b i denotes the reference value.
The standard deviation e s t d of the variable c i is calculated by
e s t d = 1 n i = 1 n ( c i e m e a n ) 2 , i = 1,2 , . . . , n
In this study, the positional data provided by the Beidou navigation receiver serve as the reference value. Based on Equations (36) and (37), we investigate the impact of two parameters—the UAV’s flight paths and its relative flight altitude—on the positioning accuracy of the scene-matching navigation system.
Figure 15a compares the positioning results of the SMNS and the Beidou navigation receiver during a rectangular flight path at a 3000 m relative altitude. Figure 15b illustrates the positioning error curve of the SMNS during a rectangular flight path at a 3000 m ground-relative altitude.
Figure 16a compares the positioning results of the SMNS and the Beidou navigation receiver during a small-radius circular flight path at a 3000 m ground-relative altitude. Figure 16b illustrates the positioning error curve of the SMNS during a small-radius circular flight path at a 3000 m ground-relative altitude.
Figure 17a compares the positioning results of the SMNS and the Beidou navigation receiver during a rectangular flight path at a 1500 m ground-relative altitude. Figure 17b illustrates the positioning error curve of the SMNS during a rectangular flight path at a 1500 m ground-relative altitude.
The corresponding analysis results are summarized in Table 3. The following can be observed:
  • When the UAV is flying at a relative altitude of 1500 m, the positioning accuracy of the SMNS is superior to that when the UAV is flying at a relative altitude of 3000 m;
  • Regardless of the flight path of the UAV, the SMNS exhibits significant errors in positioning results. This phenomenon is particularly pronounced when the UAV undergoes changes in its maneuvering state, where the positioning errors become even more noticeable.
Table 3. Statistical analysis of scene-matching location errors under three operational conditions (Condition I: SMNS @3000 m; Condition II: proposed integrated navigation system @3000 m; Condition III: proposed integrated navigation system @1500 m).
Table 3. Statistical analysis of scene-matching location errors under three operational conditions (Condition I: SMNS @3000 m; Condition II: proposed integrated navigation system @3000 m; Condition III: proposed integrated navigation system @1500 m).
NameConditionsMean (m)Standard Deviation (m)
Longitude errorCondition I−3.6361.96
Latitude error31.8048.57
Longitude errorCondition II7.4950.61
Latitude error16.0840.99
Longitude errorCondition III−0.8338.00
Latitude error1.3435.78
As shown in Figure 18, the horizontal position errors of the SMNS exhibit numerous significant gross errors, which could pose a substantial risk to the safe operation of UAVs. To address this issue, we performed data fusion on the positioning results from the SMNS and the SINS. The horizontal position errors of the integrated navigation system after fusion are illustrated in Figure 19.
As shown in Table 4, the standalone SMNS exhibits latitude errors with a mean of −9.48 m and a standard deviation of 54.79 m, while showing longitude errors with a mean of −1.87 m and a standard deviation of 63.52 m.
The SINS/SMNS integrated system demonstrates improved performance:
  • Latitude error: mean = −10.41 m; std = 26.11 m.
  • Longitude error: mean = −2.33 m; std = 34.59 m.
These results indicate a 52.34% reduction in latitude error deviation and a 45.54% reduction in longitude error deviation compared to the standalone SMNS, meeting UAV operational requirements for mission execution, precision navigation, and controlled recovery in GNSS-denied environments.

6. Conclusions

To address the positioning challenges of unmanned aerial vehicles (UAVs) in GNSS-denied environments, we proposed an integrated navigation method based on SINS/SMNS. This approach combines image preprocessing, feature matching, and multi-sensor data fusion to achieve robust navigation. The proposed method was validated through simulations using actual flight data. The results demonstrate that the SINS/SMNS integration significantly outperforms standalone SMNS in horizontal positioning accuracy, improving latitude accuracy by 52.34% and longitude accuracy by 45.54%. The achieved accuracy meets UAV operational requirements in GNSS-denied environments.

Author Contributions

Conceptualization, Y.W. and Q.W.; methodology, Y.W.; software, Y.W.; validation, Y.W.; formal analysis, Y.W.; data curation, Z.H.; writing-original draft, Y.W.; supervision, Q.W.; project administration, P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gura, D.; Rukhlinskiy, V.; Sharov, V.; Bogoyavlenskiy, A. Automated system for dispatching the movement of unmanned aerial vehicles with a distributed survey of flight tasks. J. Intell. Syst. 2021, 30, 728–738. [Google Scholar] [CrossRef]
  2. Santos, N.P.; Rodrigues, V.B.; Pinto, A.B.; Damas, B. Automatic detection of civilian and military personnel in reconnaissance missions using a UAV. In Proceedings of the 2023 IEEE International Conference on Autonomous Robot Systems and Com-petitions (ICARSC), Tomar, Portugal, 26–27 April 2023; pp. 157–162. [Google Scholar]
  3. Dong, Y.; Wang, D.; Zhang, L.; Li, Q.; Wu, J. Tightly coupled GNSS/INS integration with robust sequential Kalman filter for accurate vehicular navigation. Sensors 2020, 20, 561. [Google Scholar] [CrossRef] [PubMed]
  4. Hussain, A.; Akhtar, F.; Khand, Z.H.; Rajput, A.; Shaukat, Z. Complexity and limitations of GNSS signal reception in highly obstructed enviroments. Eng. Technol. Appl. Sci. Res. 2021, 11, 6864–6868. [Google Scholar] [CrossRef]
  5. D’Ippolito, F.; Garraffa, G.; Sferlazza, A.; Zaccarian, L. A hybrid observer for localization from noisy inertial data and sporadic position measurements. Nonlinear Anal. Hybrid Syst. 2023, 49, 101360. [Google Scholar] [CrossRef]
  6. Kinnari, J.; Verdoja, F.; Kyrki, V. GNSS-denied geolocalization of UAVs by visual matching of onboard camera images with orthophotos. In Proceedings of the 2021 20th International Conference on Advanced Robotics (ICAR), Ljubljana, Slovenia, 6–10 December 2021; pp. 555–562. [Google Scholar]
  7. Lu, Z.; Liu, F.; Lin, X. Vision-based localization methods under GPS-denied conditions. arXiv 2022, arXiv:2211.11988. [Google Scholar]
  8. Tong, P.; Yang, X.; Yang, Y.; Liu, W.; Wu, P. Multi-UAV collaborative absolute vision positioning and navigation: A survey and discussion. Drones 2023, 7, 261. [Google Scholar] [CrossRef]
  9. Mei, C.; Fan, Z.; Zhu, Q.; Yang, P.; Hou, Z.; Jin, H. A Novel scene matching navigation system for UAVs based on vision/inertial fusion. IEEE Sens. J. 2023, 23, 6192–6203. [Google Scholar] [CrossRef]
  10. Cao, S.; Lu, X.; Shen, S. GVINS: Tightly coupled GNSS–visual–inertial fusion for smooth and consistent state estimation. IEEE Trans. Robot. 2022, 38, 2004–2021. [Google Scholar] [CrossRef]
  11. Ahmedelbadawi, H.; Żugaj, M. Multi-modal Image Matching for GNSS-denied UAV Localization. In Proceedings of the 1st International Conference on Drones and Unmanned Systems (DAUS‘ 2025), Granada, Spain, 19–21 February 2025; p. 241. [Google Scholar]
  12. Velesaca, H.O.; Bastidas, G.; Rouhani, M.; Sappa, A.D. Multimodal image registration techniques: A comprehensive survey. Multimed. Tools Appl. 2024, 83, 63919–63947. [Google Scholar] [CrossRef]
  13. Fan, J.; Yang, X.; Lu, R.; Li, W.; Huang, Y. Long-term visual tracking algorithm for UAVs based on kernel correlation filtering and SURF features. Vis. Comput. 2023, 39, 319–333. [Google Scholar] [CrossRef]
  14. Liu, Z.; Xu, G.; Xiao, J.; Yang, J.; Wang, Z.; Cheng, S. A real-time registration algorithm of UAV aerial images based on feature matching. J. Imaging 2023, 9, 67. [Google Scholar] [CrossRef] [PubMed]
  15. Luo, X.; Wei, Z.; Jin, Y.; Wang, X.; Lin, P.; Wei, X.; Zhou, W. Fast automatic registration of UAV images via bidirectional matching. Sensors 2023, 23, 8566. [Google Scholar] [CrossRef] [PubMed]
  16. Mohammed, H.M.; El-Sheimy, N. Feature matching enhancement of uav images using geometric con-straints. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 42, 307–314. [Google Scholar] [CrossRef]
  17. Zhang, W.; Zhao, Y. An improved SIFT algorithm for registration between SAR and optical images. Sci. Rep. 2023, 13, 6346. [Google Scholar] [CrossRef]
  18. Deng, Y.; Deng, Y. Two-step matching approach to obtain more control points for SIFT-like very-high-resolution SAR image registration. Sensors 2023, 23, 3739. [Google Scholar] [CrossRef]
  19. Zhang, W. Combination of SIFT and Canny Edge Detection for Registration Between SAR and Optical Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  20. Wu, G.; Zhou, Z. An improved ORB feature extraction and matching algorithm. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021; pp. 7289–7292. [Google Scholar]
  21. Zhang, X.; Wang, Y.; Liu, H. Robust optical and SAR image registration based on OS-SIFT and cascaded sample consensus. IEEE Geosci. Remote. Sens. Lett. 2021, 19, 4007205. [Google Scholar] [CrossRef]
  22. Nehme, E.; Ferdman, B.; Weiss, L.E.; Naor, T.; Freedman, D.; Michaeli, T.; Shechtman, Y. Learning optimal wavefront shaping for multi-channel imaging. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 2179–2192. [Google Scholar] [CrossRef]
  23. Jhan, J.-P.; Rau, J.-Y. A generalized tool for accurate and efficient image registration of UAV multi-lens multispectral cameras by N-SURF matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2021, 14, 6353–6362. [Google Scholar] [CrossRef]
  24. Li, J.; Hu, Q.; Ai, M. RIFT: Multi-Modal Image Matching Based on Radiation-Variation Insensitive Feature Transform. IEEE Trans. Image Process. 2020, 29, 3296–3310. [Google Scholar] [CrossRef]
  25. Yu, K.; Zheng, X.; Duan, Y.; Fang, B.; An, P.; Ma, J. NCFT: Automatic Matching of Multimodal Image Based on Nonlinear Consistent Feature Transform. IEEE Geosci. Remote. Sens. Lett. 2022, 19, 8014105. [Google Scholar] [CrossRef]
  26. Yao, Y.; Zhang, Y.; Wan, Y.; Liu, X.; Yan, X.; Li, J. Multi-Modal Remote Sensing Image Matching Considering Co-Occurrence Filter. IEEE Trans. Image Process. 2022, 31, 2584–2597. [Google Scholar] [CrossRef] [PubMed]
  27. Rouse, D.M.; Hemami, S.S. Understanding and simplifying the structural similarity metric. In Proceedings of the IEEE Interna-tional Conference on Image Processing, San Diego, CA, USA, 12–15 October 2008. [Google Scholar]
  28. Yao, Y.; Zhang, B.; Wan, Y.; Zhang, Y. Motif: Multi-orientation tensor index feature descriptor for sar-optical image registration. ISPRS-Int. Arch. Photogramm. Remote. Sens. Spat. Inf. Sci. 2022, XLIII-B2-2, 99–105. [Google Scholar] [CrossRef]
  29. Yao, Y.; Zhang, Y.; Wan, Y.; Liu, X.; Guo, H. Heterologous images matching considering anisotropic weighted moment and absolute phase orientation. Geomat. Inf. Sci. Wuhan Univ. 2021, 46, 1727–1736. [Google Scholar]
  30. Zhang, Y.; Yao, Y.; Wan, Y.; Liu, W.; Yang, W.; Zheng, Z.; Xiao, R. Histogram of the orientation of the weighted phase descriptor for multi-modal remote sensing image matching. ISPRS J. Photogramm. Remote Sens. 2023, 196, 1–15. [Google Scholar] [CrossRef]
  31. Paul, S.; Pati, U.C. A comprehensive review on remote sensing image registration. Int. J. Remote Sens. 2021, 42, 5396–5432. [Google Scholar] [CrossRef]
  32. Bayoudh, K.; Knani, R.; Hamdaoui, F.; Mtibaa, A. A survey on deep multimodal learning for computer vision: Advances, trends, applications, and datasets. Vis. Comput. 2022, 38, 2939–2970. [Google Scholar] [CrossRef]
  33. Ye, Y.; Shen, L. Hopc: A novel similarity metric based on geometric structural properties for mul-ti-modal remote sensing image matching. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, III-1, 9–16. [Google Scholar]
  34. Zhou, L.; Ye, Y.; Tang, T.; Nan, K.; Qin, Y. Robust matching for SAR and optical images using multiscale convolutional gradient features. IEEE Geosci. Remote. Sens. Lett. 2021, 19, 4017605. [Google Scholar] [CrossRef]
  35. Abdulkadirov, R.; Lyakhov, P.; Butusov, D.; Nagornov, N.; Reznikov, D.; Bobrov, A.; Kalita, D. Enhancing Unmanned Aerial Vehicle Object Detection via Tensor Decompositions and Positive–Negative Momentum Optimizers. Mathematics 2025, 13, 828. [Google Scholar] [CrossRef]
  36. Dranitsyna, E.V.; Sokolov, A.I. Strapdown Inertial Navigation System Accuracy Improvement Methods Based on Inertial Measuring Unit Rotation: Analytical Review. Gyroscopy Navig. 2024, 14, 290–304. [Google Scholar] [CrossRef]
  37. Arrospide, J.; Salgado, L. Log-Gabor filters for image-based vehicle verification. IEEE Trans. Image Process. 2013, 22, 2286–2295. [Google Scholar] [CrossRef] [PubMed]
  38. Yeong, D.J.; Velasco-Hernandez, G.; Barry, J.; Walsh, J. Sensor and sensor fusion technology in autonomous vehicles: A review. Sensors 2021, 21, 2140. [Google Scholar] [CrossRef]
  39. Mei, L.; Wang, C.; Wang, H.; Zhao, Y.; Zhang, J.; Zhao, X. Fast template matching in multi-modal image under pixel distribution mapping. Infrared Phys. Technol. 2022, 127, 104454. [Google Scholar] [CrossRef]
  40. Liu, X.; Li, J.-B.; Pan, J.-S. Feature point matching based on distinct wavelength phase congruency and log-gabor filters in infrared and visible images. Sensors 2019, 19, 4244. [Google Scholar] [CrossRef]
  41. Hisham, M.; Yaakob, S.N.; Raof, R.; Nazren, A.A.; Wafi, N. Template matching using sum of squared difference and normalized cross correlation. In Proceedings of the 2015 IEEE student conference on research and development (SCOReD), Kuala Lumpur, Malaysia, 13–14 December 2015; pp. 100–104. [Google Scholar]
  42. Po, L.M.; Guo, K. Transform-Domain Fast Sum of the Squared Difference Computation for H.264/AVC Rate-Distortion Opti-mization. IEEE Trans. Circuits Syst. Video Technol. 2007, 17, 765–773. [Google Scholar] [CrossRef]
  43. Zhao, F.; Huang, Q.; Gao, W. Image matching by normalized cross-correlation. In Proceedings of the 2006 IEEE international conference on acoustics speech and signal processing proceedings, Toulouse, France, 14–19 May 2006; p. II-II. [Google Scholar]
  44. Barrera, F.; Lumbreras, F.; Sappa, A.D. Multimodal template matching based on gradient and mutual information using scale-space. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 2749–2752. [Google Scholar]
  45. Lu, X.X. A review of solutions for perspective-n-point problem in camera pose estimation. Proc. J. Phys. Conf. Ser. 2018, 1087, 052009. [Google Scholar] [CrossRef]
Figure 2. Automatic orthorectification process with SINS and digital elevation model (DEM).
Figure 2. Automatic orthorectification process with SINS and digital elevation model (DEM).
Sensors 25 03379 g002
Figure 3. The coordinates of the control points (specifically the four corner points and the center point).
Figure 3. The coordinates of the control points (specifically the four corner points and the center point).
Sensors 25 03379 g003
Figure 4. Multi-scale Log-Gabor filtering pipeline for maximum index map generation.
Figure 4. Multi-scale Log-Gabor filtering pipeline for maximum index map generation.
Sensors 25 03379 g004
Figure 5. Template-matching process using sliding window search on satellite reference coordinates.
Figure 5. Template-matching process using sliding window search on satellite reference coordinates.
Sensors 25 03379 g005
Figure 6. Real-time position compensation framework for scene matching.
Figure 6. Real-time position compensation framework for scene matching.
Sensors 25 03379 g006
Figure 7. Integrated navigation system verification platform architecture.
Figure 7. Integrated navigation system verification platform architecture.
Sensors 25 03379 g007
Figure 8. UAV platform and key subsystems: (a) Day–night electro-optical reconnaissance payload integrating a visible light camera (1920 × 1080), an infrared thermal imager (1280 × 1024), and a laser rangefinder (5 m accuracy); (b) A laser strapdown inertial navigation system (LSINS) with ≤0.05°/h gyro bias; (c) A custom high-performance computing module incorporating an NVIDIA Jetson AGX Xavier in a 3D-printed enclosure. The NVIDIA Jetson AGX Xavier is designed and manufactured by NVIDIA (Santa Clara, CA, United States).
Figure 8. UAV platform and key subsystems: (a) Day–night electro-optical reconnaissance payload integrating a visible light camera (1920 × 1080), an infrared thermal imager (1280 × 1024), and a laser rangefinder (5 m accuracy); (b) A laser strapdown inertial navigation system (LSINS) with ≤0.05°/h gyro bias; (c) A custom high-performance computing module incorporating an NVIDIA Jetson AGX Xavier in a 3D-printed enclosure. The NVIDIA Jetson AGX Xavier is designed and manufactured by NVIDIA (Santa Clara, CA, United States).
Sensors 25 03379 g008
Figure 9. First orthorectification result for UAV in right turn (UAV attitude: yaw = 252.772°; pitch = 1.172°; roll = 24.375°). (a) Original image (long-wave infrared image). (b) Orthorectified image (scaled to match reference figure); (c) satellite reference map (satellite images sourced from ArcGIS and stored onboard UAV).
Figure 9. First orthorectification result for UAV in right turn (UAV attitude: yaw = 252.772°; pitch = 1.172°; roll = 24.375°). (a) Original image (long-wave infrared image). (b) Orthorectified image (scaled to match reference figure); (c) satellite reference map (satellite images sourced from ArcGIS and stored onboard UAV).
Sensors 25 03379 g009
Figure 10. Second orthorectification result for UAV in level flight (UAV attitude: yaw = 273.507°; pitch = 2.093°; roll = 0.549°). (a) Original image (long-wave infrared image). (b) Orthorectified image (scaled to match reference figure). (c) Satellite reference map (satellite images sourced from ArcGIS and stored onboard UAV).
Figure 10. Second orthorectification result for UAV in level flight (UAV attitude: yaw = 273.507°; pitch = 2.093°; roll = 0.549°). (a) Original image (long-wave infrared image). (b) Orthorectified image (scaled to match reference figure). (c) Satellite reference map (satellite images sourced from ArcGIS and stored onboard UAV).
Sensors 25 03379 g010
Figure 11. Third orthorectification result for UAV in climbing flight (UAV attitude: yaw = 354.88°; pitch = 9.116°; roll = 0.397°). (a) Original image (long-wave infrared image). (b) Orthorectified image (scaled to match reference figure). (c) Satellite reference map (satellite images sourced from ArcGIS and stored onboard UAV).
Figure 11. Third orthorectification result for UAV in climbing flight (UAV attitude: yaw = 354.88°; pitch = 9.116°; roll = 0.397°). (a) Original image (long-wave infrared image). (b) Orthorectified image (scaled to match reference figure). (c) Satellite reference map (satellite images sourced from ArcGIS and stored onboard UAV).
Sensors 25 03379 g011
Figure 12. Image matching and mosaic results for UAV in roll flight.
Figure 12. Image matching and mosaic results for UAV in roll flight.
Sensors 25 03379 g012
Figure 13. Image matching and mosaic results for UAV in level flight.
Figure 13. Image matching and mosaic results for UAV in level flight.
Sensors 25 03379 g013
Figure 14. Image matching and mosaic results for UAV in climbing flight.
Figure 14. Image matching and mosaic results for UAV in climbing flight.
Sensors 25 03379 g014
Figure 15. Comparison of positioning results between the SMNS and the Beidou navigation receiver during a rectangular flight path at a 3000 m relative altitude. (a) SMNS and Beidou positioning results. (b) Longitude error and latitude error.
Figure 15. Comparison of positioning results between the SMNS and the Beidou navigation receiver during a rectangular flight path at a 3000 m relative altitude. (a) SMNS and Beidou positioning results. (b) Longitude error and latitude error.
Sensors 25 03379 g015
Figure 16. Comparison of positioning results between the SMNS and the Beidou navigation receiver during a small-radius circular flight path at a 3000 m relative altitude. (a) SMNS and Beidou positioning results. (b) Longitude error and latitude error.
Figure 16. Comparison of positioning results between the SMNS and the Beidou navigation receiver during a small-radius circular flight path at a 3000 m relative altitude. (a) SMNS and Beidou positioning results. (b) Longitude error and latitude error.
Sensors 25 03379 g016
Figure 17. Comparison of positioning results between the SMNS and the Beidou navigation receiver during a rectangular flight path at a 1500 m relative altitude. (a) SMNS and Beidou positioning results. (b) Longitude error and latitude error.
Figure 17. Comparison of positioning results between the SMNS and the Beidou navigation receiver during a rectangular flight path at a 1500 m relative altitude. (a) SMNS and Beidou positioning results. (b) Longitude error and latitude error.
Sensors 25 03379 g017
Figure 18. Horizontal position errors of the SMNS. (a) Latitude error. (b) Longitude error.
Figure 18. Horizontal position errors of the SMNS. (a) Latitude error. (b) Longitude error.
Sensors 25 03379 g018
Figure 19. Horizontal position errors of the SINS/SMNS integrated navigation system. (a) Latitude error. (b) Longitude error.
Figure 19. Horizontal position errors of the SINS/SMNS integrated navigation system. (a) Latitude error. (b) Longitude error.
Sensors 25 03379 g019
Table 1. Integrated navigation system verification platform: experimental equipment specifications and performance metrics (components: laser inertial navigation system, Beidou navigation receiver, atmospheric data sensor system, day–night electro-optical reconnaissance payload, and high-performance processor module).
Table 1. Integrated navigation system verification platform: experimental equipment specifications and performance metrics (components: laser inertial navigation system, Beidou navigation receiver, atmospheric data sensor system, day–night electro-optical reconnaissance payload, and high-performance processor module).
Device NameMetric NameParameter
Laser inertial
navigation system
Gyroscope bias stability≤0.05°/h
Accelerometer bias stability≤100 μg
Beidou navigation
receiver
Velocity accuracy≤0.05 m/s
Position accuracy≤0.1 m
Atmospheric data
sensor system
Airspeed≤1 m/s
Barometric altitude≤10 m
Day–night electro-optical
reconnaissance payload
Visible light resolution1920 × 1080
Infrared resolution1280 × 1024
Laser ranging accuracy5 m
High-performance
processor module
GPUVolta architecture with 512 CUDA cores
CPU8-core Carmel Armv8.2 64-bit CPU 32 GB
RAM256-bit LPDDR4
External storage1 TB SSD
Table 2. Statistical results of actual processing time for template matching, Log-Gabor filtering, orthorectification, and KF algorithms on a high-performance processor across different ROI sizes.
Table 2. Statistical results of actual processing time for template matching, Log-Gabor filtering, orthorectification, and KF algorithms on a high-performance processor across different ROI sizes.
ROI Size (Pixel)Template Matching (ms)Log-Gabor Filtering (ms)Orthorectification (ms)KF (ms)
2000 × 200067.3334.623.960.02638
3000 × 3000126.5354.264.060.02624
3800 × 3800209.612383.514.070.02736
5400 × 5400472.39207.884.080.02675
Table 4. Statistical analysis results of the positioning errors for the integrated navigation system based on the SINS/SMNS.
Table 4. Statistical analysis results of the positioning errors for the integrated navigation system based on the SINS/SMNS.
MethodsError NameMean (m)Standard Deviation (m)
SMNSLatitude error−9.4854.79
Longitude error−1.8763.52
SINS/SMNSLatitude error−10.4126.11
Longitude error−2.3334.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Wang, Q.; Hao, Z.; Chen, P. An Integrated Navigation Method Based on the Strapdown Inertial Navigation System/Scene-Matching Navigation System for UAVs. Sensors 2025, 25, 3379. https://doi.org/10.3390/s25113379

AMA Style

Wang Y, Wang Q, Hao Z, Chen P. An Integrated Navigation Method Based on the Strapdown Inertial Navigation System/Scene-Matching Navigation System for UAVs. Sensors. 2025; 25(11):3379. https://doi.org/10.3390/s25113379

Chicago/Turabian Style

Wang, Yukun, Qiang Wang, Zhonghu Hao, and Puhua Chen. 2025. "An Integrated Navigation Method Based on the Strapdown Inertial Navigation System/Scene-Matching Navigation System for UAVs" Sensors 25, no. 11: 3379. https://doi.org/10.3390/s25113379

APA Style

Wang, Y., Wang, Q., Hao, Z., & Chen, P. (2025). An Integrated Navigation Method Based on the Strapdown Inertial Navigation System/Scene-Matching Navigation System for UAVs. Sensors, 25(11), 3379. https://doi.org/10.3390/s25113379

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop