Next Article in Journal
Biocompound and Lake Pigment Extraction from Invasive Alien Plant Biomass for Sustainable Ink Applications
Previous Article in Journal
Validation of a Novel Variable-Cam System: Electromyographic and Kinetic Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Sensor-Assisted Navigation for UAVs in Power Inspection: A Fusion Approach Using LiDAR, IMU and GPS

1
Dali Bureau, EHV Power Transmission Company of China Southern Power Grid, Dali 671014, China
2
Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(6), 2632; https://doi.org/10.3390/app16062632
Submission received: 5 February 2026 / Revised: 4 March 2026 / Accepted: 4 March 2026 / Published: 10 March 2026

Abstract

High-precision localization is essential for autonomous navigation and environment perception of unmanned aerial vehicles (UAVs) in complex power inspection scenarios. To overcome the limited accuracy and accumulated drift of conventional GPS-based single-sensor localization, this paper proposes a LiDAR–IMU–GPS-aided navigation method that combines a tightly coupled front-end and a loosely coupled back-end. The front-end employs an improved Lie-group-based UKF-SLAM framework to explicitly handle the nonlinearities of rotational motion, thereby improving the stability of local pose estimation. The back-end integrates GPS absolute constraints, loop closure detection, and point cloud registration via pose graph optimization, which effectively suppresses long-term accumulated drift. The framework achieves accurate and robust localization for UAV power inspection. Experiments on public benchmark datasets and real-world power inspection scenarios demonstrate the effectiveness of the proposed method. On the MH_02_easy sequence, the absolute trajectory error is reduced from 0.521 m to 0.170 m compared with ROVIO, while in a real inspection sequence the cumulative error is reduced by more than 99% after back-end optimization. Moreover, the system maintains stable navigation under GPS-degraded conditions, indicating strong robustness and practical applicability.

1. Introduction

High-precision localization is one of the key enabling technologies for autonomous navigation and environment perception of unmanned aerial vehicles (UAVs) in complex environments [1,2,3,4]. In industrial applications such as power line inspection, more stringent requirements are imposed on the robustness, accuracy, and real-time performance of localization systems [5]. Conventional UAV localization solutions mainly rely on the Global Positioning System (GPS) [6]. However, in environments characterized by complex terrain, urban canyons, or severe multipath effects, civilian-grade GPS typically provides positioning accuracy in the range of 2–5 m, and its signals are prone to blockage and interference, making it difficult to meet the centimeter- to decimeter-level accuracy demands of power inspection tasks.
Moreover, localization systems based on a single sensor modality inherently suffer from performance limitations. For example, inertial measurement units (IMUs) are susceptible to error accumulation during long-term operation, while LiDAR-based localization in large-scale or long-range scenarios may experience feature degradation or matching failures, leading to significant localization drift [7,8,9].
In recent years, multi-sensor fusion has emerged as an effective approach to improving the localization accuracy of UAVs [6,10]. By integrating the complementary strengths of multiple sensors such as LiDAR, IMU, and GPS, it becomes possible to achieve continuous and stable high-precision localization even in environments where GPS signals are degraded or unavailable. Among these approaches, simultaneous localization and mapping (SLAM) has been widely recognized as a core methodology for autonomous navigation in unmanned systems and has been extensively applied in fields such as autonomous driving and robotic navigation.
According to different back-end processing paradigms, SLAM methods can generally be categorized into filtering-based approaches and optimization-based approaches. The Extended Kalman Filter (EKF) has been widely adopted in early systems due to its high computational efficiency; however, its reliance on linearization assumptions may introduce significant errors in nonlinear motion models and can even lead to system divergence. In contrast, the Unscented Kalman Filter (UKF) approximates nonlinear distributions through sigma-point sampling, thereby avoiding linearization errors and making it more suitable for handling nonlinearities in attitude estimation [11,12].
In power inspection missions, UAVs are often required to operate in environments with complex structures and strong electromagnetic interference, such as high-voltage transmission lines and substations, which poses significant challenges to the robustness and environmental adaptability of localization systems [13]. Existing studies have shown that loosely coupled LiDAR–IMU approaches can achieve relatively stable localization performance; however, many of these systems do not sufficiently incorporate absolute positioning information from GPS, making it difficult to effectively suppress accumulated drift during long-duration missions [14,15].
Furthermore, conventional EKF-based fusion methods are prone to model inconsistency in rotational estimation due to the strong nonlinearity of quaternion representations, which may lead to degradation in estimation accuracy [16,17]. In contrast, UKF methods formulated on Lie groups provide a more principled treatment of three-dimensional rotational kinematics, thereby improving the consistency and reliability of attitude estimation. Additionally, the proposed UKF-SLAM framework improves computational efficiency and robustness by directly handling rotational motion on the SE(3) manifold, avoiding the issues of quaternion-based methods. This formulation enables more accurate state estimation in highly nonlinear environments, such as power inspection scenarios, where accurate pose and motion tracking are critical for UAV navigation.
To address the localization requirements of UAV-based power inspection, this paper proposes a high-precision aided navigation system based on loosely coupled multi-sensor fusion. The system adopts a three-sensor integration architecture comprising LiDAR, IMU, and GPS. In the front-end, loosely coupled LiDAR–IMU fusion is employed to achieve locally accurate pose estimation, while in the back-end, a pose graph optimization framework is introduced to construct a globally consistent trajectory by jointly incorporating GPS absolute position constraints and loop closure information.
At the algorithmic level, this paper develops an improved UKF-SLAM approach formulated on Lie group operations. By modeling rotational kinematics on Lie groups, the proposed method avoids linearization errors associated with quaternion representations and enhances the stability of state estimation in dynamic environments.
It is worthwhile to highlight several aspects of the contributions here:
  • We propose a navigation architecture for power inspection that combines a tightly coupled LiDAR-IMU front-end (UKF-SLAM on Lie groups) with a loosely coupled back-end (GPS and loop closure pose graph optimization), achieving both high local accuracy and global consistency;
  • We design a UKF-SLAM algorithm formulated on Lie group operations to enhance the mathematical rigor and numerical stability of rotational estimation in the front-end;
  • We construct a pose graph optimization model incorporating GPS constraints, loop closure detection, and point cloud registration, which effectively suppresses long-term accumulated drift in a loosely coupled manner.
This paper is organized as follows. Section 2 presents an aided localization method based on UKF filtering for LiDAR–IMU fusion. Section 3 describes an aided localization approach based on pose graph optimization for LiDAR–GPS fusion. Section 4 details the experimental setup and provides a comprehensive analysis of the results. Finally, Section 5 concludes the paper.

2. An Aided Localization Method Based on UKF Filtering for LiDAR–IMU Fusion

In this section, we present the tightly coupled LiDAR-IMU fusion front-end based on an improved UKF-SLAM formulated on Lie groups. This front-end processes raw IMU and LiDAR data jointly to provide high-frequency, locally consistent pose estimates. The Extended Kalman Filter (EKF) performs well in many motion estimation tasks; however, its reliance on linearization can introduce significant approximation errors, which may lead to system divergence in severe cases [18]. In contrast, the Unscented Kalman Filter (UKF) approximates the state distribution through a set of sampled sigma points, thereby avoiding direct linearization of nonlinear models.
Moreover, rotational motion is commonly represented using quaternions, which further increases the nonlinearity of the system. To address this issue, Lie group operations can be employed to map quaternion representations to three-dimensional vector spaces, thereby improving the accuracy and stability of state estimation.

2.1. Fusion Algorithm Design Workflow

2.1.1. Introduction to Data Fusion Strategy

Compared with single-sensor systems, the fusion of heterogeneous sensors can provide richer information representation, higher accuracy, improved measurement precision, and more comprehensive system performance. By evaluating and comparing the performance of single-sensor systems and multi-sensor fusion systems across different application scenarios, it can be clearly observed that multi-sensor fusion approaches exhibit overall significant advantages. The corresponding comparative results are illustrated in the Figure 1 below.
Front-end information fusion methods can be categorized into four types: information fusion strategies based on distinguishable units, fusion strategies based on feature complementarity, fusion strategies based on target attributes from different sensors, and fusion strategies based on decision-level outputs from different sensors. A detailed analysis of these strategies is presented below.
(1)
Information Fusion Strategy Based on Distinguishable Units
FSBDU refers to a fusion process in which data from distinguishable units of different sensors are directly integrated, followed by further processing of the fused data.
(2)
Information Fusion Strategy Based on Feature Complementarity
FSBCF integrates multiple target features extracted from the corresponding sensor data. Heterogeneous sensors capture complementary features of the same target at different and often uncorrelated scales, which are then fused and utilized for subsequent classification and recognition tasks.
(3)
Fusion Strategy Based on Target Attributes from Different Sensors
FSBTA is a distributed data processing approach in which each sensor extracts target parameters and identifies targets independently to generate a target list. Multiple target lists are then fused to obtain reliable and accurate target information, thereby reducing false alarms and missed detections.
(4)
Fusion Strategy Based on Decisions from Different Sensors
FSBMD performs preliminary decisions on target location, attributes, and classification using data from individual sensors, and then comprehensively combines the decisions obtained from multiple sensors through dedicated fusion strategies to produce the final fused result. FSBMD can generally be categorized into decision fusion, decision making, confidence-based fusion, and probabilistic fusion.

2.1.2. Study on Alignment-Based Fusion Strategies

(1)
Temporal Alignment Algorithm
To address the inconsistency in output frequencies among sensors such as LiDAR (10 Hz), IMU (100 Hz), and GPS (1 Hz), an appropriate time synchronization scheme must be designed to enable effective multi-source data fusion [19,20,21]. Because errors in time alignment can lead to inaccurate position estimates. When the timestamps of different sensors are not synchronized, the sensor measurements may no longer correspond to the state at the same moment. For example, LiDAR and IMU may observe the same object at the same time, but due to differences in sampling times, the data points used in the fusion algorithm may not match spatially, which in turn affects the final state estimation [22]. In this study, a timestamp-based sorting method is adopted, as shown in Figure 2 below. This approach primarily utilizes software-level interpolation to perform temporal alignment and synchronization of sensor measurements with different sampling rates. This strategy ensures temporal consistency across data streams and provides reliable inputs for subsequent fusion algorithms.
(2)
Spatial Alignment Algorithm
The proposed method first estimates an initial pose state of the system motion and defines a GPS-based error bound, which is then used as a robust kernel to suppress noisy observations, thereby obtaining an initial fused pose (Initial Pose). Subsequently, trajectory interpolation information generated from IMU pre-integration is incorporated, and the system pose state is further refined through a loosely coupled optimization framework. Finally, high-precision pose estimates are produced as the measurement output. The relevant process is shown in Figure 3 below.

2.2. Optimization Process Based on UKF-SLAM

2.2.1. State Vector and System Model

Assuming that the IMU biases follow a random walk model and considering that the LiDAR has tracked n landmark points, the system state vector can be defined as follows:
χ = R T , υ T , x T , b g T , b a T , p 1 T ,   , p n T T
Here, R S O ( 3 ) , x R 3 , and υ R 3 denote the orientation, position, and velocity of the platform, respectively. b g R 3 and b a R 3 represent the biases of the gyroscope (angular velocity) and the accelerometer, respectively. p 1   p n R 3 denotes the three-dimensional positions of the n LiDAR-tracked landmark points expressed in the global coordinate frame.
Based on the IMU measurements, the continuous-time kinematic model of the system can be expressed as follows:
R ˙ = R w b g + η g
υ ˙ = R a b a + η a + g
x ˙ = υ
b g ˙ = η b g , b a ˙ = η b a
p i ˙ = 0 , i = 1 , , n
The angular velocity w and acceleration a measured by the IMU both contain bias terms and additive Gaussian white noise:
η B = η g T , η a T , η b g T , η b a T T
The observations obtained by the LiDAR consist of a set of tracked landmark points p i . Based on the standard perspective projection model, the observation of the i -th landmark, denoted as y i , can be expressed as follows:
y i = 1 ω i u i υ i + n y i
where ( u i , υ i , ω i ) T is computed as follows:
u i υ i ω i = R C B T R W B T p i x W B x C B
In the above formulation, R C B and x C B denote the rotation and translation from the camera coordinate frame C to the body frame B , respectively, while R W B and x W B represent the corresponding transformation from the world coordinate frame W to the body frame B . The observation noise n y i is assumed to follow a zero-mean Gaussian distribution.

2.2.2. UKF-SLAM Based on Lie Group Operations

In this section, a UKF-SLAM system is constructed based on the aforementioned formulations.
(1)
Unscented Kalman Filter (UKF)
The Unscented Kalman Filter (UKF) is a state estimation method designed for nonlinear systems. It approximates the probability distribution of state variables using a set of deterministically sampled sigma points, thereby avoiding explicit linearization of nonlinear models [11,23].
These sigma points are propagated directly through the nonlinear motion equations, and the mean and covariance of the state estimate are reconstructed via weighted summation. This approach maintains estimation accuracy while reducing errors introduced by linearization.
To accommodate the state estimation requirements of the Unscented Kalman Filter (UKF) in nonlinear systems, an augmented state vector is first defined. Let the original state vector be denoted as χ R S ; the augmented state vector and its corresponding covariance matrix can then be constructed as follows:
χ a = χ T , η B T T , P a = P S × S 0 M × 12 0 12 × M Q 12 × 12 N × N
where η B R 12 denotes the system noise vector, and Q is the noise covariance matrix, whose structure is given by:
Q = Q g 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 Q a 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 Q b g 0 3 × 3 0 3 × 3 0 3 × 3 0 3 × 3 Q b a 12 × 12
The dimension of the augmented state is N = S + 12 . Based on this, a symmetric sampling strategy is employed to generate 2 N + 1 sigma points:
σ 0 = χ a
σ i = χ a + λ + N P a i , i = 1 , 2 , , N
σ i = χ a λ + N P a i , i = N + 1 , , 2 N
Here, λ denotes the scaling parameter, and P a represents the matrix square root obtained via Cholesky decomposition. The notation · i indicates the extraction of the i -th column vector of the matrix.
Within the UKF framework, each sigma point is propagated forward through the system model function f . Based on the propagated sigma points, the a priori state estimate and the corresponding covariance matrix can be computed as follows:
χ ^ k | k 1 = i = 0 2 N w i f σ i , P ^ k | k 1 = i = 0 2 N w i f σ i χ ^ k | k 1 f σ i χ ^ k | k 1 T
where the weighting coefficients are defined as follows:
w 0 = λ λ + N , w i = λ 2 λ + N i = 1 , , 2 N
The scaling parameter is given by λ = α 2 ( N + β ) N . Typically, α is set to 10 3 , and β = 2 is chosen when the state is assumed to follow a Gaussian distribution.
Similarly, the predicted measurement and the corresponding covariance can be updated through the measurement model function h as follows:
z ^ k = i = 0 2 N w i h σ i
P x z = i = 0 2 N w i ( f ( σ i ) χ ^ k | k 1 ) h σ i z ^ k T
P z z = i = 0 2 N w i ( h ( σ i ) z ^ k ) h σ i z ^ k T
Based on the previously obtained cross-covariance matrix P x z and the predicted measurement covariance matrix P z , the Kalman gain matrix can be computed as follows:
K k = P x z P z 1
Subsequently, by combining the actual measurement z k with the predicted measurement z ^ k , the posterior state estimate can be updated as follows:
χ k = χ ^ k | k 1 + K k z k z ^ k
The corresponding posterior covariance matrix is updated as follows:
P k = P ^ k | k 1 K k P z K k T
In this formulation, χ ^ k k 1 and P ^ k k 1 denote the a priori state estimate and its corresponding covariance matrix, respectively. This update process effectively incorporates measurement information, thereby yielding an optimal estimate of the system state.
(2)
Improved UKF with Lie Group Operations
A Lie group is a class of matrix groups with continuous structure, whose mathematical properties are well suited for describing geometric transformations such as rotations and rigid-body motions [24,25]. An n -dimensional Lie group G satisfies the following fundamental properties:
  • Closure: G 1 , G 2 G , G 1 G 2 G ;
  • Associativity: G 1 , G 2 , G 3 G , ( G 1 G 2 ) G 3 = G 1 ( G 2 G 3 ) ;
  • Existence of Inverse Elements: G 1 G , G 1 1 G .
In state estimation, commonly used Lie groups include:
  • The three-dimensional rotation group S O ( 3 ) , which represents rotation matrices R ;
  • The three-dimensional special Euclidean group S E ( 3 ) which represents rigid-body transformation matrices:
    T = R p 0 T 1 S E 3
    where p R 3 denotes the translation vector.
Each Lie group G is associated with a corresponding Lie algebra g , which is a vector space tangent to the identity element. The Lie group and its Lie algebra are related through the exponential map e x p ( ) : g G and the logarithmic map l o g ( ) : G g . For the rotation group S O ( 3 ) , its Lie algebra s o ( 3 ) consists of all three-dimensional skew-symmetric matrices. Specifically, if ϕ R 3 denotes a rotation vector, then:
ϕ = l o g ( R ) , ϕ = l o g ( R ) , R = e x p ( ϕ )
Here, the operator · maps a three-dimensional vector to a skew-symmetric matrix, while · denotes the corresponding inverse mapping.
To extend the Unscented Kalman Filter (UKF) to Lie group spaces, it is necessary to redefine a Gaussian distribution that is compatible with manifolds. A Gaussian distribution defined on a Lie group G can be expressed as follows:
Θ X = α e x p 1 2 l o g G X υ T P 1 l o g G x
where l o g G ( X v ) g denotes the vector in the Lie algebra associated with the Lie group element X v , P is the covariance matrix, and α is a normalization constant.
This distribution is associated with the Lie group through the exponential map as follows:
X = μ e x p G ( ε ^ )
Here, μ G denotes the mean on the Lie group, and ε ^ g represents a zero-mean Gaussian perturbation in the Lie algebra. Based on this formulation, the prediction and update processes of the Kalman filter can be extended to the S E ( 3 ) manifold.
Specifically, the state propagation equation on the S E ( 3 ) manifold is given by:
χ S E ( 3 ) | k = T C W e x p S E 3 R W C χ S E ( 3 ) | k 1
Here, T C W S E ( 3 ) denotes the transformation between coordinate frames, and R C W S O ( 3 ) represents the rotational component. Correspondingly, sigma point sampling on the S E ( 3 ) manifold is performed as follows:
σ S E 3 0 = χ S E 3
σ S E 3 i = χ S E 3 e x p S E 3 λ + N P a i , i = 1 , 2 , , N
σ S E 3 i = χ S E 3 e x p S E 3 λ + N P a i , i = N + 1 , , 2 N
In the above formulation, P a denotes the augmented covariance matrix, and λ is the scaling parameter. The sampled sigma points are propagated through the system model; however, it should be noted that on the S E ( 3 ) manifold, the computation of the mean is no longer a linear weighted sum. Instead, iterative procedures or Riemannian centroid methods must be employed to compute the mean and adjust the covariance on the manifold, in order to preserve the consistency of the estimation.
On the S E ( 3 ) manifold, the a priori mean and covariance must be obtained through statistical operations defined on the manifold. The mean μ ^ S E ( 3 ) is defined as the solution to a minimization problem of the weighted sum of squared geodesic distances:
μ ^ S E 3 = argmin ρ S E 3 i = 0 2 N w i d 2 f σ S E 3 i , ρ
where d ( ) denotes the geodesic distance on S E ( 3 ) . The covariance matrix is then computed in the Lie algebra space via the logarithmic map as follows:
P ^ S E 3 = i = 0 2 N w i l o g μ S E 3 f σ S E 3 i l o g μ S E 3 f σ S E 3 i T
Here, l o g μ ( X ) denotes the tangent-space vector that maps the Lie group element X to the tangent space at μ .
During the measurement update stage, the predicted measurement mean z ^ S E 3 is obtained in an analogous manner:
z ^ S E 3 = argmin ρ S E 3 i = 0 2 N w i d 2 h σ S E 3 i , ρ
Subsequently, the cross-covariance matrix between the state and the observation, as well as the observation covariance matrix, are given by:
P x z | S E 3 = i = 0 2 N w i l o g μ ^ S E 3 f σ S B 3 i l o g z ^ S B 3 h σ S E 3 i T
P z z | S E 3 = i = 0 2 N w i l o g z ^ S B 3 h σ S E 3 i l o g z ^ S E 3 h σ S E 3 i T
Based on the above matrices, the Kalman gain on the manifold can be computed, and the state and covariance are subsequently updated as follows:
K k | S E ( 3 ) = e x p S E 3 P x z | S E 3 P z z | S E 3 1
χ k | S E ( 3 ) = μ ^ S E 3 + e x p S E 3 K k | S E ( 3 ) l o g z ^ S E 3 z k | S E 3
P k | S E ( 3 ) = P ^ S E 3 + e x p S E 3 K k | S E ( 3 ) P z z | S E 3 K k | S E ( 3 ) T
where z denotes the actual measurement.
In summary, the workflow of the improved UKF algorithm based on Lie group operations can be summarized as follows:
  • Construct the augmented state on the S E ( 3 ) manifold and generate sigma points;
  • Propagate the sigma points through the nonlinear system equations, and compute the a priori estimate using geodesic mean and covariance calculations;
  • Predict the measurements using the observation model, and compute the cross-covariance and the observation covariance;
  • Fuse the measurement residuals using the Kalman gain to update the posterior state and covariance;
  • Iteratively execute the above steps to achieve continuous state estimation of the dynamic system on the manifold.

3. An Aided Localization Study Based on Pose Graph Optimization for LiDAR–GPS Fusion

This section describes the loosely coupled back-end optimization module, which refines the front-end poses by incorporating GPS measurements and loop closure constraints through pose graph optimization.

3.1. Construction of the Pose Graph Optimization Model

The pose state variable at an arbitrary time i is defined as ξ i = [ t i T , ϕ i T ] T = x i , y i , z i , ϕ 1 i , ϕ 2 i , ϕ 3 i T , where t i = [ x i , y i , z i ] T R 3 denotes the position in the Cartesian coordinate system, and ϕ i = [ ϕ 1 i , ϕ 2 i , ϕ 3 i ] T R 3 represents the Lie algebra parameters corresponding to the rotation matrix.
Based on this state representation, GPS measurements, loop closure detection results, and point cloud registration constraints can be jointly incorporated to construct various geometric constraints within the pose graph optimization model.

3.1.1. Construction of GPS Coordinate Constraints

The raw GPS data are first preprocessed by sequentially transforming them from the geodetic coordinate system to the Earth-Centered Earth-Fixed (ECEF) coordinate system, and then further into the East–North–Up (ENU) coordinate system [26,27,28]. After coordinate transformation, linear interpolation is applied to temporally align the GPS data with the LiDAR-synchronized timestamps. Let the GPS coordinate observation at time instant i after temporal synchronization be denoted as:
t g i = t g x i , t g y i , t g z i T
The GPS residual term at time instant i is defined as the difference between the estimated position t i and the GPS-observed position t g i :
r g p s i = t i t g i
This residual depends only on the translational component of the state. Taking the derivative with respect to the state variable ξ i = [ t i T , ϕ i T ] T yields the Jacobian matrix:
r g p s i T ξ i = t i t g i t i ϕ i = I 3 × 3 0 3 × 3 3 × 6
where I 3 × 3 denotes the identity matrix. Based on this residual, the objective function corresponding to the GPS constraint can be formulated as follows:
F g p s ( ξ i ) = r g p s i 2 2

3.1.2. Construction of Loop Closure Constraints

Using a GPS-assisted loop closure detection method, candidate loop closure node pairs can be identified between a historical time instant k and the current time instant i (with k < i ). Let T k , T i S E ( 3 ) denote the pose transformation matrices at time instants k and i , respectively, and let the corresponding state variables be ξ k = [ t k T , ϕ k T ] T and ξ i = [ t i T , ϕ i T ] T .
The relationship between the pose transformation matrices and the state variables can be expressed as follows:
T k = R k t k 0 1 ξ k = t k , ϕ k T = x k , y k , z k , ϕ 1 k , ϕ 2 k , ϕ 3 k T T i = R i t i 0 1 ξ i = t i , ϕ i T = x i , y i , z i , ϕ 1 i , ϕ 2 i , ϕ 3 i T
where R k , R i S O ( 3 ) are the rotation matrices, t k , t i R 3 denote the translation vectors, and ϕ k , ϕ i R 3 represent the corresponding Lie algebra coordinates.
When the loop closure detection condition is satisfied, the poses T k and T i can be regarded as corresponding to the same spatial location. Accordingly, the loop closure residual term is constructed as follows:
r l o o p k , i = l n R k 1 R i t k t i 6 × 1
In this formulation, l n ( ) : S O ( 3 ) s o ( 3 ) denotes the logarithmic map of the Lie group, and ( ) : s o ( 3 ) R 3 maps a skew-symmetric matrix to a three-dimensional vector. This residual jointly characterizes the consistency constraints of both rotational and translational components in the loop closure.
According to the correspondence between Lie algebra elements and rotation parameters, the Lie algebra can be parameterized as ϕ = θ a , where a R 3 is a unit rotation axis vector and θ R denotes the rotation angle (in radians) about this axis. Under this parameterization, the left Jacobian matrix associated with the Baker–Campbell–Hausdorff (BCH) formula on the Lie group S O ( 3 ) can be expressed as:
J l ϕ = s i n θ θ I + 1 s i n θ θ a a T + 1 c o s θ θ a
Here, a denotes the skew-symmetric matrix corresponding to the vector a . This Jacobian matrix is subsequently employed in pose graph optimization to perform linearization of the rotational residuals.
b.
By differentiating the residual equation, the Jacobian matrix J l o o p i can be derived as follows:
r l o o p k , i T ξ i = l n R k 1 R i t k t i t i ϕ i = l n R k 1 R i t i l n R k 1 R i ϕ i t k t i t i t k t i ϕ i = 0 3 × 3 l n R k 1 R i ϕ i I 3 × 3 0 3 × 3
The computation of ln R k 1 R i ϕ i is given as follows:
ln R k 1 R i ϕ i = lim ϕ 0 ln R k 1 R i exp ϕ i ln R k 1 R i ϕ i = = lim ϕ 0 ln R k 1 R i + J r 1 ln R k 1 R i ϕ i ln R k 1 R i ϕ i = = J r 1 ln R k 1 R i = J l 1 ln R k 1 R i
Combining the above two expressions, we obtain:
J l o o p i = r l o o p k , i T ξ i = 0 3 × 3 J l 1 ln R k R i I 3 × 3 0 3 × 3 6 × 6
b.
By differentiating the residual equation, the Jacobian matrix J l o o p k can be derived as follows:
r l o o p k , i T ξ k = ln R k 1 R i t k t i t k ϕ k = ln R k 1 R i t k ln R k 1 R i ϕ k t k t i t k t k t i ϕ k = 0 3 × 3 l n R k 1 R i ϕ k I 3 × 3 0 3 × 3
The computation of ln R k 1 R i ϕ k is given as follows:
ln R k 1 R i ϕ k = lim ϕ 0 ln R k 1 exp ϕ k R i ln R k 1 R i ϕ k = lim ϕ 0 ln R k 1 R i exp R i T ϕ k ln R k 1 R i ϕ k = J l 1 ln R k 1 R i R i T = J l 1 ln R k 1 R i R i T
Combining the above two expressions, we obtain:
J l o o p k = r l o o p k , i T ξ k = 0 3 × 3 J l 1 ln R k 1 R i R i T I 3 × 3 0 3 × 3 6 × 6

3.1.3. Construction of Point Cloud Registration Constraints

Based on a vision-enhanced point cloud registration algorithm, the relative pose transformation between two consecutive time instants i 1 and i (with i 1 ) can be obtained. Let the pose transformation matrices at these two time instants be denoted as T i 1 , T i S E ( 3 ) , and let their corresponding Lie algebra coordinates be ξ i 1 , ξ i R 6 . According to the definition of point cloud registration, the relative pose transformation can be expressed as T i 1 . i = T i 1 1 T i .
Expanding the pose transformation matrices and the corresponding state variables yields:
T i 1 = R i 1 t i 1 0 1 ,     T i = R i t i 0 1 ,     T i 1 , i = R i 1 , i t i 1 , i 0 1 ξ i 1 = t i 1 , ϕ i 1 T , ξ i = t i , ϕ i T
Therefore, expanding T i 1 . i = T i 1 1 T i yields:
R i 1 , i t i 1 , i 0 1 = R i 1 1 R i 1 1 t i 1 0 1 R i t i 0 1 = R i 1 1 R i R i 1 t i t i 1 0 1
However, due to the presence of errors, the above relation does not hold exactly. Accordingly, the registration residual r r e g i s t r a t i o n i 1 , i is constructed as follows:
r r e g i s t r a t i o n i 1 , i = ln R i 1 , i 1 R i 1 1 R i t i t i 1 R i 1 t i 1 , i 6 × 1
  • By differentiating the residual equation, the Jacobian matrix J r e g i s t r a t i o n i 1 can be derived as follows:
    r registration i 1 , i T ξ i 1 = ln R i 1 , i 1 R i 1 1 R i t i t i 1 R i 1 t i 1 , i 6 × 1 t i 1 ϕ i 1 = ln R i 1 , i 1 R i 1 1 R i t i 1 ln R i 1 , i 1 R i 1 1 R i ϕ i 1 t i t i 1 R i 1 t i 1 , i t i 1 t i t i 1 R i 1 t i 1 , i ϕ i 1 = 0 3 × 3 ln R i 1 , i 1 R i 1 1 R i ϕ i 1 I 3 × 3 R i 1 t i 1 , i
The computation of ln R i 1 , i 1 R i 1 1 R i ϕ i 1 is given as follows:
ln R i 1 , i 1 R i 1 1 R i ϕ i 1 = lim ϕ 0 ln R i 1 , i 1 R i 1 1 exp ϕ i 1 R i ln R i 1 , i 1 R i 1 1 R i ϕ i 1 = lim ϕ 0 ln R i 1 , i 1 R i 1 1 R i exp R i T ϕ i 1 ln R i 1 , i 1 R i 1 1 R i ϕ i 1 = lim ϕ 0 ln R i 1 , i 1 R i 1 1 R i exp R i T ϕ i 1 ln R i 1 , i 1 R i 1 1 R i ϕ i 1 = J r 1 ln R i 1 , i 1 R i 1 1 R i R i T = J l 1 ln R i 1 , i 1 R i 1 1 R i R i T
Accordingly, the Jacobian matrix J r e g i s t r a t i o n i 1 is given by:
J r e g i s t r a t i o n i 1 = r registration i 1 , i T ξ i 1 = 0 3 × 3 J l 1 ln R i 1 , i 1 R i 1 1 R i R i T I 3 × 3 R i 1 t i 1 , i 6 × 6
  • By differentiating the residual equation, the Jacobian matrix J r e g i s t r a t i o n i can be derived as follows:
    r registration i 1 , i T ξ i = ln R i 1 , i 1 R i 1 1 R i t i t i 1 R i 1 t i 1 , i 6 × 1 t i ϕ i = ln R i 1 , i 1 R i 1 1 R i t i ln R i 1 , i 1 R i 1 1 R i ϕ i t i t i 1 R i 1 t i 1 , i t i t i t i 1 R i 1 t i 1 , i ϕ i = 0 3 × 3 ln R i 1 , i 1 R i 1 1 R i ϕ i I 3 × 3 0 3 × 3
The computation of ln R i 1 , i 1 R i 1 1 R i ϕ i is given as follows:
ln R i 1 , i 1 R i 1 1 R i ϕ i = lim ϕ i 0 ln R i 1 , i 1 R i 1 1 R i exp ϕ i ln R i 1 , i 1 R i 1 1 R i ϕ i = lim ϕ i 0 ln R i 1 , i 1 R i 1 1 R i + J r 1 ln R i 1 , i 1 R i 1 1 R i ϕ i ln R i 1 , i 1 R i 1 1 R i ϕ i = J r 1 ln R i 1 , i 1 R i 1 1 R i = J l 1 ln R i 1 , i 1 R i 1 1 R i
Accordingly, the Jacobian matrix J r e g i s t r a t i o n i is given by:
J r e g i s t r a t i o n i = r registration i 1 , i T ξ i = 0 3 × 3 J l 1 ln R i 1 , i 1 R i 1 1 R i I 3 × 3 0 3 × 3 6 × 6

3.1.4. Design of the Pose Graph Model

During the optimization process, if constraints are constructed directly from LiDAR point cloud feature correspondences, the number of matched points will increase rapidly as the number of poses grows. This leads to a substantial computational burden and is unfavorable for the design of lightweight systems.
Therefore, in the back-end, a pose graph optimization framework is adopted, in which the poses obtained from point cloud registration are treated as optimization variables, thereby avoiding large-scale point cloud matching and significantly reducing computational complexity [29,30].
The optimization model incorporates three types of constraints: GPS position constraints, loop closure constraints, and point cloud registration constraints. The resulting pose graph is illustrated as follows:
The objective function is formulated as follows:
F ( ξ 1 , ξ 2 , ξ 3 , ξ 4 , ξ 5 , ξ 6 , , ξ n ) = min ξ 1 2 ( k , i ε l o o p c l o s u r e r k , i T Σ k , i 1 r k , i + i 1 , i ε r e g i s t r a t i o n r i 1 , i T Σ i 1 , i 1 r i 1 , i + i ε g p s r i T Σ i 1 r i )
Following the standard design of pose graph optimization algorithms, the three types of constraints are incorporated into the optimization framework, which yields the complete objective function of the pose graph optimization algorithm. Additionally, in order to handle measurement uncertainties and improve the robustness of the optimization, each constraint is weighted by an information matrix that reflects the confidence level of the corresponding measurement. These weights ensure that more reliable measurements, such as GPS, have a greater impact on the optimization process, while less reliable constraints are given less importance. To mitigate the impact of potential outliers, the Huber loss function is employed as a robust kernel, with its parameter set to δ = 1.345, which provides a balance between quadratic and linear loss and effectively reduces the influence of large residuals. This enhances the stability and precision of the optimization process, ensuring that it remains both computationally efficient and resilient to data errors.

3.2. Technical Implementation Scheme

(1)
GPS Module
  • The satellite information acquisition module is responsible for receiving raw GPS satellite signal data;
  • The latitude and longitude solution module computes the UAV’s geodetic coordinates based on the received signals and provides initial information for the coordinate transformation module;
  • The coordinate transformation module converts the geodetic coordinates into the Earth-Centered Earth-Fixed (ECEF) coordinate system and subsequently transforms them into relative coordinates in the LiDAR coordinate frame, which are defined as the node coordinates in this study.
(2)
LiDAR Module
  • The point cloud acquisition module collects and generates point cloud data of the surrounding environment using an onboard mechanical LiDAR. The point cloud data are represented in the form of three-dimensional coordinates X Y Z . This module performs data acquisition and transmission, providing raw point cloud data for subsequent processing stages, and serves as the basis for lightweight front-end perception in power inspection scenarios.
  • The point cloud filtering module operates on the output of the acquisition layer. In this work, K-means clustering is adopted as a lightweight preprocessing strategy to organize local point cloud structure and suppress redundant or irrelevant small-scale noise points with relatively low computational cost. Compared with more complex density-based or model-based filtering methods, this design is more suitable for the current real-time-oriented framework. In addition, voxel grid filtering is employed to reduce the number of points, thereby decreasing algorithmic complexity and enhancing real-time performance. The combination of K-means clustering and voxel grid downsampling provides a practical balance between filtering effectiveness and computational efficiency.
  • The point cloud calibration module utilizes the raw point cloud data together with the matching results from the two preceding frames. A prediction-based approach is adopted to transform the point clouds from each scanning cycle into a unified temporal reference, producing temporally aligned point clouds that are subsequently fed into the feature extraction module, which helps improve temporal consistency before registration.
  • The feature extraction module extracts edge features and planar features from the point cloud based on curvature analysis. This curvature-based design is chosen because it can effectively preserve the dominant structural characteristics of power-line environments while maintaining relatively low implementation complexity. Compared with learning-based feature extraction methods, it avoids additional training requirements and higher computational overhead, making it more suitable for the current stage of the proposed framework. Feature sequences corresponding to different time instants are stored for matching purposes. The point cloud matching layer matches features from identical spatial locations across two time instants and forwards the results to the pose estimation layer.
  • The point cloud matching module performs feature-based matching between two time instants and employs the Levenberg–Marquardt (LM) nonlinear optimization method to decouple and estimate the UAV rotation matrix R and translation vector T . The resulting pose estimates are then passed to the localization fusion layer. This feature-based matching strategy further reduces the computational burden compared with direct dense matching, and is therefore more compatible with the efficiency requirements of UAV onboard localization.
(3)
Remaining Modules
  • The graph optimization module adopts a sliding-window-based optimization strategy. It fuses the pose transformation information between two consecutive point clouds output by the point cloud matching module with the relative GPS poses provided by the coordinate transformation module. By jointly optimizing multiple historical state variables through a least-squares framework, the pose states of multiple time instants are estimated simultaneously.
  • The inverse coordinate transformation module converts the fused pose estimates back into the Earth-Centered Earth-Fixed (ECEF) coordinate system and subsequently transforms them into geodetic coordinates (latitude and longitude), which are then output as the final positioning results. The coordinate transformation process remains as follows: First, the pose estimates (e.g., position and orientation) in the local frame are converted to ECEF coordinates using a series of geometric transformations. This involves first translating the local coordinate system’s origin to the Earth’s center and then rotating the coordinates to align with the Earth’s rotational axes. The rotation is performed using a direction cosine matrix (DCM) or quaternion-based transformations. Second, once the position is in the ECEF frame, it is transformed into geodetic coordinates (latitude, longitude, and altitude). The relationship between ECEF coordinates and geodetic coordinates is non-trivial and requires solving the following set of nonlinear equations:
Latitude equation, Longitude equation and Height equation:
t a n ϕ = z + e 2 · b · s i n 3 ϕ p e 2 · a · c o s 3 ϕ
λ = a r c t a n y x
h = p c o s ϕ N ϕ
where N ( ϕ ) is the radius of curvature in the prime vertical, and p = x 2 + y 2 . These equations are solved iteratively using the Newton–Raphson method to refine the latitude (φ), longitude (λ), and height (h) by iterating based on initial guesses.
The above figure illustrates the procedure for obtaining edge coordinates from the LiDAR module. Specifically, during the k -th execution cycle, the LiDAR acquires a point cloud set P C 0 ( k ) over one scanning period. Using the integrated navigation pose sequence T 0 ( k ) within the same period, the point clouds are temporally calibrated to generate a synchronized point cloud set P C 1 ( k ) , in which all points are referenced to a common time instant. Subsequently, K-means clustering is applied for denoising, resulting in the filtered point cloud set P C 2 ( k ) .
Line and planar features Q ( k ) are then extracted from the processed point cloud and matched with the corresponding features Q ( k 1 ) from the previous cycle. Based on the matched features, an error function X ( k ) is constructed with the rotation matrix R and translation vector T as optimization variables. If the condition X ( k ) < T   (where T denotes a predefined accuracy threshold) is not satisfied, the rotation matrix R ( k ) and translation vector T ( k ) are refined using the Levenberg–Marquardt (LM) optimization method. This iterative optimization process continues until the convergence criterion is met, after which the optimized R ( k ) and T ( k ) are output as the relative pose between the two nodes.
Figure 4 above illustrates the pose graph optimization process. In this framework, the GPS-derived estimates are incorporated as node constraints, while the LiDAR-derived pose estimates are treated as edge constraints between nodes. A pose graph consisting of six node constraints and five inter-node edge constraints is constructed. Using a sliding-window strategy, the optimization continuously incorporates the most recent measurements while discarding older state variables. The Levenberg–Marquardt (LM) method is employed to iteratively optimize the pose graph, ultimately yielding optimal estimates of multiple state variables, from which the final poses are obtained.
The iterative algorithm for nonlinear least-squares optimization proceeds as follows. First, the nonlinear problem is linearized to construct an equivalent linear system. Then, the increment of the optimization variables is computed from this linear system and used to update the current estimate. This updated estimate is taken as the starting point of the next iteration. When the maximum number of iterations is reached, or when the objective function value falls below a predefined threshold, the optimization is considered to have converged. At that point, the corresponding optimization variables are taken as the final solution. The following Figure 5 shows the flowchart of the LiDAR-based pose estimation algorithm.

4. Experimental Setup and Result Analysis

4.1. Experimental Evaluation of UKF-Based LiDAR–IMU-Aided Localization

The experimental evaluation in this section is based on the Lie-group-based UKF-SLAM framework introduced in Section 2.2.2, with the objective of validating its effectiveness in improving the accuracy and stability of system state estimation.
The improved UKF algorithm based on Lie group operations follows the formulation presented in Section 2.2.2. For reference and to facilitate reproducibility, the overall workflow of the proposed LiDAR–IMU–GPS fusion framework, including both the front-end UKF-SLAM and the back-end pose graph optimization, is summarized in Algorithm 1.
Algorithm 1: LiDAR–IMU–GPS Fusion for UAV Localization
Input:
  -   LiDAR   point   clouds   P k at time k
  -   IMU   measurements   I k   =   { w m ,   a m } between k − 1 and k
  -   GPS   readings   g k (latitude, longitude, altitude)
Output:
  -   Optimized   pose   trajectory   T k ∈SE(3) for each time step k
1:   Initialize   state   vector   χ 0   ( pose ,   velocity ,   biases ,   landmarks ) ,   covariance   P 0 , empty pose graph
2:  for each time step k = 1, 2, … do
3:    // ---------- Front-end: LiDAR–IMU fusion via Lie-group UKF ----------
4 :     Perform   IMU   pre - integration   over   [ t k 1 ,   t k ] using Equations (2)–(6)
5 :     Generate   sigma   points   σ i on SE(3) manifold (Equations (28)–(30))
6:    Propagate sigma points through motion model (Equations (2)–(6))
7 :   Compute   predicted   mean   μ ^ S E 3   and   covariance   P ^ S E 3 using geodesic distance (Equations (31) and (32))
8 :     if   new   LiDAR   scan   P k is available then
9:      Extract edge and planar features based on curvature (Section 3.2)
10 :     Match   features   with   previous   frame   P k 1
11:    Compute observation model (Equations (8) and (9))
12:    Update UKF: compute Kalman gain, posterior state and covariance on SE(3) (Equations (33)–(38))
13:    end if
14:
15:    // ---------- Back-end: Pose graph optimization (triggered every N frames) ----------
16:    if k mod N = 0 then
17 :     Add   current   pose   T k as a new node in the graph
18:    if GPS data available then
19 :   Add   GPS   constraint   edge :   residual   r g p s i (Equation (40)), Jacobian (Equation (41))
20:    end if
21 :       Add   odometry   edge   between   T k 1   and   T k using LiDAR registration residual (Equation (54)), Jacobians (Equations (57) and (60))
22:    if loop closure detected between current frame and historical frame l then
23:      Add loop closure edge: residual r l o o p k , i (Equation (44)), Jacobians (Equations (48) and (51))
24:    end if
25:     Solve nonlinear least-squares problem using Levenberg–Marquardt (Equation (61))
26:    end if
27: end for
28 : return   optimized   poses   T k
The following Figure 6 presents the validation results of the algorithm on the test dataset.
Experimental results demonstrate that the proposed method achieves effective and robust state estimation, with localization accuracy that is comparable to or superior to that of current state-of-the-art approaches. The experimental results are shown in Figure 7 below.
Table 1 presents a comparison of the absolute trajectory error (ATE) between the proposed method and three representative state-of-the-art approaches (VIN-MONO, ROVIO, and NormalUKF) across five data sequences. The results indicate that the proposed method achieves superior performance on multiple sequences, particularly attaining the lowest error on the MH_02_easy sequence (0.170). Overall, its localization accuracy is comparable to or better than that of the best-performing baseline methods.

4.2. Experimental Evaluation of Pose Graph Optimization-Based LiDAR-GPS-Aided Localization

4.2.1. Data Analysis

Experiments are conducted on data sequences containing loop closures to verify the feasibility of the loop closure detection mechanism. The experimental results are shown in Table 2 below.
During operation, ten loop closure points are detected based on GPS coordinate distances and point cloud distance metrics. The fitting errors of all detected loop closures are below the predefined upper threshold (≤0.20). According to the loop closure detection results, ten edges corresponding to loop closure constraints are incorporated into the back-end optimization for further computation.
On Dataset Sequence 2, the front-end odometry is evaluated without back-end optimization. The original localization accuracy error is 4.535629 m, and the cumulative drift reaches 54.519383 m [here, cumulative error refers to the RMSE of the absolute pose error (APE) accumulated over the entire trajectory, which quantifies the total positional drift relative to the ground truth], indicating that the odometry trajectory has significantly deviated from the ground truth. After enabling back-end optimization with loop closure detection, the pose graph incorporates 10 loop-closure constraint edges, 1390 registration-based pose constraint edges, and jointly optimizes 1391 pose nodes. As a result, the localization accuracy error is reduced to 4.455809 m, while the cumulative drift is substantially decreased to 6.830040 m. These results demonstrate that back-end optimization with loop closure detection can effectively correct the localization trajectory in a timely manner and significantly reduce accumulated localization drift.
After further incorporating GPS-based back-end optimization, the pose graph contains a total of 2791 edges, including 1391 edges derived from GPS position constraints, 10 edges from loop closure constraints, and 1390 edges from point cloud registration constraints, with 1391 pose vertices in total. The resulting localization accuracy error is reduced to 1.173358 m. Since the incorporated GPS measurements exhibit only instantaneous (non-accumulating) errors rather than cumulative drift, the overall accumulated error of the system is reduced to 0.086931 m, indicating a substantial suppression of long-term drift.
A comparison with the GPS-assisted back-end optimization experiment without loop closure detection shows no significant degradation in localization accuracy. This is likely due to the relatively small number of loop closure constraints, resulting in comparable performance with or without loop closure detection when GPS constraints are included in the back-end optimization. The following Table 3 demonstrates the effect of back-end optimization and loop closure detection on eliminating cumulative errors in dataset sequence 2, and the following Figure 8 shows the error curves and XYZ errors of various methods.

4.2.2. Experimental Summary

  • The system architecture design for UAV-based LiDAR–GPS fusion localization comprises multiple functional modules, including a satellite signal acquisition module, a latitude–longitude computation module, a coordinate transformation module, a point cloud acquisition module, a point cloud filtering module, a point cloud calibration module, a feature extraction module, a point cloud matching module, a graph optimization module, and an inverse coordinate transformation module.
  • The algorithmic framework for UAV-based LiDAR–GPS fusion localization is established by integrating several key components, including the conversion of GPS geodetic coordinates into node coordinates for pose graph optimization, the algorithmic design for estimating the relative pose between two LiDAR point cloud frames, and the subsequent coupling of these estimates into a pose graph optimization framework. Through joint optimization, more accurate geodetic position estimates (latitude and longitude) are ultimately obtained.

5. Conclusions

This paper proposes a multi-sensor navigation method for UAV-based power inspection that integrates a tightly coupled LiDAR-IMU front-end (Lie-group UKF-SLAM) with a loosely coupled back-end (pose graph optimization with GPS and loop closure). The front-end enhances local estimation stability and accuracy, while the back-end effectively suppresses long-term drift using GPS constraints and loop closure.
Experimental results demonstrate that the proposed method achieves competitive performance on both public benchmark datasets and self-collected power inspection scenarios. In particular, in Section 4.1, on the MH_02_easy sequence, the proposed method reduces the absolute trajectory error (ATE) from 0.521 m (ROVIO) to 0.170 m, corresponding to an error reduction of approximately 67.4%. In Section 4.2, for Dataset Sequence 2, when only the front-end odometry is used, the original accuracy error is 4.535629 m and the cumulative error is 54.519383 m; after enabling loop-closure-based back-end optimization together with GPS-assisted back-end optimization, the cumulative error is reduced to 0.086931 m, demonstrating a reduction of more than 99% in cumulative drift relative to the front-end-only result. In addition, under GPS-degraded conditions, the system is able to maintain stable navigation through loosely coupled LiDAR–IMU fusion, which further verifies its robustness and practical applicability in complex power inspection environments.
Future work will focus on the following directions. First, further efforts will be devoted to improving the system’s adaptability in dynamic environments and under strong electromagnetic interference, with an emphasis on exploring robustness enhancement strategies for multi-modal sensor fusion. Second, deep learning-based feature extraction and matching methods will be investigated to improve the efficiency and accuracy of point cloud registration. Third, the deployment and optimization of the proposed system on edge computing platforms will be explored to better meet the requirements of real-time onboard processing and low power consumption in UAV applications. Fourth, this work can be extended to small unmanned surface vehicles equipped with autonomous navigation systems, thereby achieving a cross-domain breakthrough. Fifth, we will continue to complete field experiments and promote the application of this algorithm on real UAVs to meet corresponding engineering requirements. Sixth, to further strengthen the statistical validation of the proposed method, future work will include repeated experimental runs under consistent conditions to compute key uncertainty descriptors such as variance and confidence intervals. Seventh, we plan to deploy the proposed system on embedded platforms to evaluate onboard processing capability, and optimize the code for parallel execution and low-latency inference.

Author Contributions

Conceptualization, A.W. and H.M.; methodology, W.Y.; software, X.D.; validation, Y.Y.; formal analysis, S.L.; investigation, S.L.; resources, H.M.; writing—review and editing, J.L.; funding acquisition, A.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Project of China Southern Power Grid Company Limited (CGYKJXM20240218).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the first author.

Acknowledgments

We sincerely appreciate the valuable comments and suggestions provided by the editors and reviewers to enhance this research.

Conflicts of Interest

Authors Anjun Wang, Wenbin Yu, Xuexing Dong, Yang Yang, and Shizeng Liu were employed by Dali Bureau, EHV Power Transmission Company of China Southern Power Grid. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

References

  1. Tavasci, L.; Nex, F.; Gandolfi, S. Reliability of real-time kinematic (RTK) positioning for low-cost drones’ navigation across global navigation satellite System (GNSS) critical environments. Sensors 2024, 24, 6096. [Google Scholar] [CrossRef]
  2. Elamin, A.; Abdelaziz, N.; El-Rabbany, A. A GNSS/INS/LiDAR integration scheme for UAV-based navigation in GNSS-challenging environments. Sensors 2022, 22, 9908. [Google Scholar] [CrossRef]
  3. Chang, Y.; Cheng, Y.; Manzoor, U.; Murray, J. A review of UAV autonomous navigation in GPS-denied environments. Robot. Auton. Syst. 2023, 170, 104533. [Google Scholar] [CrossRef]
  4. Xie, L.; Jiang, C.; Sun, Q.; Wang, H.; Song, Q.; Guan, G. The global map’s creating and positioning of substation inspection robot based on adaptive Monte Carlo particle filter algorithm. Electr. Power Eng. Technol. 2019, 38, 16–23. [Google Scholar]
  5. Zhou, W.; Liu, G. Review of overhead line defect inspection based on deep learning and UAV images. Electr. Power Eng. Technol. 2024, 43, 73–82. [Google Scholar]
  6. Gyagenda, N.; Hatilima, J.V.; Roth, H.; Zhmud, V. A review of GNSS-independent UAV navigation techniques. Robot. Auton. Syst. 2022, 152, 104069. [Google Scholar] [CrossRef]
  7. Mahdi, A.E.; Azouz, A.; Abdalla, A.E.; Abosekeen, A. A machine learning approach for an improved inertial navigation system solution. Sensors 2022, 22, 1687. [Google Scholar] [CrossRef]
  8. Buchanan, R.; Agrawal, V.; Camurri, M.; Dellaert, F.; Fallon, M. Deep imu bias inference for robust visual-inertial odometry with factor graphs. IEEE Robot. Autom. Lett. 2022, 8, 41–48. [Google Scholar] [CrossRef]
  9. Lai, L.; Li, L.; Wang, H.; Yuan, J.; Fan, W.; Zhao, D. Enhanced LiDAR-inertial SLAM with adaptive intensity feature extraction and fusion. Measurement 2025, 253, 117738. [Google Scholar] [CrossRef]
  10. Yang, J.-C.; Lin, C.-J.; You, B.-Y.; Yan, Y.-L.; Cheng, T.-H. Rtlio: Real-time lidar-inertial odometry and mapping for UAVS. Sensors 2021, 21, 3955. [Google Scholar] [CrossRef]
  11. Julier, S.J.; Uhlmann, J.K. Unscented filtering and nonlinear estimation. Proc. IEEE 2004, 92, 401–422. [Google Scholar] [CrossRef]
  12. Crassidis, J.L.; Markley, F.L. Unscented filtering for spacecraft attitude estimation. J. Guid. Control Dyn. 2003, 26, 536–542. [Google Scholar] [CrossRef]
  13. da Silva, M.F.; Honório, L.M.; Marcato, A.L.M.; Vidal, V.F.; Santos, M.F. Unmanned aerial vehicle for transmission line inspection using an extended Kalman filter with colored electromagnetic interference. ISA Trans. 2020, 100, 322–333. [Google Scholar] [CrossRef]
  14. Shan, T.; Englot, B.; Meyers, D.; Wang, W.; Ratti, C.; Rus, D. Lio-sam: Tightly-coupled lidar inertial odometry via smoothing and mapping. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 5135–5142. [Google Scholar]
  15. He, X.; Pan, S.; Gao, W.; Lu, X. LiDAR-inertial-GNSS fusion positioning system in urban environment: Local accurate registration and global drift-free. Remote Sens. 2022, 14, 2104. [Google Scholar] [CrossRef]
  16. Markley, F.L. Attitude estimation or quaternion estimation? J. Astronaut. Sci. 2004, 52, 221–238. [Google Scholar] [CrossRef]
  17. Pittelkau, M.E. An analysis of the quaternion attitude determination filter. J. Astronaut. Sci. 2003, 51, 103–120. [Google Scholar] [CrossRef]
  18. Li, X.R.; Jilkov, V.P. Survey of maneuvering target tracking. Part V. Multiple-model methods. IEEE Trans. Aerosp. Electron. Syst. 2005, 41, 1255–1321. [Google Scholar] [CrossRef]
  19. Yang, J.; Han, X.; Deng, W.; Jin, H.; Zhang, B. Laser Pulse-Driven Multi-Sensor Time Synchronization Method for LiDAR Systems. Sensors 2025, 25, 7555. [Google Scholar] [CrossRef]
  20. Gao, Y.; Zhao, L. VE-LIOM: A Versatile and Efficient LiDAR-Inertial Odometry and Mapping System. Remote Sens. 2024, 16, 2772. [Google Scholar] [CrossRef]
  21. Gao, J.; Sha, J.; Wang, Y.; Wang, X.; Tan, C. A fast and stable GNSS-LiDAR-inertial state estimator from coarse to fine by iterated error-state Kalman filter. Robot. Auton. Syst. 2024, 175, 104675. [Google Scholar] [CrossRef]
  22. Bono, F.M.; Polinelli, A.; Radicioni, L.; Benedetti, L.; Castelli-Dezza, F.; Cinquemani, S.; Belloli, M. Wireless Accelerometer Architecture for Bridge SHM: From Sensor Design to System Deployment. Future Internet 2025, 17, 29. [Google Scholar] [CrossRef]
  23. Julier, S.; Uhlmann, J.; Durrant-Whyte, H.F. A new method for the nonlinear transformation of means and covariances in filters and estimators. IEEE Trans. Autom. Control 2002, 45, 477–482. [Google Scholar] [CrossRef]
  24. Celledoni, E.; Owren, B. Lie group methods for rigid body dynamics and time integration on manifolds. Comput. Methods Appl. Mech. Eng. 2003, 192, 421–438. [Google Scholar] [CrossRef]
  25. Müller, A.; Terze, Z. The significance of the configuration space Lie group for the constraint satisfaction in numerical time integration of multibody systems. Mech. Mach. Theory 2014, 82, 173–202. [Google Scholar] [CrossRef]
  26. Bai, Y.B.; Holden, L.; Kealy, A.; Zaminpardaz, S.; Choy, S. A hybrid indoor/outdoor detection approach for smartphone-based seamless positioning. J. Navig. 2022, 75, 946–965. [Google Scholar] [CrossRef]
  27. Steger, C.R.; Steger, B.; Schär, C. HORAYZON v1.2: An efficient and flexible ray-tracing algorithm to compute horizon and sky view factor. Geosci. Model Dev. 2022, 15, 6817–6840. [Google Scholar] [CrossRef]
  28. Baasch, K.-N.; Icking, L.; Ruwisch, F.; Schön, S. Coordinate frames and transformations in GNSS ray-tracing for autonomous driving in urban areas. Remote Sens. 2022, 15, 180. [Google Scholar] [CrossRef]
  29. Ćwian, K.; Nowicki, M.R.; Wietrzykowski, J.; Skrzypczyński, P. Large-scale LiDAR SLAM with factor graph optimization on high-level geometric features. Sensors 2021, 21, 3445. [Google Scholar] [CrossRef]
  30. Kretzschmar, H.; Stachniss, C. Information-theoretic compression of pose graphs for laser-based SLAM. Int. J. Robot. Res. 2012, 31, 1219–1230. [Google Scholar] [CrossRef]
Figure 1. Sensor Characteristics and Fusion Results.
Figure 1. Sensor Characteristics and Fusion Results.
Applsci 16 02632 g001
Figure 2. Schematic Diagram of the Temporal Alignment Algorithm.
Figure 2. Schematic Diagram of the Temporal Alignment Algorithm.
Applsci 16 02632 g002
Figure 3. Schematic Diagram of the Spatial Alignment Algorithm.
Figure 3. Schematic Diagram of the Spatial Alignment Algorithm.
Applsci 16 02632 g003
Figure 4. Schematic Diagram of Pose Graph Optimization.
Figure 4. Schematic Diagram of Pose Graph Optimization.
Applsci 16 02632 g004
Figure 5. Flowchart of the LiDAR-Based Pose Estimation Algorithm.
Figure 5. Flowchart of the LiDAR-Based Pose Estimation Algorithm.
Applsci 16 02632 g005
Figure 6. Validation Results on the Test Dataset.
Figure 6. Validation Results on the Test Dataset.
Applsci 16 02632 g006
Figure 7. Variation in Localization Error Curves.
Figure 7. Variation in Localization Error Curves.
Applsci 16 02632 g007
Figure 8. Accumulated Error Curve. (a) Error Curve of Front-end Odometry; (b) XYZ Errors of Front-end Odometry; (c) Error Curve of Front-end Odometry + Back-end Optimization with Loop Closure Detection; (d) XYZ Errors of Front-end Odometry + Back-end Optimization with Loop Closure Detection; (e) Error Curve of Front-end Odometry + GPS-based Back-end Optimization; (f) XYZ Errors of Front-end Odometry + GPS-based Back-end Optimization; (g) Error Curve of Front-end Odometry + Back-end Optimization with Loop Closure Detection + GPS-based Back-end Optimization; (h) XYZ Errors of Front-end Odometry + Back-end Optimization with Loop Closure Detection + GPS-based Back-end Optimization.
Figure 8. Accumulated Error Curve. (a) Error Curve of Front-end Odometry; (b) XYZ Errors of Front-end Odometry; (c) Error Curve of Front-end Odometry + Back-end Optimization with Loop Closure Detection; (d) XYZ Errors of Front-end Odometry + Back-end Optimization with Loop Closure Detection; (e) Error Curve of Front-end Odometry + GPS-based Back-end Optimization; (f) XYZ Errors of Front-end Odometry + GPS-based Back-end Optimization; (g) Error Curve of Front-end Odometry + Back-end Optimization with Loop Closure Detection + GPS-based Back-end Optimization; (h) XYZ Errors of Front-end Odometry + Back-end Optimization with Loop Closure Detection + GPS-based Back-end Optimization.
Applsci 16 02632 g008
Table 1. Experimental Results on the Dataset.
Table 1. Experimental Results on the Dataset.
SequenceMH_01_EasyMH_02_EasyMH_03_MediumMH_04_DifficultMH_05_Difficult
VIN-MONO0.1140.1860.2020.2570.254
ROVIO0.1720.5210.3860.5220.418
NormalUKF0.2840.1990.2460.4900.351
Proposed0.1390.1700.2430.3230.261
Table 2. Loop Closure Detection Results on Dataset Sequence 2.
Table 2. Loop Closure Detection Results on Dataset Sequence 2.
Loop Closure IndexNew FrameOld FrameFitting Error
19831530.156819
29881580.0504666
313539440.0826466
413589490.0233855
513639540.0472858
613689590.0849739
713739640.0943137
813789690.103733
913839730.107383
1013889770.0551427
Table 3. Effect of Back-End Optimization with Loop Closure Detection on Eliminating Accumulated Drift for Dataset Sequence 2.
Table 3. Effect of Back-End Optimization with Loop Closure Detection on Eliminating Accumulated Drift for Dataset Sequence 2.
MethodAttributeRPEATEAPEAttributeValue
Front-end odometryMax18.938824110.030900111.690937----
Mean2.80753142.33358742.502190----
Median1.83312639.17852139.217980----
Min0.5294290.0000000.000001----
RMSE4.53562954.28076154.519383----
SSE555.44221940984434134557----
STD3.56226133.97452734.145673----
Front-end odometry + back-end optimization with loop closure detectionMax20.32671716.01869516.129760----
Mean2.5032855.4704325.468591Number of loop closure edges10
Median1.4560173.9025933.899106Number of optimization edges1400
Min0.6235400.0000000.000001Number of vertices1391
RMSE4.4558096.8290476.830040Error before optimization--
SSE536.06436964870.564889.4Error after optimization--
STD3.6861644.0878184.091939Optimization time0.921837
Front-end odometry + GPS-based back-end optimizationMax2.2742380.7482670.800248----
Mean1.0772700.0433100.055482Number of loop closure edges0
Median0.9988230.0269020.038200Number of optimization edges2781
Min0.2878640.0016360.007302Number of vertices1391
RMSE1.1840770.0761190.086910Error before optimization1103180
SSE37.8550058.05959510.506733Error after optimization23.5963
STD0.4914540.0625970.066896Optimization time0.110304
Front-end odometry + back-end optimization with loop closure detection + GPS-based back-end optimizationMax2.2742380.7482670.800247----
Mean1.0566820.0438510.055488Number of loop closure edges10
Median0.9833740.0274040.038182Number of optimization edges2791
Min0.2453640.0016380.007325Number of vertices1391
RMSE1.1733580.0764250.086931Error before optimization1157560
SSE37.1727548.12454810.511893Error after optimization23.7683
STD0.5100900.0625930.017162Optimization time0.17204
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, A.; Yu, W.; Dong, X.; Yang, Y.; Liu, S.; Liu, J.; Mei, H. Multi-Sensor-Assisted Navigation for UAVs in Power Inspection: A Fusion Approach Using LiDAR, IMU and GPS. Appl. Sci. 2026, 16, 2632. https://doi.org/10.3390/app16062632

AMA Style

Wang A, Yu W, Dong X, Yang Y, Liu S, Liu J, Mei H. Multi-Sensor-Assisted Navigation for UAVs in Power Inspection: A Fusion Approach Using LiDAR, IMU and GPS. Applied Sciences. 2026; 16(6):2632. https://doi.org/10.3390/app16062632

Chicago/Turabian Style

Wang, Anjun, Wenbin Yu, Xuexing Dong, Yang Yang, Shizeng Liu, Jiahao Liu, and Hongwei Mei. 2026. "Multi-Sensor-Assisted Navigation for UAVs in Power Inspection: A Fusion Approach Using LiDAR, IMU and GPS" Applied Sciences 16, no. 6: 2632. https://doi.org/10.3390/app16062632

APA Style

Wang, A., Yu, W., Dong, X., Yang, Y., Liu, S., Liu, J., & Mei, H. (2026). Multi-Sensor-Assisted Navigation for UAVs in Power Inspection: A Fusion Approach Using LiDAR, IMU and GPS. Applied Sciences, 16(6), 2632. https://doi.org/10.3390/app16062632

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop