Next Article in Journal
Challenges and Prospects of Sensing Technology for the Promotion of Tele-Physiotherapy: A Narrative Review
Previous Article in Journal
Algorithmic Generation of Parameterized Geometric Models of the Aortic Valve and Left Ventricle
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Digital Terrain Construction Based on IMU and LiDAR Fusion Perception

1
Institute of Automotive Engineering, Jiangsu University, Zhenjiang 212013, China
2
National Key Laboratory for Intelligent and Green Vehicles and Transportation, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(1), 15; https://doi.org/10.3390/s25010015
Submission received: 16 October 2024 / Revised: 9 December 2024 / Accepted: 18 December 2024 / Published: 24 December 2024
(This article belongs to the Section Radar Sensors)

Abstract

:
To address the shortcomings of light detection and ranging (LiDAR) sensors in extracting road surface elevation information in front of a vehicle, a scheme for digital terrain construction based on the fusion of an Inertial Measurement Unit (IMU) and LiDAR perception is proposed. First, two sets of sensor coordinate systems were configured, and the parameters of LiDAR and IMU were calibrated. Then, a terrain construction system based on the fusion perception of IMU and LiDAR was established, and improvements were made to the state estimation and mapping architecture. Terrain construction experiments were conducted in an academic setting. Finally, based on the output information from the terrain construction system, a moving average-like algorithm was designed to process point cloud data and extract the road surface elevation information at the vehicle’s trajectory position. By comparing the extraction effects of four different sliding window widths, the 4 cm width sliding window, which yielded the best results, was ultimately selected, making the extracted road surface elevation information more accurate and effective.

1. Introduction

With the continuous advancement of vehicle suspension technology, intelligent suspension systems are capable of autonomously sensing the terrain ahead, making decisions, and executing complex suspension control tasks [1,2]. The realization of this capability relies heavily on sensor technology. As an indispensable component of intelligent driving systems, sensor technology provides real-time and precise environmental information for intelligent suspension systems. However, the performance of the intelligent suspension system is directly affected by the accuracy of the sensors and the quality of the algorithms [3,4]. With the ongoing development of technologies such as LiDAR and visual sensors, it has become possible to accurately acquire elevation information for the road surface in front of the vehicle. However, research on acquiring elevation information for the road surface in front of the vehicle is currently limited. Therefore, it is necessary to conduct digital terrain research on front road surface perception.
For terrain-detection technologies [5,6,7,8,9], researchers are utilizing sensor technologies for terrain measurement, which can be mainly divided into three research directions: vision-based terrain-detection methods, LiDAR-based terrain-detection methods, and terrain-detection methods based on multi-sensor fusion. Vision-based terrain-detection methods have advantages such as a low cost and low energy consumption, which have led many scholars to conduct extensive research on map construction based on visual sensors. Klein et al. [10] utilized the concepts of vision and keyframes to enhance the real-time capability of mapping, although this method is primarily applicable for map construction and navigation, making it inadequate for high-precision mapping needs. Raul et al. [11] proposed a vision-based map construction algorithm that simultaneously processes feature detection and tracking, local map construction, and loop-closure detection optimization, thereby improving the accuracy and robustness of the algorithm to achieve high-precision positioning and navigation. However, this method struggles to acquire elevation information for the road surface in front of the vehicle. Additionally, since visual sensors are sensitive to lighting, texture, and rapid motion, their application scenarios may be limited [12,13,14,15]. Thus, vision-based terrain-detection methods are mainly applied in indoor environments and relatively structured roadways.
LiDAR-based terrain-detection methods offer advantages such as long measurement distances, high ranging accuracy, high resolution, and immunity to lighting conditions. Consequently, LiDAR is more advantageous for identifying road surface information. Zhang et al. [16] proposed the real-time LiDAR Odometry and Mapping (LOAM) algorithm, which enables real-time odometry and mapping functions using LiDAR. However, this algorithm lacks loop-closure detection and cannot handle large-scale rotational transformations. Shan et al. [17] extended the LOAM algorithm to develop the Lightweight and Ground-Optimized LiDAR Odometry and Mapping on Variable Terrain (LeGO-LOAM) algorithm. This method optimizes the algorithm’s lightweight nature and ground adaptability through point cloud classification and feature extraction. It also incorporates loop-closure detection and pose graph optimization, enhancing the system’s robustness. Nevertheless, the ground point extraction algorithm of this method does not meet the needs of complex terrain and requires a precise LiDAR installation. To reduce drift errors during the mapping process, Lin et al. [18] proposed a loop-closure detection-based algorithm to correct accumulated errors in the map. This method calculates the 2D histograms of keyframes and uses Normalized Cross-Correlation (NCC) as a similarity metric between adjacent keyframes. Experimental results show that this approach improves the reliability and robustness of the algorithm. Wang et al. [19] introduced a lightweight 3D localization and mapping method for solid-state LiDAR, named Solid-State LiDAR Simultaneous Localization and Mapping (SSL SLAM). Building on previous research, this method employs a scan-to-map approach to enhance performance and a sliding window technique to reduce computational costs, ensuring real-time capability. With the ongoing enhancement of LiDAR performance, high-beam LiDARs can extract clearer contour information. As LiDARs are mass-produced, their costs will continue to decrease. Therefore, LiDAR technology is advancing towards higher performance, lower cost, and broader applications to meet market demands in various fields [20,21,22].
Terrain-detection methods based on multi-sensor fusion are more commonly used. This approach leverages the advantages of combining multiple sensors to compensate for the shortcomings of individual sensors, thereby improving data accuracy and robustness, and achieving more reliable terrain analysis and technical support in complex and variable environments. Nguyen-Ngoc et al. [23] proposed a tightly coupled GPS/INS/LIDAR integration system that uses Factor Graph Optimization (FGO) to integrate measurement data from various sensors into the system, enabling high-precision localization and mapping. However, the multi-sensor fusion system faces challenges related to the time synchronization and spatial calibration of different sensors, as well as high computational resource requirements, and the algorithm suffers from error accumulation issues. Therefore, further improvements are needed in terms of real-time performance and robustness. Almalioglu et al. [24] developed a framework based on the Unscented Kalman Filter (UKF) that integrates millimeter-wave radar and IMU for low-cost indoor 6DoF pose estimation. However, the accuracy, resolution, and processing speed of the millimeter-wave radar used are difficult to match with those of LiDAR. Su et al. [25] proposed the LiDAR-based sensor fusion SLAM for ground robots on complex terrain (GR-LOAM) algorithm, which achieves high precision and robustness in positioning and mapping for ground robots on complex terrain by fusing LiDAR, IMU, and encoder data. This method has high demands for real-time performance and accuracy, but there is no in-depth study on the road elevation information. Shan et al. [26] proposed the Tightly Coupled LiDAR Inertial Odometry via Smoothing and Mapping (LIO-SAM) algorithm, which is a tightly coupled SLAM algorithm based on LiDAR and IMU and an extension of LeGO-LOAM. This framework achieves tightly coupled LiDAR-inertial odometry through smoothing and mapping by optimizing the LiDAR odometry factors, IMU pre-integration factors, GPS factors, and loop-closure detection factors to obtain globally consistent pose estimates, thus improving the system accuracy and robustness. In summary, for extracting road elevation information in complex terrain, road perception systems can effectively reduce extraction errors and improve the system accuracy and robustness not only through multi-sensor fusion methods, but also through improved ground point extraction algorithms. Based on the issues identified in the above research, this paper selects high-precision LiDAR to reduce measurement errors in road elevation information systems. Additionally, an improved framework for road point cloud mapping is proposed to obtain high-precision road point cloud information. On this basis, a moving average-like algorithm is proposed to balance both accuracy and efficiency, ensuring that the extracted road elevation information is more precise and effective.
The main contributions of this paper are summarized as follows:
(1)
A terrain construction system based on the fusion of IMU and LiDAR perception is established. This system includes the architecture for state estimation and mapping, IMU pre-integration factors, LiDAR odometry factors, and loop-closure detection factors.
(2)
Considering the large computational load of the original architecture, this paper improves the existing terrain construction system by adding a new functional module to the algorithm framework. This module saves each point cloud frame with complete road surface details and stitches multiple point clouds together to generate a point cloud file containing the complete road surface information.
(3)
A new moving average-like algorithm is proposed to address the challenges in extracting road elevation information. By comparing the extraction effects of sliding windows with different widths, the optimal window width is determined, thereby improving the accuracy of road elevation information extraction.
The remainder of the paper is organized as follows: Section 2 describes the configuration of the sensor coordinate systems and completes the parameter calibration between the LiDAR and IMU. Section 3 builds a terrain construction model based on the fusion of IMU and LiDAR perception. Building on this model, improvements are made to the state estimation and mapping architecture, and terrain construction experiments are conducted at a school facility. Section 4 utilizes the point cloud library to extract point cloud information about the front tire trajectory, and designs moving average-like algorithm to determine the optimal sliding window width for extraction effectiveness. By comparing it with the Gaussian filter algorithm, the superiority of the moving average-like algorithm is validated. Section 5 concludes the paper.

2. Sensor Coordinate Systems and Parameter Calibration

2.1. Sensor Coordinate System Configuration

The terrain construction system contains two main coordinate systems: the LiDAR coordinate system and the IMU coordinate system. Figure 1 illustrates the relationship between the LiDAR coordinate system and the IMU coordinate system.
(1)
LiDAR coordinate system
Taking the geometric center of the LiDAR device as the origin, a right-handed coordinate system is employed where the vehicle’s longitudinal forward direction is the x-axis, the horizontal left direction is the y-axis, and the vertical upward direction is the z-axis. The coordinate system is denoted by l, with the axes labeled as O l l x l y l z .
(2)
IMU coordinate system
The IMU coordinate system is considered the body coordinate system, using a right-handed coordinate system where the vehicle’s longitudinal backward direction is the x-axis, the horizontal left direction is the y-axis, and the vertical downward direction is the z-axis. The coordinate system is denoted by b, with the axes labeled as O b b x b y b z .

2.2. Sensor Parameters Calibration

The calibration between the LiDAR and the IMU is aimed at eliminating errors and discrepancies between them. Through calibration, targets in the LiDAR coordinate system can be transformed into the corresponding body coordinate system of the IMU, ensuring the accuracy and consistency of measurement results [27]. Specifically, the following objectives can be achieved:
(1)
Coordinate System Alignment
LiDAR and IMU each have their own coordinate systems. Through calibration, the transformation relationship between them can be determined, thereby allowing their measurement results to be compared and analyzed in the same coordinate system.
(2)
Angle Compensation
The measurement results of the IMU may be subject to interference from factors such as vibrations and acceleration, leading to errors in angle measurement. By calibrating with the LiDAR, the angular measurement errors of the IMU can be accurately compensated, thereby improving the accuracy of attitude estimation.
(3)
Position Calibration
There may be minor installation errors in the positions of the LiDAR and IMU, leading to slight discrepancies in their spatial locations. Calibration can determine the positional offsets between them, allowing for the adjustment of their location data and thereby improving the accuracy of positioning.
The relationship between the LiDAR and IMU is solved through the loosely coupled approach using IMU pre-integration and LiDAR odometry, as illustrated in Figure 2. For two keyframes k and k + 1, the rotation matrix obtained from the IMU pre-integrating between the frames is q b k + 1 b k , while the relative rotation computed from the LiDAR odometry is q l k + 1 l k . Since both describe the same rotation, the following constraint relationship [28] holds:
q b k + 1 b k q l b = q l b q l k + 1 l k
To facilitate calculations, this paper introduces the left and right multiplication matrices for quaternions. For a quaternion q = s ,   v T in Hamilton form, the following matrices are used:
q L = s v T v s I + v x , q R = s v T v s I v x
where s is the scalar part and v is the vector part of the quaternion.
To convert the quaternion representation into matrix operations, the constraints can be expressed as follows:
q b k + 1 b k L q l k + 1 l k R q l b = 0
Therefore, a set of linear equations can be constructed using a group of keyframes, typically choosing 10 to 20 keyframes to form an overdetermined system. By solving the least squares problem, the extrinsic rotation parameters represented by quaternions can be obtained. Since obtaining precise results for translation vectors is challenging, manual measurement is often used to determine them.

3. Ground Construction Model Based on Sensor Fusion

3.1. State Estimation and Mapping Architecture

First, we define the coordinate systems and symbols used for vehicle state estimation and mapping. The coordinate system w is defined as the world coordinate system, constructed with the origin of the coordinate system b from the first set of keyframes received as the starting point. The vehicle’s driving state will be represented in the world coordinate system w. The IMU is closely connected to the vehicle, so the coordinate system of the IMU perfectly overlaps with the vehicle’s coordinate system. Therefore, the transformation matrix between them is the identity matrix. The system state x of the vehicle can be expressed as follows:
x = R T , p T , v T , b T T
where R is the rotation matrix, p is the position vector, v is the velocity, b is the IMU bias, and T is the transpose. The transformation matrix T represents the transformation from the vehicle coordinate system to the global coordinate system, given by
T = R | P
The overall architecture of state estimation is shown in Figure 3.
The system receives sensor data from LiDAR and IMU and estimates the vehicle’s state based on the observations from these sensors. This paper uses a factor graph to model the problem, as it is better suited for inference compared to Bayesian networks. Under the assumption of Gaussian noise models, the overall architecture can be represented using the following nonlinear least squares optimization, aiming to minimize the total residual of the system by solving for the state variables:
X = arg m i n X F ( z , X )
F z , X = k B r b ( z b k + 1 b k , X ) b k b k + 1 2 + k L r l ( z l k + 1 l k , X ) l k l k + 1 2
where X denotes all the state variables to be optimized in the factor graph, and z ^ represent all the observations. r is the residual computation function for the corresponding constraints, and Σ 2 is the vector norm calculated using the covariance corresponding to the constraints. B and L represent the constraint observations from the IMU and LiDAR, respectively.

3.2. IMU Pre-Integration Factor

The measurements of IMU acceleration and angular velocity are defined as follows:
a m b = R w b a w g w + b a + n a
ω m b = ω b + b g + n g
where a m b and ω m b are the raw acceleration and angular velocity data measured by the IMU in the body coordinate system, respectively. g w denotes the gravitational acceleration in the world coordinate system. n a and n g are the acceleration noise and angular velocity noise of the IMU during measurement, respectively. b a and b g are the acceleration bias and angular velocity bias of the IMU, respectively. R w b is the rotation matrix from the world coordinate system to the body coordinate system.
After optimizing the state variables, the system predicts the current latest state using the data obtained from the IMU measurements. In this paper, the state variables at the k-th keyframe are defined as x b k w . The IMU model is expressed as follows:
p b k + 1 w = p b k w + v b k w Δ t + 1 2 g w Δ t 2 + R b k w t k t k + 1 R b t b k a m t b b a t n a d t 2
v b k + 1 w = v b k w + g w Δ t + R b k w t k t k + 1 R b t b k a m t b b a t n a d t
q b k + 1 w = q b k w t k t k + 1 q b t b k ω m t b b g t n g d t
Using the model from the above equation, the velocity, position, and attitude of the body at the k + 1-th frame can be estimated based on the velocity, position, and attitude of the body at the k-th frame. Since the time difference Δt between the two keyframes is very small, the acceleration bias and angular velocity bias can be neglected.
For computational convenience, this paper uses the midpoint method to discretize the integration over continuous time. Specifically, the state propagation is carried out by averaging the IMU data at time i and the IMU data at the adjacent time j.
a ¯ = 1 2 R b i w a m i b b a + R b j w a m j b b a + g w
ω ¯ = 1 2 ω m i b + ω m j b b g
where a ¯ and ω ¯ represent the average of the IMU data at time i and the IMU data at the adjacent time j. Substituting these values into the continuous-time IMU model yields the discretized IMU motion model as follows:
p b j w = p b i w + v b i w Δ t + 1 2 a ¯ δ t 2
v b j w = v b i w + a ¯ δ t
q b j w = q b i w 1 1 2 ω ¯ δ t
To address the real-time issues associated with re-integration when updating state variables after optimization, this paper adopts a pre-integration approach. This method treats the IMU data between adjacent keyframes as an independent factor, unrelated to changes in other state variables, and only dependent on the acceleration and angular velocity biases of the IMU. This achieves decoupling from the state variables. By doing so, it avoids the time delay caused by re-integration after updating the state variables, improving the system’s real-time performance. By constructing independent IMU factors, state updates and integration processes can be handled more flexibly, making the system more stable and efficient during operation. Decoupling the state variable x b k w from the continuous-time IMU model yields
Δ p b k + 1 b k = t k t k + 1 R b t b k a m t b b a t n a d t 2
Δ v b k + 1 b k = t k t k + 1 R b t b k a m t b b a t n a d t
Δ q b k + 1 b k = t k t k + 1 q b t b k ω m t b b g t n g d t
where Δ p b k + 1 b k , Δ v b k + 1 b k , and Δ q b k + 1 b k are the pre-integrated quantities for position, velocity, and attitude, respectively, between the k-th frame and k + 1-th frame. At this point, because the pre-integrated quantities are in the body coordinate system, they do not change due to updates in the state variables. After discretizing the pre-integrated IMU quantities using the midpoint method, the pre-integration of the IMU at time i and the adjacent time j is given by
Δ p b j b k = Δ p b i b k + Δ v b i b k δ t + 1 2 a ¯ δ t 2
Δ v b j b k = Δ v b i b k + a ¯ δ t
Δ q b j b k = Δ q b i b k 1 1 2 ω ¯ δ t
where
a ¯ = 1 2 R b i b k a m i b b a + R b j b k a m j b b a
ω ¯ = 1 2 ω m i b + ω m j b b g

3.3. LiDAR Odometer Factor

When a new LiDAR frame is received, feature extraction is first performed. Edge and plane features are extracted by evaluating the roughness of points in a local region. Points with higher roughness are classified as edge features, while points with lower roughness are classified as plane features. At time i, the edge and plane features extracted from the LiDAR scan are defined as F i e and F i p , respectively. All the features extracted at time i constitute the LiDAR frame F i , which is located in the body coordinate system, where
F i = F i e , F i p
To avoid excessive computational load from adding every LiDAR frame to the factor graph, this paper uses a keyframe approach. When the vehicle’s pose changes exceed a predetermined threshold (displacement: 0.8 m, angle: 10°), F i + 1 is selected as a keyframe. The newly saved keyframe, F i + 1 , is associated with the new vehicle state node in the factor graph, while LiDAR frames between two keyframes are ignored. This approach not only balances map density and memory consumption, but also helps maintain a relatively sparse factor graph. Suppose a new state node, x i + 1 , is added to the factor graph, and its corresponding LiDAR keyframe is F i + 1 . The steps to generate the LiDAR odometry factor are as follows:
Firstly, a point cloud map containing a fixed number of recent LiDAR scans is constructed, and n latest keyframes, referred to as sub-keyframes, are extracted for estimation. Then, using the transformation matrices T i n , , T i associated with the set of sub-keyframes F i n , , F i , the sub-keyframes are transformed into the world coordinate system. The transformed sub-keyframes are merged to form a voxel map, M i . Since two types of features, namely, edge features and plane features, were extracted in the previous feature extraction step, the voxel map M i consists of two sub-voxel maps, denoted as the edge feature voxel map, M i e , and the plane feature voxel map, M i p . The relationship between the LiDAR frame and the voxel map is expressed as follows:
M i = M i e , M i p
where
M i e = F i e F i 1 e F i n e
M i p = F i p F i 1 p F i n p
where F i e and F i p are edge features and plane features in the world coordinate system, respectively.
Next, for the newly received LiDAR frame, F i + 1 , we match the scan with the voxel map M i . We then transform the edge and plane features F i + 1 e and F i + 1 p extracted from the LiDAR scan from the body coordinate system to the world coordinate system to obtain F i + 1 e and F i + 1 p , where the transformation matrix is T ~ i + 1 from the IMU. Then, we find the corresponding edge and plane relationships in M i e and M i p with the features in F i + 1 e and F i + 1 p . The distance between a feature and its corresponding edge or plane can be calculated using the following formula:
d e k = | ( p i + 1 , k e p i , u e ) × ( p i + 1 , k e p i , v e ) | | p i , u e p i , v e |
d p k = ( p i + 1 , k p p i , u p ) ( p i , u p p i , v p ) × ( p i , u p p i , w p ) | ( p i , u p p i , v p ) × ( p i , u p p i , w p ) |
where k, u, v, and w are the indices of the feature point correspondences. p i + 1 e is an edge feature point in F i + 1 e , while p i , u e and p i , v e are two points on the edge line corresponding to this edge feature point in M i . For a plane feature point in F i + 1 p , the corresponding plane patch in M i is defined by three points: p i , u p , p i , v p , and p i , w p .
Finally, we use the Gauss–Newton (G-N) method to solve the following residual minimization problem to obtain the optimal transformation matrix [29]:
min T i + 1 p i + 1 , k e F i + 1 e d e k + p i + 1 , k p F i + 1 p d p k
Subsequently, the relative transformation Δ T i , i + 1 between x i and x i + 1 can be obtained using the following equation, which represents the LiDAR odometry factor connecting these two poses:
Δ T i , i + 1 = T i T T i + 1

3.4. Loop-Closure Factor

In this paper, a loop-closure detection method [30,31] based on the Euclidean distance is used, where the index m is set to 15, and the search distance for the loop closure from the new state x i + 1 is set to 20 m. When a new state, x i + 1 , is added to the factor graph, the graph is first searched and previous states close to x i + 1 in the Euclidean space are found. For example, for x 5 , the submap F i + 1 is first transformed into the world coordinate system along with past keyframes, and then F i + 1 is scanned and matched with the keyframes F 5 m , , F 5 , , F 5 + m . Finally, the relative transformation, Δ T 5 , i + 1 , is added to the factor graph as a loop-closure factor.

3.5. Terrain Construction Experiment Based on an Improved Architecture

To verify the effectiveness and accuracy of the terrain construction system, this paper established a Chery experimental vehicle platform equipped with an IMU and LiDAR sensors. The RS-Ruby Lite 80-line LiDAR was used to collect environmental point cloud data, while the 3DM-GX5-AHRS IMU provided motion and attitude information for the vehicle such as acceleration, angular velocity, and heading angle. A segment of a straight road within the campus featuring a speed bump was selected. The vehicle traveled along this road at a constant speed. During the vehicle’s movement, data collected using various sensors were input to the ECU and used for state estimation and mapping on the ROS platform. Figure 4 shows the road with the speed bump on the campus, and Figure 5 presents the point cloud map of the road with the speed bump.
From Figure 5, basic road surface information can be obtained, with the red circle highlighting the point cloud information of the speed bump. Since the architecture does not require detailed road surface information, a significant number of “irrelevant” planar points were filtered out during the mapping process to reduce data processing load. This, however, is clearly detrimental to the generation of road surface information.
To obtain detailed road surface point cloud data, this paper introduces a new functional module into the algorithmic architecture. This module saves each point cloud file with complete road surface details and stitches together multiple point clouds to create a single point cloud file with comprehensive road surface information. Basis on this, road surface elevation data were obtained. Figure 6 shows the point cloud map of the road segment with a speed bump generated using the improved architecture, with the red circle highlighting the point cloud information of the speed bump.
By comparing Figure 5 and Figure 6, we can clearly observe that the improved architecture generates a larger number of point clouds. This enhancement not only excels on flat road surfaces, but also achieves a refined representation of speed bumps. This signifies that we could meet the precision requirements necessary for extracting road information. Based on this, formal experiments can be further conducted to verify the feasibility and effectiveness of this improved architecture. Figure 7 illustrates the uneven road surface within the campus used for the formal experiment, where the driver traveled in a straight line along the road at a certain speed to perform terrain construction. Figure 8 shows the point cloud map generated using the original architecture.
As shown in Figure 8, the road surface information is presented at a basic level. However, the point cloud accuracy is insufficient for the extraction of road information addressed in this paper; hence, further experiments were conducted using the improved architecture. Figure 9 displays the point cloud map generated using the improved architecture.
As observed in Figure 9, the number of point clouds generated by the improved architecture was significantly greater than that of the original architecture. The point cloud count increased from 3,501,510 to 45,551,861, providing a complete depiction of the road surface details, which facilitated the extraction of road information.

4. Road Information Extraction Model Based on Point Cloud Library and Moving Average-like Algorithm

4.1. Extraction of Point Cloud Information for Front Tire Trajectories

In point cloud map processing, extracting point cloud information for the front tire trajectory positions is crucial for determining the road surface elevation where the wheels are about to pass. To achieve this, this paper employed the Passthrough algorithm [32,33] from the PCL (point cloud library) to filter the point cloud information corresponding to the front tire trajectory positions. The Passthrough algorithm is a commonly used point cloud processing method that filters point cloud data based on specific conditions to extract data within a certain range. Figure 10 illustrates the effect of the Passthrough algorithm.
The process of the Passthrough algorithm is illustrated in Figure 11. The algorithm filters point cloud data by setting a range, retaining only the point cloud data that meet the specified conditions, and discarding the data that do not. Taking point cloud data in a three-dimensional space as an example, the Passthrough algorithm can impose constraints on the coordinates of points along the x, y, and z axes separately. By setting minimum and maximum values for each axis, a cubic or rectangular range is specified to filter the required point cloud data. This filtering method efficiently extracts the point cloud information corresponding to the front tire trajectory positions.

4.2. Moving Average-like Algorithm

The moving average algorithm [34] is a commonly used signal processing technique for smoothing time series data. However, the results of the moving average algorithm are influenced by historical records, whereas road information is objective and historical road data should not affect the generation of the next frame of road information. Therefore, it is necessary to eliminate the impact of historical factors in the algorithm design and develop a moving average-like algorithm suitable for road elevation information extraction. When selecting the width of the sliding window, it should be based on the tire contact area, ensuring it is neither too large nor too small, balancing both precision and efficiency. This study compared moving average-like algorithms with sliding window widths of 4 cm, 6 cm, 8 cm, and 10 cm to choose the width that provides the best results as the final algorithm.
Using a sliding window width of 4 cm as an example, the detailed algorithm design process is explained as follows:
(1)
Initial grid region creation: Starting from the longitudinal distance x = 0 m, first draw a grid area with a length of 0.04 m (in the x direction) and a width of 0.205 m (in the y direction), with no restriction in the z direction. In this grid, x ( 0 ,   0.04 ) m. Then, move forward by 0.01 m and draw another grid area with the same dimensions: length 0.04 m (in the x direction) and width 0.205 m (in the y direction), with no restriction in the z direction. For this grid, x ∈ (0.01, 0.05) m. Repeat this process until you reach the final grid area with a length of 0.04 m (in the x direction), and a width of 0.205 m (in the y direction), with no restriction in the z direction. For this final grid, x ( 19.96 ,   20 ) m. In the end, a total of 1997 grid areas, each 4 cm in width, will be obtained.
(2)
Point assignment to grid regions: Iterate through the points and place each point into the corresponding grid areas based on its x value. A single point may fall into multiple grid areas. For example, a point with an x value of 3 cm will be included in the grid areas spanning 0~4 cm, 1~5 cm, 2~6 cm, and 3~7 cm. Therefore, this point needs to be placed into all of these overlapping grid areas simultaneously.
(3)
Data storage: Create a vector container to store the x, y, and z data of the points for each grid region.
(4)
Calculate averages: In a single grid, calculate the average x value and the average z value of all points within the grid. These averages represent the x and z values for that particular grid.
(5)
Generate elevation information: Perform the same operation for the other grids, iterating through all grids. This will yield 1997 points with x values as the horizontal coordinates and z values as the vertical coordinates. By plotting these points and connecting them on the x–z coordinate plane, you will obtain the elevation information of the road surface.

4.3. Results and Discussion

Based on the terrain construction experiment, when the vehicle travels in a straight line, the trajectory equation is as follows:
x 0 m , 20 m
y [ 0.88 m , 0.675 m ]
Processing the complete point cloud map according to the preset trajectory equation, we successfully filtered out the point cloud information of the front tire trajectory position through the Passthrough algorithm. Figure 12 shows the top view of the point cloud at the front tire trajectory position after processing.
This filtered point cloud data provided essential support for the further analysis of the road surface elevation at the wheel’s path, contributing to enhancing the accuracy and reliability of road condition assessments.
Next, this paper used the moving average-like algorithm and Gaussian filter algorithm [35] to extract surface elevation information from the selected point cloud data, thereby validating the superiority of the moving average-like algorithm in extracting surface elevation information.
Figure 13 shows the road surface elevation information generated using the moving average-like algorithm with sliding window widths of 4 cm, 6 cm, 8 cm, and 10 cm. From Figure 13, it can be seen that the moving average-like algorithm preserves some high-frequency variation details, preventing the loss of important data. Additionally, as the width of the sliding window decreases, the details of the road elevation information are better presented. This algorithm optimizes the smoothing effect by using sliding windows of different widths, avoiding excessive smoothing or distortion of the data.
Generally, the larger the window width, the stronger the smoothing effect, but it also results in the loss of more local details, which reduces the consistency or accuracy of the data. A larger window averages out changes over a broader range, potentially masking smaller but important variations in the data. Additionally, for highly dynamic or rapidly changing scenarios, a larger window may lead to excessive smoothing, thus overlooking important local fluctuations or features. Therefore, reducing the sliding window width can yield more accurate road elevation information. However, a smaller window is not always better: if the window is too small, the algorithm may fail to effectively remove the noise, leading to chaotic and unreliable results. In this paper, a 4 cm sliding window was chosen, as it effectively preserved enough road detail while eliminating the interference caused by noise.
Figure 14 shows the road surface elevation information generated using the Gaussian filter algorithm. The algorithm effectively smooths the data, removes noise, and retains the overall trend of the data, making it suitable for most smoothing needs. However, by comparing it with the ground truth road elevation data, it can be observed that the algorithm may cause excessive smoothing at the edges and in detailed areas, leading to information loss, especially in the case of high-frequency information, which makes the extracted road elevation data less accurate.
After summarizing the information from Figure 13 and Figure 14 into Table 1, it can be observed that the road surface elevation data generated using the moving average algorithm with a sliding window width of 4 cm was closest to the actual road surface, achieving a congruence of 97.6%. Therefore, the road surface elevation information with the sliding window width of 4 cm, which showed the best results, was selected for the final analysis.

5. Conclusions

In response to challenges related to the difficulty of obtaining elevation information for the road ahead of a vehicle, this paper proposes a digital terrain construction scheme based on the fusion perception of IMU and LiDAR. This scheme establishes a terrain construction system using the fusion of IMU and LiDAR, and introduces improvements to the original framework to extract more detailed road surface point cloud information. Based on the output information from the terrain construction system, the moving average-like algorithm is designed to process the point cloud data and extract the road surface elevation information at the vehicle’s front tire trajectory. By comparing different sliding window width algorithm schemes, the optimal width was selected to ensure that the extracted road surface elevation information is more accurate and effective.

Author Contributions

Conceptualization, C.H. and Y.W.; methodology, C.H.; software, Y.W.; validation, C.H. and Y.W.; formal analysis, C.H., Y.W. and X.S.; investigation, S.Y.; resources, X.S.; data curation, S.Y.; writing—original draft preparation, Y.W.; writing—review and editing, C.H.; visualization, S.Y.; supervision, X.S.; project administration, C.H.; funding acquisition, C.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the General Program of the National Natural Science Foundation of China under grant number 52472432 and the International Cooperation Project (BZ2022050).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data are contained within the article.

Acknowledgments

Thank you to the National Key Laboratory for Intelligent and Green Vehicles and Transportation of Tsinghua University for their support of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, B.; Wang, Y.; Papaioannou, G.; Du, H. Sensor Fusion and Advanced Controller for Connected and Automated Vehicles. Sensors 2023, 23, 7015. [Google Scholar] [CrossRef]
  2. Hasanujjaman, M.; Chowdhury, M.Z.; Jang, Y.M. Sensor Fusion in Autonomous Vehicle with Traffic Surveillance Camera System: Detection, Localization, and AI Networking. Sensors 2023, 23, 3335. [Google Scholar] [CrossRef] [PubMed]
  3. Ahangari Sisi, Z.; Mirzaei, M.; Rafatnia, S. Estimation of vehicle suspension dynamics with data fusion for correcting measurement errors. Measurement 2024, 231, 114438. [Google Scholar] [CrossRef]
  4. Zhang, Z.H.; Zhao, J.T.; Huang, C.Y.; Li, L. Learning visual semantic map-matching for loosely multi-sensor fusion localization of autonomous vehicles. IEEE Trans. Intell. Vehicles 2023, 8, 358–367. [Google Scholar] [CrossRef]
  5. Sun, Y.; Luo, Y.; Zhang, Q.; Xu, L.; Wang, L.; Zhang, P. Estimation of Crop Height Distribution for Mature Rice Based on a Moving Surface and 3D Point Cloud Elevation. Agronomy 2022, 12, 836. [Google Scholar] [CrossRef]
  6. Almalki, F.A.; Angelides, M.C. Autonomous flying IoT: A synergy of machine learning, digital elevation, and 3D structure change detection. Comput. Commun. 2022, 190, 154–165. [Google Scholar] [CrossRef]
  7. Moura Coelho, R.; Gouveia, J.; Botto, M.A.; Krebs, H.I.; Martins, J. Real-time walking gait terrain classification from foot-mounted Inertial Measurement Unit using Convolutional Long Short-Term Memory neural network. Expert Syst. Appl. 2022, 203, 117306. [Google Scholar] [CrossRef]
  8. Chen, Y.X.; Chen, L.; Huang, C.; Lu, Y.; Wang, C. A dynamic tire model based on HPSO-SVM. Int. J. Agric. Biol. Eng. 2019, 12, 36–41. [Google Scholar] [CrossRef]
  9. Song, F.; Yang, Z.; Gao, X.; Dan, T.; Yang, Y.; Zhao, W.; Yu, R. Multi-scale feature-based land cover change detection in mountainous terrain using multi-temporal and multi-sensor remote sensing images. IEEE Access 2018, 6, 77494–77508. [Google Scholar] [CrossRef]
  10. Klein, G.; Murray, D. Parallel tracking and mapping for small and workspaces. In Proceedings of the 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, Nara, Japan, 13–16 November 2007; IEEE: Piscataway, NJ, USA, 2007; pp. 225–234. [Google Scholar]
  11. Raul, M.; Jose, M.; Juan, D.T. ORB-SLAM: A versatile and accurate monocular SLAM system. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar]
  12. Gao, Y.; Li, X.; Wang, X.V.; Wang, L.; Gao, L. A Review on Recent Advances in Vision-based Defect Recognition towards Industrial Intelligence. J. Manuf. Syst. 2022, 62, 753–766. [Google Scholar] [CrossRef]
  13. Zhang, Y.Y.; Zhang, B.; Shen, C.; Liu, H.L.; Huang, J.C.; Tian, K.P.; Tang, Z. Review of the field environmental sensing methods based on multi-sensor information fusion technology. Int. J. Agric. Bio-Log. Eng. 2024, 17, 1–13. [Google Scholar]
  14. Cai, C.; Zou, Y.; Pan, Z.; Liu, Z.; Gao, S. A novel lane detection algorithm based on multi-feature fusion and window search. J. Jiangsu Univ. 2023, 44, 386–391. [Google Scholar]
  15. Zhou, X.; Sun, J.; Tian, Y.; Wu, X.H.; Dai, C.X.; Li, B. Spectral classification of lettuce cadmium stress based on in-formation fusion and VISSA-GOA-SVM algorithm. J. Food Process Eng. 2019, 42, e13085. [Google Scholar] [CrossRef]
  16. Zhang, J.; Singh, S. LOAM: Lidar Odometry and Mapping in Real-time. Robot. Sci. Syst. 2014, X, 007. [Google Scholar] [CrossRef]
  17. Shan, T.; Englot, B. LeGO-LOAM: Lightweight and Ground-Optimized Lidar Odometry and Mapping on Variable Terrain. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4758–4765. [Google Scholar]
  18. Lin, J.; Zhang, F. A fast, complete, point cloud based loop closure for LiDAR odometry and mapping. arXiv 2019, arXiv:1909.11811. [Google Scholar]
  19. Wang, H.; Wang, C.; Xie, L. Lightweight 3-D Localization and Mapping for Solid-State LiDAR. IEEE Robot. Autom. Lett. 2021, 6, 1801–1807. [Google Scholar] [CrossRef]
  20. Xu, H. Semi-supervised manifold learning based on polynomial mapping for localization in wireless sensor networks. Signal Process. 2020, 172, 107570.1–107570.11. [Google Scholar] [CrossRef]
  21. Droeschel, D.; Schwarz, M.; Behnke, S. Continuous mapping and localization for autonomous navigation in rough terrain using a 3D laser scanner. Robot. Auton. Syst. 2017, 88, 104–115. [Google Scholar] [CrossRef]
  22. Xu, J.; Liu, H.; Shen, Y.; Zeng, X.; Zheng, X. Individual nursery trees classification and segmentation using a point cloud-based neural network with dense connection pattern. Sci. Hortic. 2024, 328, 112945. [Google Scholar] [CrossRef]
  23. Nguyen-Ngoc, T.-T.; Phi, T.-D.; Phan-Nguyen, P.-Q.; Nguyen, V.-H. Tightly-coupled GPS/INS/Lidar Integration for Road Vehicles. In Proceedings of the 2023 International Symposium on Electrical and Electronics Engineering (ISEEE 2023), Galati, Romania, 26–28 October 2023; pp. 156–160. [Google Scholar]
  24. Almalioglu, Y.; Turan, M.; Lu, C.X.; Trigoni, N.; Markham, A. Milli-RIO: Ego-Motion Estimation with Low-Cost MillimetreWave Radar. IEEE Sens. J. 2021, 21, 3314–3323. [Google Scholar] [CrossRef]
  25. Su, Y.; Wang, T.; Shao, S.; Yao, C.; Wang, Z. GR-LOAM: LiDAR-based sensor fusion SLAM for ground robots on complex terrain. Robot. Auton. Syst. 2021, 140, 103759. [Google Scholar] [CrossRef]
  26. Shan, T.; Englot, B.; Meyers, D.; Wang, W.; Ratti, C.; Rus, D. LIO-SAM: Tightly-coupled Lidar Inertial Odometry via Smoothing and Mapping. arXiv 2020, arXiv:2007.00258. [Google Scholar]
  27. Li, S.Q.; Sun, X.; Zhang, M.A. Vehicle recognition technology at urban intersections based on the fusion of LiDAR and cameras. J. Jiangsu Univ. 2024, 45, 621–628. [Google Scholar]
  28. Li, Y.; Yang, S.; Xiu, X.; Miao, Z. A Spatiotemporal Calibration Algorithm for IMU–LiDAR Navigation System Based on Similarity of Motion Trajectories. Sensors 2022, 22, 7637. [Google Scholar] [CrossRef]
  29. Li, C.X.; Jiang, H.B.; Wang, C.Y.; Ma, S.D. Vehicle pose estimation method based on sensor information fusion. J. Jiangsu Univ. 2022, 43, 636–644. [Google Scholar]
  30. Yang, N.; Cheng, L. SLAM algorithm for large scenes based on efficient loopback detection. J. Shenyang Univ. Technol. 2024, 43, 45–51. [Google Scholar]
  31. Chen, W.; Shang, G.; Ji, A.; Zhou, C.; Wang, X.; Xu, C.; Li, Z.; Hu, K. An Overview on Visual SLAM: From Tradition to Semantic. Remote Sens. 2022, 14, 3010. [Google Scholar] [CrossRef]
  32. Rusu, R.B.; Cousins, S. 3D is here: Point Cloud Library (PCL). In Proceedings of the 2011 IEEE International Conference on Robotics and Automation (ICRA), Shanghai, China, 9–13 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 1–4. [Google Scholar]
  33. Cao, B.; Mendoza, R.C.; Philipp, A.; Gohring, D. LiDAR-based object-level SLAM for autonomous vehicles. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 4397–4404. [Google Scholar]
  34. Wang, C.; Li, Y. Road surface modeling using LiDAR data: A method based on moving average filter. J. Geogr. Inf. Sci. 2018, 34, 1113–1123. [Google Scholar]
  35. Zhan, Q.; Liang, Y.; Cai, Y.; Xiao, Y. Pattern Detection in Airborne LiDAR Data Using Laplacian of Gaussian Filter. Geo-Spat. Inf. Sci. 2011, 14, 184–189. [Google Scholar] [CrossRef]
Figure 1. LiDAR coordinate system and IMU coordinate system.
Figure 1. LiDAR coordinate system and IMU coordinate system.
Sensors 25 00015 g001
Figure 2. LiDAR and IMU calibration diagram.
Figure 2. LiDAR and IMU calibration diagram.
Sensors 25 00015 g002
Figure 3. Overall architecture of state estimation.
Figure 3. Overall architecture of state estimation.
Sensors 25 00015 g003
Figure 4. Road segment with speed bump on campus.
Figure 4. Road segment with speed bump on campus.
Sensors 25 00015 g004
Figure 5. Point cloud map of road segment with speed bump.
Figure 5. Point cloud map of road segment with speed bump.
Sensors 25 00015 g005
Figure 6. Point cloud map of road segment with speed bump generated using improved architecture.
Figure 6. Point cloud map of road segment with speed bump generated using improved architecture.
Sensors 25 00015 g006
Figure 7. Uneven road surface used for the formal experiment.
Figure 7. Uneven road surface used for the formal experiment.
Sensors 25 00015 g007
Figure 8. Point cloud map generated using the original architecture.
Figure 8. Point cloud map generated using the original architecture.
Sensors 25 00015 g008
Figure 9. Improved architecture-generated point cloud maps.
Figure 9. Improved architecture-generated point cloud maps.
Sensors 25 00015 g009
Figure 10. Effect of the Passthrough algorithm.
Figure 10. Effect of the Passthrough algorithm.
Sensors 25 00015 g010
Figure 11. Flowchart of the Passthrough algorithm.
Figure 11. Flowchart of the Passthrough algorithm.
Sensors 25 00015 g011
Figure 12. Top view of point cloud data for the front tire trajectory position.
Figure 12. Top view of point cloud data for the front tire trajectory position.
Sensors 25 00015 g012
Figure 13. Road surface elevation information generated using moving average-like algorithm.
Figure 13. Road surface elevation information generated using moving average-like algorithm.
Sensors 25 00015 g013
Figure 14. Road surface elevation information generated using Gaussian filter algorithm.
Figure 14. Road surface elevation information generated using Gaussian filter algorithm.
Sensors 25 00015 g014
Table 1. Algorithm scheme comparison.
Table 1. Algorithm scheme comparison.
Algorithmic ApproachCongruence with the Actual Road Surface
Sliding window width of 4 cm97.6%
Sliding window width of 6 cm90.3%
Sliding window width of 8 cm82.8%
Sliding window width of 10 cm74.5%
Gaussian filter71.4%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, C.; Wang, Y.; Sun, X.; Yang, S. Research on Digital Terrain Construction Based on IMU and LiDAR Fusion Perception. Sensors 2025, 25, 15. https://doi.org/10.3390/s25010015

AMA Style

Huang C, Wang Y, Sun X, Yang S. Research on Digital Terrain Construction Based on IMU and LiDAR Fusion Perception. Sensors. 2025; 25(1):15. https://doi.org/10.3390/s25010015

Chicago/Turabian Style

Huang, Chen, Yiqi Wang, Xiaoqiang Sun, and Shiyue Yang. 2025. "Research on Digital Terrain Construction Based on IMU and LiDAR Fusion Perception" Sensors 25, no. 1: 15. https://doi.org/10.3390/s25010015

APA Style

Huang, C., Wang, Y., Sun, X., & Yang, S. (2025). Research on Digital Terrain Construction Based on IMU and LiDAR Fusion Perception. Sensors, 25(1), 15. https://doi.org/10.3390/s25010015

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop