High-Precision SLAM Based on the Tight Coupling of Dual Lidar Inertial Odometry for Multi-Scene Applications

Featured Application: This paper fuses different sensors to form a general high-precision SLAM framework for multi-scene applications. The algorithm framework in this paper can be extended to the ﬁelds of autonomous driving, robot navigation, and 3D reconstruction. Abstract: Simultaneous Localization and Mapping (SLAM) is an essential feature in many applications of mobile vehicles. To solve the problem of poor positioning accuracy, single use of mapping scene, and unclear structural characteristics in indoor and outdoor SLAM, a new framework of tight coupling of dual lidar inertial odometry is proposed in this paper. Firstly, through external calibration and an adaptive timestamp synchronization algorithm, the horizontal and vertical lidar data are fused, which compensates for the narrow vertical ﬁeld of view (FOV) of the lidar and makes the characteristics of vertical direction more complete in the mapping process. Secondly, the dual lidar data is tightly coupled with an Inertial Measurement Unit (IMU) to eliminate the motion distortion of the dual lidar odometry. Then, the value of the lidar odometry after correcting distortion and the pre-integrated value of IMU are used as constraints to establish a non-linear least-squares objective function. Joint optimization is then performed to obtain the best value of the IMU state values, which will be used to predict the state of IMU at the next time step. Finally, experimental results are presented to verify the effectiveness of the proposed method.


Introduction
Simultaneous localization and mapping (SLAM) require building a map of an unknown environment by a mobile vehicle and simultaneously localizing the vehicle in such a map [1][2][3]. SLAM is essential for vehicles to fulfill many tasks, including vehicle rescue [4] and exploration [5]. The perception of the unknown external environment by various onboard sensors provides vital information for SLAM. Thus, integrating different sensors to develop a practical SLAM framework that can be applied in multiple scenes is essential.
Generally, both vision-based and lidar-based SLAM [6][7][8][9][10] are used. Although the vision-based SLAM can obtain high-precision positioning [11][12][13], the vision sensors are vulnerable to light change in the environment and cannot work in dark or untextured scenes. In comparison, lidar is not affected by light and can usually measure the angle and distance of obstacles with higher accuracy. Therefore, this paper focuses on the design of lidar-based SLAM to adapt the multi-scene applications.
In recent years, many different lidar-based SLAM schemes have been proposed. Among them, the Lidar Odometry and Mapping in Real-time (LOAM) method, where a single-line lidar and a motor are used to form a multiline lidar to realize low-drift and a complex mechanical structure, the localization accuracy is relatively low [28,29]. More recently, a new design for a 3D sensor system was introduced [30], involving construction from a 2D range scanner coupled with a passive linkage mechanism, such as a spring. However, the system requires reasonably continual excitation to induce motion of the sensor head. Therefore, the system is not considered appropriate for electric ground vehicles operating with infrequent accelerations on smooth terrain. In order to solve the above problems, some researchers have started investigating the use of multiple 3D lidars for better coverage of the environment. In [31], a high-precision lidar odometry system was proposed to achieve robust and real-time operation under challenging perceptual conditions. In this system, the two lidars are mounted at 180 degrees to each other to make up for the self-similar areas with low lidar observability. The authors of [32] proposed a system to achieve robust and simultaneous extrinsic calibration, odometry, and mapping for multiple lidars, while [33] proposed a scheme to combine multiple lidars with complementary FOV for feature-based lidar-inertia odometry and mapping. While the above methods can work well in single-scene applications, their adaptability to different environments was not considered, especially the environments that need to perceive height information.
This paper proposes a high-precision SLAM framework based on tight coupling of dual lidar inertial to overcome the above problems. In this framework, the horizontal lidar and the vertical lidar are integrated to broaden the FOV. Then the dual lidar and IMU are tightly coupled to improve the localization accuracy. In the outdoor environment, this paper introduces GPS measurements to further improve positioning accuracy. The main contributions of this paper include:

1.
A general high-precision SLAM framework is provided by fusing different sensors. It can adapt to multi-scene applications, such as a corridor with fewer features, stairs with height, and complex outdoor environments.

2.
The horizontal and vertical lidars are fused by external calibration and adaptive time synchronization algorithms to solve the narrow vertical FOV of the lidar.

3.
To improve the positioning accuracy of SLAM in the environment with height information (e.g., stairs), the dual lidar odometry and IMU are tightly coupled. The dual lidar odometry measurement and IMU pre-integration are jointly optimized to obtain more accurate IMU state values, which will be used to eliminate the motion distortion of the dual lidar to improve the localization accuracy. 4.
In addition, several practical experiments are carried out to verify the feasibility and effectiveness of the proposed method.

IMU State Prediction and Pre-Integration
Usually, the frequency of IMU is much higher than lidar. IMU can obtain the acceleration and angular velocity at each time step, which can be used to predict the next state of IMU through integration.
In general, the IMU states at time t can be modeled as: where X W t represents the state in the world frame W at time t; p t , v t , and R t represent the position, velocity, and rotation matrix, respectively, at time t; andâ t andω t are the acceleration and angular velocity of the raw IMU measurements (â t andω t are affected by a slowly varying bias, b t ).
The next state, X W t+∆t , is predicted through the integration of IMU, where ∆t is the interval between two consecutive IMU measurements. The state prediction value of the IMU is then used to infer the motion of the vehicle. The velocity, position, and rotation of the vehicle at time t + ∆t can be computed by Equation (2) [27]: where g is gravitational acceleration; b a t is the varying bias of the accelerationâ t ; and n a t is the white noise of the accelerationâ t . The quaternion q t under Hamilton notation (which corresponds to R t ) is used; ⊗ is used for the multiplication of two quaternions; b ω t is the varying bias of the angular velocityω t ; and n ω t is the white noise of the angular velocityω t . The motion state of IMU between the i-th and the j-th time steps can be represented by the IMU pre-integrated measurement value ∆p ij , ∆v ij , ∆q ij , which can be computed by: where ∆p ij , ∆v ij , ∆q ij is the position, velocity, and rotation matrix of the IMU, respectively, which has the covariance C I i I j in the error-state model. Due to space limitations, we invite readers to refer to the descriptions in [26,34] to understand the detailed derivation of IMU pre-integration.

Segmentation and Feature Extraction
Before extracting features, to filter out the noise, the point cloud is segmented to reduce noise interference. D t = {d 1 , d 2 , . . . , d n } is the point cloud information acquired at time t; d τ is a point in D t ; D t is projected onto a range image; the point d τ represents a pixel of the image; and the pixel value r τ represents the Euclidean distance from the point d τ to the lidar. An image-based segmentation method [17] is applied to the range image to group points into many clusters. Points of the same category have the same label (the ground is a particular category). Features are then extracted from ground points and segmented points, with the corresponding edge points and plane points obtained by calculating the curvature E of each point. To evenly extract features from all directions, the range image is divided horizontally into several equal sub-images. Then, the points are sorted in each row of the sub-image based on their curvature values, E. The curvature E is computed by Equation (4).
where S is the set of continuous points d τ from the same row of the range image; λ is a point in the set S; E th is a threshold which is set to distinguish different types of features; the points are called with E > E th edge features; and the points are called with E < E th planar features.

Hardware System Description
The configuration of our system is shown in Figure 1, with the left model obtained through Unigraphics NX and the designed equipment in our paper shown on the right. The designed equipment is a scanning arm composed of various sensors that can be installed on a backpack or a vehicle. The sensor suite used in the scanning arm includes two 10 HZ Velodyne VLP-16 lidars, a 200 HZ YIS510A IMU, and a 5 HZ HY-269 GPS. The panoramic camera prepares for the follow-up research. The lidar has a horizontal FOV of 360 • with 0.4 • resolution, a vertical FOV 15 • with 2 • resolution, and an accuracy of ±3 cm. The Roll/Pitch of IMU accuracy is 0.25 • , HY-269 GPS is a small-volume positioning and directional receiver, and the positioning accuracy of GPS is 1.2 m.

Hardware System Description
The configuration of our system is shown in Figure 1, with the left model obtained through Unigraphics NX and the designed equipment in our paper shown on the righ The designed equipment is a scanning arm composed of various sensors that can be in stalled on a backpack or a vehicle. The sensor suite used in the scanning arm includes tw 10 HZ Velodyne VLP-16 lidars, a 200 HZ YIS510A IMU, and a 5 HZ HY-269 GPS. Th panoramic camera prepares for the follow-up research. The lidar has a horizontal FOV o 360° with 0.4° resolution, a vertical FOV 15° with 2° resolution, and an accuracy of ±3 cm The Roll/Pitch of IMU accuracy is 0.25°, HY-269 GPS is a small-volume positioning and directional receiver, and the positioning accuracy of GPS is 1.2 m.

System Overview
In this paper, horizontal and vertical lidars are fused to make up for the shortcoming of lidar's narrow FOV. When indoors, accurate positioning information is obtained through the tight coupling of dual lidar inertial odometry. When outdoors, GPS measure ments are added, providing a relatively accurate initial position for positioning, thereb further improving outdoor positioning accuracy. Figure 2 provides a brief overview of our proposed framework. Firstly, after obtain ing the IMU data, the IMU state prediction and pre-integration are updated, as detailed in Section 2.1. IMU pre-integration is used to construct a factor graph constraint, which will be used for joint optimization. Secondly, the dual lidar point clouds are obtained through external calibration and an adaptive timestamp synchronization algorithm, a detailed in Section 3.3. Since lidar has continuous measurement, the lidar measurement i obtained in continuous motion and it is almost certain that motion distortion will occur When dual lidar point cloud is received, deskewing is applied on the dual lidar raw poin cloud to obtain deskewed dual lidar point cloud [14]. To be ground-optimized and reduc the amount of calculation, point cloud segmentation is applied to filter out noise (e.g moving people or objects) and the planar points and edge points are distinguished b

System Overview
In this paper, horizontal and vertical lidars are fused to make up for the shortcomings of lidar's narrow FOV. When indoors, accurate positioning information is obtained through the tight coupling of dual lidar inertial odometry. When outdoors, GPS measurements are added, providing a relatively accurate initial position for positioning, thereby further improving outdoor positioning accuracy. Figure 2 provides a brief overview of our proposed framework. Firstly, after obtaining the IMU data, the IMU state prediction and pre-integration are updated, as detailed in Section 2.1. IMU pre-integration is used to construct a factor graph constraint, which will be used for joint optimization. Secondly, the dual lidar point clouds are obtained through external calibration and an adaptive timestamp synchronization algorithm, as detailed in Section 3.3. Since lidar has continuous measurement, the lidar measurement is obtained in continuous motion and it is almost certain that motion distortion will occur. When dual lidar point cloud is received, deskewing is applied on the dual lidar raw point cloud to obtain deskewed dual lidar point cloud [14]. To be ground-optimized and reduce the amount of calculation, point cloud segmentation is applied to filter out noise (e.g., moving people or objects) and the planar points and edge points are distinguished by calculating the curvature of the lidar point, as detailed in Section 2.2. After feature extraction, the dual lidar point cloud is registered to construct the lidar odometer factor. Then, the sliding window factor graph optimization method of local finite frame is used, with the IMU pre-integration factor and lidar odometer factor jointly optimized to realize the mutual parameter optimization of the dual lidar and IMU. Joint optimization is taken to obtain a maximum a posteriori (MAP) estimation of the IMU states, as detailed in Section 3.4, avoiding the drift from IMU propagation. To further improve outdoor positioning accuracy, GPS is introduced in this framework. When GPS is loosely coupled with IMU through an Unscented Kalman Filter (UKF), GPS measurements are used to further improve the positioning accuracy of IMU, with the optimized IMU state then used for the state estimation of IMU at the next time step. GPS measurements are added, providing a relatively accurate initial position for localization, thereby further improving outdoor localization accuracy [25]. However, the drift of lidar inertial odometry grows very slowly; in practice, GPS measurements are used when the estimated position covariance is larger than the received GPS position covariance. Finally, the output of the figure is the global map and localization. accuracy, GPS is introduced in this framework. When GPS is loosely coupled with IMU through an Unscented Kalman Filter (UKF), GPS measurements are used to further improve the positioning accuracy of IMU, with the optimized IMU state then used for the state estimation of IMU at the next time step. GPS measurements are added, providing a relatively accurate initial position for localization, thereby further improving outdoor localization accuracy [25]. However, the drift of lidar inertial odometry grows very slowly; in practice, GPS measurements are used when the estimated position covariance is larger than the received GPS position covariance. Finally, the output of the figure is the global map and localization.

Fusion of Horizontal Lidar and Vertical Lidar
To accurately fuse the horizontal and vertical lidar data, their individual coordinate systems must be transformed into a unified coordinate system, which is termed "frame coordinate system". The horizontal lidar coordinate system is used as the frame coordinate system; the vertical lidar coordinate system is registered in this frame coordinate system through external parameter calibration, with the external parameters obtained by the CAD model of the scanning arm. This paper adopts the adaptive time synchronization algorithm to ensure that the timestamps of the horizontal and vertical lidars can be simultaneously output. The fusion process of dual lidar is divided into two parts: (1) external parameter calibration of the horizontal and vertical lidars and (2) the adaptive time synchronization algorithm. Both parts are described as follows.

External Parameter Calibration of Horizontal Lidar and Vertical Lidar
Assuming that 1 L is the horizontal lidar coordinate system and 2 L is the vertical lidar coordinate system, the horizontal lidar coordinate system is used as the frame coordi-

Fusion of Horizontal Lidar and Vertical Lidar
To accurately fuse the horizontal and vertical lidar data, their individual coordinate systems must be transformed into a unified coordinate system, which is termed "frame coordinate system". The horizontal lidar coordinate system is used as the frame coordinate system; the vertical lidar coordinate system is registered in this frame coordinate system through external parameter calibration, with the external parameters obtained by the CAD model of the scanning arm. This paper adopts the adaptive time synchronization algorithm to ensure that the timestamps of the horizontal and vertical lidars can be simultaneously output. The fusion process of dual lidar is divided into two parts: (1) external parameter calibration of the horizontal and vertical lidars and (2) the adaptive time synchronization algorithm. Both parts are described as follows.

External Parameter Calibration of Horizontal Lidar and Vertical Lidar
Assuming that L 1 is the horizontal lidar coordinate system and L 2 is the vertical lidar coordinate system, the horizontal lidar coordinate system is used as the frame coordinate system and the vertical lidar coordinate system L 2 is registered in this frame coordinate system L 1 through the rotation matrix R L 1 L 2 and translation matrix H L 1 L 2 , which can be computed by Equation (5): where R L 1 L 2 is the rotation matrix and H L 1 L 2 is the translation matrix which can be obtained by external parameters. From the hardware structure of the scanning arm, the vertical lidar coordinate system first rotates around the y-axis and then translates to the horizontal lidar coordinate system, where where the angle of the rotation matrix θ = −77.94 • and the translation matrix H L 1 The point cloud of the lidar after external parameter calibration in the frame coordinate system is shown in Figure 3, with the point clouds of the vertical and horizontal lidars not affecting each other.
where the angle of the rotation matrix 77.94 θ°= − and the translat The point cloud of the lidar after external parameter calibration in the fra nate system is shown in Figure 3, with the point clouds of the vertical and horiz not affecting each other.

The Adaptive Time Synchronization Algorithm
In this paper, the lidar time synchronization is achieved through hardwa chronization and GNSS is used as the master clock. However, there is a tim tween the two lidars. As the time offset is constant, an adaptive timestamp sy tion algorithm is used to solve it. The adaptive time synchronization algorithm match its timestamp and realize the simultaneous localization and mapping o The timestamps of the dual lidar is shown in Figure 4, where lidar_201/scan the timestamp of the horizontal lidar and lidar_202/scan represents the times vertical lidar.

The Adaptive Time Synchronization Algorithm
In this paper, the lidar time synchronization is achieved through hardware time synchronization and GNSS is used as the master clock. However, there is a time offset between the two lidars. As the time offset is constant, an adaptive timestamp synchronization algorithm is used to solve it. The adaptive time synchronization algorithm is used to match its timestamp and realize the simultaneous localization and mapping of the lidars. The timestamps of the dual lidar is shown in Figure 4, where lidar_201/scan represents the timestamp of the horizontal lidar and lidar_202/scan represents the timestamp of the vertical lidar. H is the translation matrix which can be obtained by external parameters. From the hardware structure of the scanning arm, the vertical lidar coordinate system first rotates around the y -axis and then translates to the horizontal lidar coordinate system, where 1 2 cos 0 sin where the angle of the rotation matrix 77.94 θ°= − and the translation matrix

The Adaptive Time Synchronization Algorithm
In this paper, the lidar time synchronization is achieved through hardware time synchronization and GNSS is used as the master clock. However, there is a time offset between the two lidars. As the time offset is constant, an adaptive timestamp synchronization algorithm is used to solve it. The adaptive time synchronization algorithm is used to match its timestamp and realize the simultaneous localization and mapping of the lidars. The timestamps of the dual lidar is shown in Figure 4, where lidar_201/scan represents the timestamp of the horizontal lidar and lidar_202/scan represents the timestamp of the vertical lidar.  The output of the adaptive time synchronization algorithm only depends on the timestamp, not on the arrival time of dual lidar point cloud messages. It means that the adaptive time synchronization algorithm can be safely used on dual lidar point cloud messages that have suffered arbitrary processing delays. As shown in Figure 5, time goes from left to right, with the first row representing the horizontal lidar timestamp and the second row representing the vertical lidar timestamp. Each dot represents the lidar point cloud with the timestamp. The blue dot represents the pivot of the lidar point clouds, with the broken line linking the dual lidar point clouds in a set. Suppose the horizontal lidar point clouds' queue is PL P and the vertical lidar point clouds' queue is PV P , with lidar point clouds inserted in a topic-specific queue (PL P or PV P ) as they arrive. Once each topic-specific queue contains at least one message, the latest message is found at the head of the queues, known as the pivot. Assume that the queue with pivots is PL P , which matches PV P , being the smallest difference between the timestamps, T. The two queues are combined into a set; finally, this set is synchronous output. point clouds inserted in a topic-specific queue ( P PL or P PV ) as they arrive. Once each topic-specific queue contains at least one message, the latest message is found at the head of the queues, known as the pivot. Assume that the queue with pivots is P PL , which matches P PV , being the smallest difference between the timestamps, T . The two queues are combined into a set; finally, this set is synchronous output.

Tight Coupling of Dual Lidar and IMU
Although the fusion of dual lidar makes up for the shortcomings of lidar's narrow FOV, the localization accuracy is improved. However, lidars mounted on moving vehicles suffer from motion distortion, which directly affects the localization accuracy. IMU can accurately measure the three-axis acceleration and three-axis angular velocity of moving vehicles, and provide a priori information for lidar odometry. However, in the case of long or short distances, the IMU error will continue to accumulate, which directly affects positioning accuracy. In this paper, the tight coupling of dual lidar and IMU is proposed to solve the above problems. The process is divided into two parts: (1) external parameter calibration of horizontal lidar and IMU and time synchronization, and (2) joint optimization, which are described as follows:

External Parameter Calibration of Horizontal Lidar and IMU and Time Synchronization
Similar to Section 3.3, the external parameters of the lidar and IMU are also obtained based on the CAD model. To accurately fuse data from the horizontal lidar, the vertical lidar, and the IMU, the IMU coordinate system I is also registered in this frame coordi- Η . It can be seen from the structure of the scanning arm in Figure 1 that the IMU is located directly under the horizontal lidar. Therefore, the IMU coordinate system can be registered in the horizontal lidar coordinate system only by translation, where where the translation matrix

Tight Coupling of Dual Lidar and IMU
Although the fusion of dual lidar makes up for the shortcomings of lidar's narrow FOV, the localization accuracy is improved. However, lidars mounted on moving vehicles suffer from motion distortion, which directly affects the localization accuracy. IMU can accurately measure the three-axis acceleration and three-axis angular velocity of moving vehicles, and provide a priori information for lidar odometry. However, in the case of long or short distances, the IMU error will continue to accumulate, which directly affects positioning accuracy. In this paper, the tight coupling of dual lidar and IMU is proposed to solve the above problems. The process is divided into two parts: (1) external parameter calibration of horizontal lidar and IMU and time synchronization, and (2) joint optimization, which are described as follows:

External Parameter Calibration of Horizontal Lidar and IMU and Time Synchronization
Similar to Section 3.3, the external parameters of the lidar and IMU are also obtained based on the CAD model. To accurately fuse data from the horizontal lidar, the vertical lidar, and the IMU, the IMU coordinate system I is also registered in this frame coordinate system L 1 through the rotation matrix R L 1 I and the translation matrix H L 1 I . It can be seen from the structure of the scanning arm in Figure 1 that the IMU is located directly under the horizontal lidar. Therefore, the IMU coordinate system can be registered in the horizontal lidar coordinate system only by translation, where where the translation matrix H L 1 I = [−0.62.66, 0, −0.125]. In this paper, YIS510A IMU is used, which supports trigger correction of its internal clock by PPS signal. Lidar-IMU time synchronization is achieved through hardware time synchronization, with GNSS used as the master clock. GNSS can output PPS signals to unify the clock source of lidar-IMU and achieve time synchronization.

Joint Optimization
In this paper, a fixed-lag smoother and marginalization are introduced to obtain the optimal state. The fixed-lag smoother maintains a certain amount of IMU state in the sliding window, and the sliding window can effectively control the amount of calculation. When the new state enters the sliding window, the past state is marginalized. The state variable to be estimated for the whole window is X = [X W η , . . . , X W κ , Z L 1 I ], where X W η is the state of IMU at the starting point η of the sliding window; X W κ is the state of IMU at the end of the sliding window; and Z L 1 I = [R L 1 I , H L 1 I ] is the external parameter between lidar and IMU. Then, the following cost function with a Mahalanobis norm is minimized to obtain the MAP estimation of the states X, where u o (X) is the prior information from marginalization [26]. u φ (χ, X) is the residual of the relative lidar constraints that can be represented as point-to-plane distance [26], where χ ∈ Q β is the residual for each relative lidar measurement with the previous correspondences. β ∈ {η + 1, . . . , κ}, η + 1 and κ are the timestamps of the lidar sweep next to the starting one and the current lidar sweep in the window, with the covariance matrix C χ L β determined by the lidar accuracy [35]. u γ (Z ς ς+1 , X) is the residual of the IMU constraints, where u γ (Z ς ς+1 , X) can be obtained by IMU state prediction (Equation (2)) and IMU pre-integration (Equation (3)), and ∆q ς,ς+1 −1 ⊗ q −1 ς ⊗ q ς+1 xyz stands for the vector part of a quaternion.
With the continuous-time linearized propagation of the error states and the IMU noise parameters, the covariances C I ς I ς+1 of the pre-integration measurements and biases can be estimated. The cost function, in the form of a non-linear least-square, can be solved by the Gauss-Newton algorithm. Ceres Solver [36] is used for solving this nonlinear problem.
It can be seen from Section 3.2 that the new states X are obtained by the joint optimization, which is used as the next state of the IMU, avoiding the drift from IMU propagation. The state of the IMU is applied to deskewing, thereby eliminating the motion distortion of the dual lidar.

Data Acquisition Equipment
To verify the performance of the proposed high-precision SLAM framework based on the tight coupling of Dual Lidar Inertial Odometry (HSDLIO) for multi-scene applications, this paper describes a series of experiments to qualitatively and quantitatively analyze our proposed framework. The device runs on an industrial computer with a processor of i7-9700 through the port and network port, and the system configured of the industrial computer is ubuntu16.04. The scanning arm is installed on the backpack, with the industrial computer and battery installed in the backpack. The small display screen of the backpack shows the localization and mapping in real-time, with the backpack equipment shown in Figure 6. Data of three representative environments are collected, including the featureless corridor, stairs with height, and complex outdoor environments. To illustrate the effectiveness of the HSDLIO algorithm, LeGO-LOAM and LIO-SAM algorithms are compared with it.

Data Acquisition Equipment
To verify the performance of the proposed high-precision SLAM framework based on the tight coupling of Dual Lidar Inertial Odometry (HSDLIO) for multi-scene applications, this paper describes a series of experiments to qualitatively and quantitatively analyze our proposed framework. The device runs on an industrial computer with a processor of i7-9700 through the port and network port, and the system configured of the industrial computer is ubuntu16.04. The scanning arm is installed on the backpack, with the industrial computer and battery installed in the backpack. The small display screen of the backpack shows the localization and mapping in real-time, with the backpack equipment shown in Figure 6. Data of three representative environments are collected, including the featureless corridor, stairs with height, and complex outdoor environments. To illustrate the effectiveness of the HSDLIO algorithm, LeGO-LOAM and LIO-SAM algorithms are compared with it.

Indoor Experiment 1
Indoor experiment 1 was carried out in a corridor with few feature points. Since there was no GPS signal indoors, SLAM was mainly carried out through the tight coupling of dual lidar odometry and IMU.
To highlight the performance of the HSDLIO algorithm, LeGO-LOAM and LIO-SAM

Indoor Experiment 1
Indoor experiment 1 was carried out in a corridor with few feature points. Since there was no GPS signal indoors, SLAM was mainly carried out through the tight coupling of dual lidar odometry and IMU.
To highlight the performance of the HSDLIO algorithm, LeGO-LOAM and LIO-SAM algorithms are compared with it. Figure 7a shows the environment map of the corridor, while Figure 7b-d shows the overall mapping of LeGO-LOAM, LIO-SAM, and HSDLIO algorithms in the corridor scene, with the white points representing backpack motion trajectories. Compared with the HSDLIO algorithm in Figure 7d, the overall mapping of LeGO-LOAM (Figure 7b) has a significant deviation and the structure is unclear. Figure 8a,b shows that the mapping of LeGO-LOAM is incomplete because the lidar has a considerable drift in a scene with fewer feature points. The mapping results of LIO-SAM (Figure 7c) is close to the real environment, with its structure complete. The main reason is that LIO-SAM tightly couples the horizontal lidar and IMU to make up for the lidar drift. However, Figure 9a,b indicates the mapping result of HSDLIO is more abundant than LIO-SAM in the top surface and ground information. The reason is that HSDLIO fuses horizontal and vertical lidars so that its angle of FOV in both horizontal and vertical directions is 360 • . The dual lidars and IMU are tightly coupled to make up for the lidar drift.    In the long corridor scene, compared with LeGO-LOAM and LIO-SAM, the st of HSDLIO mapping is more complete and the point cloud information on the and top surface is richer. To verify the positioning accuracy of the HSDLIO, the traj of LeGO-LOAM, LIO-SAM, and HSDLIO are compared in Figure 10. It can be ob in Figure 10   In the long corridor scene, compared with LeGO-LOAM and LIO-SAM, the structure of HSDLIO mapping is more complete and the point cloud information on the ground and top surface is richer. To verify the positioning accuracy of the HSDLIO, the trajectories of LeGO-LOAM, LIO-SAM, and HSDLIO are compared in Figure 10. It can be observed in Figure 10 that the LeGO-LOAM trajectory has drifted. The main reason is that in the LeGO-LOAM algorithm, the lidar odometry error continues to increase in a long corridor environment with fewer feature points. Both LIO-SAM and HSDLIO use the tight coupling of IMU and lidar for positioning, which improves positioning accuracy. HSDLIO further improves the accuracy of positioning through the tight coupling of dual lidars and IMU.
To further improve the positioning accuracy of HSDLIO, the relative translational error (RTE)-the distance from the start point to the end point-is introduced. In all experiments, the data collection started and ended at the same point; when the positioning accuracy of the device is very high, the start and end point will coincide. The RTE, when the backpack returns to the start point, is shown in Table 1. The RTE can be computed by Equation (10): where x o = x s − x e , y o = y s − y e , z o = z s − z e , x s , y s , and z s are the coordinates of the starting point and x e , y e , and z e are the coordinates of the end point. The result of LeGO-LOAM is not shown because its trajectory has severely shifted. Table 1 shows that the error of HSDLIO in the x and y direction is similar to LIO-SAM, but the accuracy in the z direction is higher than LIO-SAM and the RTE of HSDLIO is smaller than LIO-SAM. In the long corridor scene, compared with LeGO-LOAM and LIO-SAM, the stru of HSDLIO mapping is more complete and the point cloud information on the gr and top surface is richer. To verify the positioning accuracy of the HSDLIO, the trajec of LeGO-LOAM, LIO-SAM, and HSDLIO are compared in Figure 10. It can be obs in Figure 10 that the LeGO-LOAM trajectory has drifted. The main reason is that LeGO-LOAM algorithm, the lidar odometry error continues to increase in a long co environment with fewer feature points. Both LIO-SAM and HSDLIO use the tigh pling of IMU and lidar for positioning, which improves positioning accuracy. HS further improves the accuracy of positioning through the tight coupling of dual lidar IMU. To further improve the positioning accuracy of HSDLIO, the relative transla error (RTE)-the distance from the start point to the end point-is introduced. In a periments, the data collection started and ended at the same point; when the positi accuracy of the device is very high, the start and end point will coincide. The RTE, the backpack returns to the start point, is shown in Table 1. The RTE can be comput Equation (10):

Indoor Experiment 2
Indoor experiment 2 aimed to show the effectiveness of HSDLIO in the indoor environment with a certain height information. In this experiment, the backpack was carried up and down four flights of stairs. Figure 11a shows the stairs scene, while Figure 11b-d shows the overall mapping of LeGO-LOAM, LIO-SAM, and HSDLIO algorithms. The white points are backpack localization trajectories. LeGO-LOAM (Figure 11b) cannot obtain the structure of the stairs. Because the vertical FOV of the horizontal lidar is relatively narrow, the mapping result of LIO-SAM (Figure 11c) does not show the height information. With the help of the vertical lidar and the tight coupling of the IMU, the mapping of HSDLIO (Figure 11d) is the most complete and accurate, which is closer to the real stairs' scene. Figure 12 shows the trajectories obtained by LeGO-LOAM, LIO-SAM, and HSDLIO. The red, black, and green lines represent the trajectories of HSDLIO, LeGO-LOAM, and LIO-SAM, respectively, with the trajectory of HSDLIO approximating the real trajectory. LeGO-LOAM has no IMU correction, which causes the drift of the trajectory. Although the trajectory of LIO-SAM does not drift, due to the narrow vertical FOV of the horizontal lidar, a significant error occurs in the z direction information. Comparing the height information of LIO-SAM and HSDLIO in the z direction in Figure 13, it can be seen that when going up and down four floors of stairs, the localization results of LIO-SAM provide no height information and also indicate a great drift, while HSDLIO provides accurate height information. Therefore, the tight coupling of dual lidars and IMU makes the advantages of HSDLIO clearly apparent. Table 2 shows that the error of HSDLIO in the x and y direction is smaller than LIO-SAM, but the accuracy in the z direction is much higher than LIO-SAM and the RTE of HSDLIO is much smaller than LIO-SAM.

Indoor Experiment 2
Indoor experiment 2 aimed to show the effectiveness of HSDLIO in the indoor environment with a certain height information. In this experiment, the backpack was carried up and down four flights of stairs. Figure 11a shows the stairs scene, while Figure 11b-d shows the overall mapping of LeGO-LOAM, LIO-SAM, and HSDLIO algorithms. The white points are backpack localization trajectories. LeGO-LOAM (Figure 11b) cannot obtain the structure of the stairs. Because the vertical FOV of the horizontal lidar is relatively narrow, the mapping result of LIO-SAM (Figure 11c) does not show the height information. With the help of the vertical lidar and the tight coupling of the IMU, the mapping of HSDLIO (Figure 11d) is the most complete and accurate, which is closer to the real stairs' scene.  LeGO-LOAM has no IMU correction, which causes the drift of the trajectory. Although the trajectory of LIO-SAM does not drift, due to the narrow vertical FOV of the horizontal lidar, a significant error occurs in the z direction information. Comparing the height information of LIO-SAM and HSDLIO in the z direction in Figure 13, it can be seen that when going up and down four floors of stairs, the localization results of LIO-SAM provide no height information and also indicate a great drift, while HSDLIO provides accurate Appl. Sci. 2022, 12, x FOR PEER REVIEW 14 height information. Therefore, the tight coupling of dual lidars and IMU makes the vantages of HSDLIO clearly apparent. Table 2 shows that the error of HSDLIO in th and y direction is smaller than LIO-SAM, but the accuracy in the z direction is m higher than LIO-SAM and the RTE of HSDLIO is much smaller than LIO-SAM.

Outdoor Experiment
The outdoor experiment was carried out on a city street. This scene has rich feature points, so the drift of lidar odometry grows very slowly. The localization accuracy of LeGO-LOAM, LIO-SAM, and HSDLIO have very little difference, but the mapping of HSDLIO can provide finer structural details.     The trajectories obtained by LeGO-LOAM, LIO-SAM, and HSDLIO are shown in Figure 14. Because there is more feature information outdoors, the drift of lidar odometry grows very slowly. LeGO-LOAM can obtain relatively accurate localization only through horizontal lidar. In LIO-SAM and HSDLIO, GPS measurement is introduced by loosely coupling with IMU through UKF, which can provide relatively accurate initial positioning. The localization accuracy of HSDLIO and LIO-SAM is higher than LeGO-LOAM. The available data in Table 3 further supports the performance of the HSDLIO algorithm. Appl

Outdoor Experiment
The outdoor experiment was carried out on a city street. This scene has rich feature points, so the drift of lidar odometry grows very slowly. The localization accuracy of LeGO-LOAM, LIO-SAM, and HSDLIO have very little difference, but the mapping of HSDLIO can provide finer structural details.
The trajectories obtained by LeGO-LOAM, LIO-SAM, and HSDLIO are shown in Figure 14. Because there is more feature information outdoors, the drift of lidar odometry grows very slowly. LeGO-LOAM can obtain relatively accurate localization only through horizontal lidar. In LIO-SAM and HSDLIO, GPS measurement is introduced by loosely coupling with IMU through UKF, which can provide relatively accurate initial positioning. The localization accuracy of HSDLIO and LIO-SAM is higher than LeGO-LOAM. The available data in Table 3 further supports the performance of the HSDLIO algorithm.  The 3D city street scene is shown in Figure 15a. In Figure 15a-d, the marks in the  The 3D city street scene is shown in Figure 15a. In Figure 15a-d, the marks in the white circle represent trees and the marks in the yellow circle represent the building structure. Figure 15b-d shows the overall mapping of LeGO-LOAM, LIO-SAM, and HSDLIO algorithms in the city street scene, with the white points being their localization trajectories. Figure 15b,c shows that the overall effect of the mapping lacks building height information, while the structure of buildings and trees is blurred. The mapping details are analyzed at the marker, compared with the mapping results of LeGO-LOAM and LIO-SAM (Figure 15b,c), while the HSDLIO mapping result is shown in Figure 15d. It can be seen in Figure 15d that the structure of the trees and buildings is complete and has finer details. To further demonstrate the advantages of HSDLIO mapping, the mapping details of the building are shown in Figure 16, with the mapping results of LeGO-LOAM and LIO-SAM lacking the structural details and height information of the building. However, the mapping result of HSDLIO show finer structural details and more height information of the building. To further demonstrate the advantages of HSDLIO mapping, the mapping details of the building are shown in Figure 16, with the mapping results of LeGO-LOAM and LIO-SAM lacking the structural details and height information of the building. However, the mapping result of HSDLIO show finer structural details and more height information of the building.
structure are finer and more complete, including finer structural details.
To further demonstrate the advantages of HSDLIO mapping, the mapping details of the building are shown in Figure 16, with the mapping results of LeGO-LOAM and LIO-SAM lacking the structural details and height information of the building. However, the mapping result of HSDLIO show finer structural details and more height information of the building.

Conclusions
This paper proposes a high-precision SLAM framework for multi-scene applications. In this framework, dual lidars are fused to make up for the shortcomings of lidars' narrow FOV and hence improve the completeness of mapping. Meanwhile, dual lidars and IMU are tightly coupled to improve the localization accuracy. Extensive experiments were carried out and the results showed that compared with the commonly used LeGO-LOAM and LIO-SAM methods, our proposed method can produce more precise localization and more accurate mapping results with more details.

Conclusions
This paper proposes a high-precision SLAM framework for multi-scene applications. In this framework, dual lidars are fused to make up for the shortcomings of lidars' narrow FOV and hence improve the completeness of mapping. Meanwhile, dual lidars and IMU are tightly coupled to improve the localization accuracy. Extensive experiments were carried out and the results showed that compared with the commonly used LeGO-LOAM and LIO-SAM methods, our proposed method can produce more precise localization and more accurate mapping results with more details.

Conflicts of Interest:
The authors declare no conflict of interest.