Next Article in Journal
Comparison of Phenological Parameters Extracted from SIF, NDVI and NIRv Data on the Mongolian Plateau
Next Article in Special Issue
LiDAR Point Clouds Usage for Mapping the Vegetation Cover of the “Fryderyk” Mine Repository
Previous Article in Journal
Corrections of Mesoscale Eddies and Kuroshio Extension Surface Velocities Derived from Satellite Altimeters
Previous Article in Special Issue
A Multi-Scale Feasibility Study into Acid Mine Drainage (AMD) Monitoring Using Same-Day Observations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust LiDAR SLAM Method for Underground Coal Mine Robot with Degenerated Scene Compensation

1
School of College of Surveying and Mapping Science and Technology, Xi’an University of Science and Technology, Xi’an 710054, China
2
Shanxi Key Laboratory of Mine Electromechanical Equipment Intelligent Detection and Control, School of Mechanical Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(1), 186; https://doi.org/10.3390/rs15010186
Submission received: 27 September 2022 / Revised: 21 December 2022 / Accepted: 26 December 2022 / Published: 29 December 2022
(This article belongs to the Special Issue Remote Sensing Solutions for Mapping Mining Environments)

Abstract

:
Simultaneous localization and mapping (SLAM) is the key technology for the automation of intelligent mining equipment and the digitization of the mining environment. However, the shotcrete surface and symmetrical roadway in underground coal mines make light detection and ranging (LiDAR) SLAM prone to degeneration, which leads to the failure of mobile robot localization and mapping. To address these issues, this paper proposes a robust LiDAR SLAM method which detects and compensates for the degenerated scenes by integrating LiDAR and inertial measurement unit (IMU) data. First, the disturbance model is used to detect the direction and degree of degeneration caused by insufficient line and plane feature constraints for obtaining the factor and vector of degeneration. Second, the degenerated state is divided into rotation and translation. The pose obtained by IMU pre-integration is projected to plane features and then used for local map matching to achieve two-step degenerated compensation. Finally, a globally consistent LiDAR SLAM is implemented based on sliding window factor graph optimization. The extensive experimental results show that the proposed method achieves better robustness than LeGO-LOAM and LIO-SAM. The absolute position root mean square error (RMSE) is only 0.161 m, which provides an important reference for underground autonomous localization and navigation in intelligent mining and safety inspection.

1. Introduction

Autonomous localization and navigation in an unknown underground environment are important and well-researched yet still active fields for mobile robots and are the key to the automation of intelligent mining equipment. With the rise of big data, cloud computing and artificial intelligence, intellectualization, as a disruptive innovative technology, has become the core driving force for the reform of basic industries around the world. Using intelligent mining equipment to drive the transformation and upgrading of the traditional mining industry can essentially enhance the core competitiveness of mining enterprises and promote the development of the traditional mining industry towards the goal of high efficiency, safety and sustainability [1].
Intelligent mining is the core technology to achieve high-quality development of the coal industry, which requires fast, accurate, automatic and full coverage of spatial data acquisition [2]. However, the traditional single-point measurement and stand-by three-dimensional (3D) laser scanning [3] data acquisition methods are labor-intensive, low in efficiency and time-consuming, which cannot meet the requirements of autonomous operation of a robot under complex conditions in coal mines. A mining robot based on SLAM technology can accurately and quickly construct a 3D map of an underground coal mine, and the map provides flexible and reliable assistance to coal mining robots for intelligent navigation and obstacle avoidance, which can be applied to work in the hazardous area under coal mines, automatic patrol inspection, remote dispatching, etc. To speed up the development of underground space exploration technology, the Defense Advanced Research Projects Agency (DARPA) launched the Underground Space Challenge in 2018 [4,5], which aims to revolutionize the level of rescue and exploration in underground space. However, the high dust level, weak texture and low illuminance in the underground environment lead to unstable feature point extraction in visual SLAM. LiDAR SLAM, on the other hand, has good accuracy and robustness. However, it suffers from motion distortion and point cloud degeneration, being easy to degenerate in the narrow mining environment. The defects of LiDAR SLAM can be compensated for by fusing IMU information, which is not affected by structural features and great changes in the environment and can provide high-precision pose estimation in a short time. However, an IMU faces error accumulation over time. To compensate for the defects of a single sensor, multi-sensor fusion methods are increasingly being used to improve the robustness of state estimation in complex environments [6,7,8].
Recently, with the continuous improvement of autonomous driving solutions and the implementation of the Xihe Plan (Seamless Navigation and Positioning Service in All Space, All Time, Indoor and Outdoor) of the Beidou Project, the demand for intelligent sensing is increasing [9,10,11,12]. Extensive studies have investigated the autonomous navigation and map building of robots, which can be divided into loose-coupled based methods and tight-coupled based methods. (1) The loose-coupled based method estimates pose by LiDAR and IMU data, which is efficient, but the error feedback mechanism is not fully utilized. Representative works are LOAM, LeGO-LOAM, Cartographer, etc. LOAM [13] is a pioneering framework for real-time odometry and mapping by applying point-to-line and point-to-plane registration. However, it uses only LiDAR odometry, which has no back-end optimization and loopback detection. LeGO-LOAM [14] performs clustering and segmentation on plane points to extract ground features and applies two-step optimization for pose estimation, which improves the operation efficiency on the lightweight platform. However, IMU data are only used to provide prior information for scan matching and point cloud distortion removal, which is not used as an observation constraint for optimization. Cartographer [15] uses the Unscented Kalman Filter (UKF) algorithm to fuse multi-source data for pose estimation. However, the back-end optimization is still constrained by point cloud matching, and it is easy to fail in the degenerated scene or when LiDAR rotates rapidly. (2) The tight-coupled based method fuses LiDAR and IMU observations together and then estimates the pose of the mobile robot. Its advantage is that the observation data of LiDAR and the IMU are fully utilized to reduce the cumulative error. However, it increases the amount of calculation, and the failure of one sensor may lead to the failure of the whole system. Koide et al. [16] proposed a real-time 3D mapping framework based on global matching cost minimization using LiDAR-IMU tight coupling, which achieves accurate and robust localization and mapping in challenging environments. FAST-LIO [17] is a proposed efficient and robust LIO framework based on tightly coupled iterative Kalman filters. However, the system discards the influence of historical data on the current state and global pose correction cannot be performed. LIOM [18] is a tight-coupled LiDAR/IMU localization and mapping framework based on graph optimization. To achieve real-time and consistent estimation, the moving sliding window method is used to marginalize the old pose. Meanwhile, the preintegration information of an IMU is used to eliminate the motion distortion of the point cloud and jointly minimize the error function of LiDAR and IMU measurement. However, this method still takes time to construct constraints and batch optimization in local windows. LINS [19] is an improvement on LIOM and proposes a tight-coupled LiDAR/IMU localization and mapping method based on an iterative error state Kalman filter algorithm, which overcomes the loss of accuracy caused by discarding high-order errors in the linearization process of an extended Kalman filter. Compared with the LIOM algorithm, LINS has improved operation efficiency. However, the accuracy of LINS depends on the structured environment. To improve accuracy and time efficiency, LiLi-OM [20] is a proposed tight-coupled LiDAR/IMU SLAM method using keyframe-based sliding window optimization, which can support solid-state LiDAR and mechanical LiDAR. However, point cloud matching between adjacent frames degenerates in symmetrical tunnels and corners. LIO-SAM [21] is a proposed tight-coupled LiDAR/IMU fusion framework via smoothing and mapping. However, the IMU data are only used for joint optimization without considering the impact of environmental degeneration on SLAM results. Kim et al. [22] developed an autonomous driving robot for underground mines using IMUs, LiDAR and encoders. It fuses three types of sensors and achieves high accuracy in estimating the location of autonomous robots in underground mines. Miller et al. [23] used multiple quadrupedal robots to explore and map the inside of a mine tunnel. They demonstrated the feasibility and functionality of the method in laboratory and field tests. Mascarich et al. [24] combined multiple sensors to develop an autonomous driving robot that performs autonomous driving tasks such as exploring the tunnel mapping environment. Since LiDAR SLAM uses the geometry structure of the environment to perform localization and mapping, it is vulnerable to geometrically degenerated environments such as open space [25] and long, straight tunnels [26,27], especially when lacking enough constraints in the shotcrete surface and symmetrical roadway in an underground coal mine. Extensive studies have investigated the degeneration problems of SLAM—f or example, eliminating the degeneration by restricting movement [28,29], minimizing spatial degeneration components in the direction of degeneration [30], adding a plane feature constraint [31,32,33,34,35], sensor fusion [36,37,38,39,40,41], etc. Zhang et al. [42] integrated IMU and odometry information into the Cartographer’s mapping process, which improved the robustness of the algorithm in a long corridor environment. The online method proposed in [43] explained the impact of environmental degeneration on pose estimation. However, this method only uses the non-degenerate direction solution and does not deal with the degenerate direction state.
Although great efforts have been devoted to achieving a robust LiDAR SLAM method for underground coal mines, the shotcrete surface and symmetrical roadway in underground coal mines make LiDAR SLAM prone to degeneration, which leads to the failure of mobile robot localization and mapping. Therefore, a robust LiDAR SLAM method is proposed for an underground coal mine robot with degenerated scene compensation, which detects and compensates for the degenerated scenes by integrating LiDAR and IMU data. Extensive experiments with qualitative and quantitative analyses have verified the accuracy and efficiency of the proposed method, which can realize accurate and reliable SLAM in underground coal mines and provide a theoretical reference and technical support for intelligent mining and safety inspection in coal mines. The main contributions of this paper are as follows:
  • The unknown linear equation is added to the state optimization equation as the disturbance model to detect the direction and degree of degeneration caused by insufficient line and plane feature constraints.
  • The IMU pose is used to compensate for ill-conditioned components in the direction of degeneration, which cannot be determined directly by scan matching. LiDAR rotation state degeneration is compensated for by projecting IMU poses onto plane features. When degeneration is also detected in the translation direction, the compensated rotation state and IMU translation state are fused into a new LIDAR pose, which is then used for scan-to-submap matching to achieve two-step degeneration compensation.
  • A tightly coupled LiDAR/IMU fusion framework is implemented based on factor graph optimization. The IMU measurements and LiDAR point cloud features are jointly optimized in a sliding window, which improves the accuracy and robustness of SLAM in the underground coal mines with the shotcrete surface and symmetric roadway environment.

2. Materials and Methods

2.1. System Configuration

In this paper, a data acquisition platform is designed based on an underground coal mine mobile robot equipped with LiDAR and an IMU, as shown in Figure 1. The proposed method defines the mobile robot coordinate system B as a reference, which is consistent with the LiDAR coordinate system L . The IMU coordinate system is defined as I . The world coordinate system is defined as W . The origin of the world coordinate system was set to the center of the mobile robot at the time of system initialization, and the Z-axis is opposite to the gravity direction in the world coordinate system. Table 1 shows the detailed specifications of the mobile robot system. An Autolabor Pro1 robot was used as the carrier to carry the VLP-16 LiDAR and Ellipse2-N IMU. External parameters between LiDAR and IMU ware calibrated in advance [44]. Allen variance was calculated to determine the degree of trust in the diagonal velocity and acceleration observations of the system. Time synchronization between the IMU and LiDAR was triggered by pulse per second (PPS). Then, the proposed method was implemented by C++ based on the robot operating system (ROS), and the nonlinear optimization was implemented using the Ceres library. The computer used in the experiment was AMD Ryzen3 3200G with DDR4 8GB.

2.2. Method outline

The outline of the proposed method is shown in Figure 2, which can be divided into three parts. (1) Preprocessing: IMU pre-integration is used to remove point cloud distortion. The non-ground point cloud is clustered and segmented from the real-time collected point cloud, and then, the line and plane features are extracted. (2) Front-end: The pose from the IMU between consecutive frames of LiDAR is used to provide an initial value for the feature matching. The disturbance model is used to detect the direction and degree of degeneration caused by insufficient line and plane feature constraints. The IMU poses are projected onto the plane features to compensate for the LiDAR rotation state. The IMU translation state is fused with the compensated rotation state to form a new LiDAR pose as the initial value for local map matching. (3) Back-end: The IMU pre-integration factor, LiDAR odometry factor and loop closure factor are used to construct the factor graph, and the new related variable nodes are optimized by the factor graph. The input of the proposed method is LiDAR point cloud and IMU data, and the output is the trajectory and map.

2.3. Data Preprocessing

The IMU pre-integration method proposed in [18] was used to obtain the relative motion between LiDAR scan frames for distortion removal. In addition, IMU poses were also used to provide initial values for LiDAR odometry optimization [45]. The acceleration and angular velocity measured by the IMU can be expressed as ω ˜ W B B ( t i n ) and α ˜ B ( t i n ) , respectively, and the IMU measurement model is as follows:
ω ˜ W B B ( t i n ) = ω W B B ( t i n ) + b g ( t i n ) + η g ( t i n )
α ˜ B ( t i n ) = R W B T ( t i n ) ( α W ( t i n ) g W ) + b α ( t i n ) + η α ( t i n )
where ω W B B ( t i n ) is the instantaneous angular velocity of I relative to W , R W B T ( t i n ) is the rotation matrix from the world coordinate system to the IMU coordinate system, α W ( t i n ) is the instantaneous acceleration in the world coordinate system, b g ( t i n ) and b α ( t i n ) are the deviation of the gyroscope from the acceleration and η g ( t i n ) and η α ( t i n ) are the random noise. According to the kinetic model of the IMU, the discrete integral method was used to integrate ω W B B ( t i n ) and α W ( t i n ) in the IMU sampling interval Δ t :
R W B ( t i n + 1 ) = R W B ( t i n ) · exp [ ( ( ω ˜ W B B ( t i n ) b g ( t i n ) η g ( t i n ) ) · Δ t ) ]
v W B ( t i n + 1 ) = v W B ( t i n ) + g W · Δ t + R W B ( t i n ) ( α ˜ B ( t i n ) b α ( t i n ) η α ( t i n ) ) · Δ t
p W B ( t i n + 1 ) = p W B ( t i n ) + v W B ( t i n ) · Δ t + 1 2 g W · Δ t 2 + 1 2 R W B ( t i n ) ( α ˜ B ( t i n ) b α ( t i n ) η α ( t i n ) ) · Δ t 2
where represents the transformation of a three-dimensional vector into an antisymmetric matrix. The re-parameterized results of IMU pre-integration between two keyframes prevent the repeated integration of IMU observations and improve the reliability of the algorithm.
The point cloud obtained by LiDAR in real-time inevitably has distortion due to the rotation mechanism of LiDAR and the nonlinear motion of the sensor. Matching with a distorted point cloud will result in a large drift error. The continuous IMU states X k W and X k + 1 W closest to the current LiDAR point cloud timestamp were obtained by pre-integration, and then, the IMU state X c u r r W at t c u r r in the world coordinate system was determined by the linear interpolation method:
X c u r r W = X k + 1 W × t c u r r t k t k + 1 t k + X k W × t k + 1 t c u r r t k + 1 t k
The motion distortion Δ p c u r r s t a r t of the current LiDAR cloud point was calculated in the start LiDAR point cloud coordinate system:
Δ p c u r r s t a r t = T W L [ p c u r r W ( p s t a r t W + v s t a r t W × t c u r r s t a r t ) ]
where T W L is the transformation matrix from the LiDAR coordinate system to the world coordinate system and t c u r r _ s t a r t is time of laser line scanning in the frame. Then, the LiDAR point cloud could be converted to the start of the LiDAR point cloud coordinate system. The coordinates after the distortion removal can be calculated by Δ p c u r r s t a r t .
In this paper, the features were described using the local roughness of the point cloud. The range image was used to calculate the roughness c of point S in its neighborhood in the same row, the non-ground points with larger roughness c were marked as line feature points, and the points with smaller roughness c were marked as plane feature points. The extraction results of line and plane feature points in the preparation roadway by the proposed method are shown in Figure 3. Many green plane feature points were extracted, but only a few blue line feature points were extracted in the underground coal mine environment. The coordinate system was the robot coordinate system, and the robot moved in the Y-axis direction. These arrows are the three-axis constraints of line and plane features in pose estimation.

2.4. Front-End Odometry

In this paper, we matched the point clouds S t and S t + 1 at moments t and t + 1 to estimate the pose T t + 1 t . Two matching points ( s t , i l , s t , j l ) were found in the line feature points L t and three matching points ( s t , i p , s t , j p , s t , k p ) were found in the plane feature points P t . We established constraints on the minimum point-to-line and point-to-plane distances, respectively.
d L t + 1 , n = ( s t + 1 , n l s t , i l ) × ( s t + 1 , n l s t , j l ) s t , i l s t , j l
d P t + 1 , n = ( s t + 1 , n p s t , i p ) · ( s t , i p s t , j p ) × ( s t , i p s t , k p ) ( s t , i p s t , j p ) × ( s t , i p s t , k p )
where d L t + 1 , n and d P t + 1 , n are the distances from the point-to-line and the point-to-plane, respectively. We obtained the function of d to construct the objective function.
f ( S t l , S t + 1 l , S t p , S t + 1 p , T t + 1 t ) = min { d L t + 1 , n + d P t + 1 , n }
The pose T t + 1 t cloud be solved by minimizing f ( S t l , S t + 1 l , S t p , S t + 1 p , T t + 1 t ) through a nonlinear iterative objective function.
We used a two-step optimization for pose estimation. Point-to-line and point-to-plane scan matching were performed to find the transformation between two scans. Considering that plane features are more numerous than line features in the underground coal mine, the pose parameters [ θ r o l l , θ p i t c h , t z ] were preferentially estimated by matching many plane features. Then, the remaining parameters [ θ y a w , t x , t y ] were estimated by line features while using [ θ r o l l , θ p i t c h , t z ] as constraints. The two-step pose estimation method optimizes the pose and effectively suppresses the pose drift. The initial value of iterative calculation was provided by IMU pre-integration between scan frames, effectively reducing the number of iterations. The Levenberg Marquardt method was used for pose optimization.
It is difficult for mobile robots to establish the data association of matching points in underground coal mines with insufficient structural features, resulting in insufficient direction constraints in the state space. The normal equation matrix of the point cloud matching solution is ill-conditioned. The degeneration factor was used to detect and judge the degeneration direction, and the linear optimization problem was constructed as follows.
arg min x A x b 2
The optimization problem corresponds to the overdetermined systems of linear equations A x = b . The optimal solution was obtained by solving the overdetermined linear equations with the least-squares method. The unknown linear equation was added to the state optimization equation as the disturbance model of the degeneration problem to detect the direction and degree of degeneration. The degeneration factor D was defined as a quantity only related to the eigenvalue λ i of A T A , as shown in the following expression:
D = λ min + 1
where λ min is the smallest eigenvalue. Singular value decomposition (SVD) was used to obtain the six eigenvalues of A T A , and their eigenvectors correspond to the [ θ r o l l , θ p i t c h , θ y a w , t x , t y , t z ] . Feature matching fails when D is less than the degeneration factor threshold D t h r , which can be obtained from the measured data of the actual environment. The statistical result of the degeneration factor in the indoor corridor environment is shown in Figure 4. We counted the degeneration factors of degeneration and non-degeneration scenes. The statistics present a clustering effect and can take the cut-off clustering value D t h r = 180 as the threshold.
Figure 3 shows the constraints of line and plane feature points extracted from the scanning point cloud acquired in real time. A large number of plane feature points provide sufficient (red, white and yellow arrows) constraints on [ θ r o l l , θ p i t c h , t z ] . Generally, these degrees of freedom do not suffer from degeneration problems. However, the extracted line feature points are few and the constraints provided for [ θ y a w , t x , t y ] are insufficient, which is prone to degeneration. The sidewalls in the roadway are unsuitable to provide vertical constraints, and relying only on the lateral constraints of the LiDAR point cloud during turns can easily lead to incorrect pose estimation.
The scan matching ill-conditioned component caused by degeneration is compensated for by the pose from the IMU. The number of line feature points extracted from the LiDAR point cloud is far smaller than the plane feature points in the underground coal mine. Therefore, the proposed method uses the pose from the IMU to compensate for the degeneration rotation. The rotation state of the IMU pre-integration is accurate in a short time, but plane constraints may not be maintained. The rotation state with sufficient constraints is obtained by feature matching after degeneration compensation. Then, the compensated rotation state and the IMU translation state are fused into a new pose from LiDAR when the translation direction of the degeneration is detected. Finally, the LiDAR pose is used as the initial value of local map matching to pose optimization. It effectively reduces the drift of system pose estimation and improves the accuracy of localization and mapping.
In this paper, q I M U and q L i D A R represent the rotation state of the IMU and LiDAR, respectively; q I M U and q L i D A R are conjugates of q I M U and q L i D A R ; α represents the degenerated vector in the direction of rotation; and the normal vectors updated by the IMU poses and plane features are represented by v I M U and v L i D A R , respectively, as shown in the following expression:
( 0 , v I M U ) = q I M U ( 0 , α ) q I M U
( 0 , v L i D A R ) = q L i D A R ( 0 , α ) q L i D A R
The rotation state of the current point cloud after degeneration compensation is represented by q L , as shown in the following expression:
q L = q L i D A R + Δ q = q L i D A R + ( v I M U v L i D A R + v I M U v L i D A R , v I M U × v L i D A R ) ( v I M U v L i D A R + v I M U v L i D A R , v I M U × v L i D A R )
where Δ q is calculated by projecting the pose from IMU preintegration to the shortest arc length parallel to the plane feature. Then, Δ q is used to compensate for ill-conditioned vectors relative to the rotation direction that cannot be estimated by feature matching. For the degeneration in the translation direction, the IMU translation vectors combined with the compensated rotation vectors are used to reconstruct the point cloud state, which can be used as the initial value of the optimization iteration for local map matching. The pose T L of the current point cloud can be represented as follows:
T L = [ 1 2 q L y 2 2 q L z 2 2 q L x q L y 2 q L w q L z 2 q L x q L z + 2 q L w q L y X L x 2 q L x q L y + 2 q L w q L z 1 2 q L x 2 2 q L z 2 2 q L y q L z 2 q L w q L x X L y 2 q L x q L z 2 q L w q L y 2 q L y q L z + 2 q L w q L x 1 2 q L x 2 2 q L y 2 X L z 0 0 0 1 ]
where X L represents the translation vector of the IMU, which can be converted into the translation of the current point cloud by the external parameters of LiDAR and the IMU. Local map matching adopts line and plane feature matching [46], which not only corrects the degeneration in the translation direction inter-frame matching but also limits the error accumulation of the LiDAR odometry.
The degeneration compensation algorithm proposed in this paper is shown in algorithm 1. The input values are the LiDAR pose TLiDAR, the IMU pose TIMU, the degeneration factor D and the degeneration factor threshold Dthr. If the degeneration factor Dr of the rotation direction is lower than the degeneration threshold Dthr, the rotation state in the IMU pose is extracted and projected to the plane features to compensate for the degeneration of the rotation direction. If the degeneration factor Dt in the translation direction is lower than the degeneration threshold Dthr, the compensated degeneration rotation state q L i D A R or the non-degeneration rotation state is fused with the IMU translation state to obtain a new LiDAR pose. Finally, the LiDAR pose is used as the initial value of local map matching to update the pose T L i D A R .
Algorithm 1. degeneration compensation
1: input: T L i D A R , T I M U , D , D t h r ;
2: output: T L i D A R ;
3:  if D r < D t h r  do
4:      Compute q L i D A R , q I M U , X L i D A R , X I M U of T L i D A R , T I M U ;
5:      Compute v L i D A R and v I M U of q L i D A R , q I M U based on (13) and (14);
6:      Compute q L i D A R of v L i D A R and v I M U based on (15);
7:  end
8:  if D t < D t h r  do
9:      Construct T L i D A R from q L i D A R and X I M U based on (16);
10:      Local map matching and updated T L i D A R ;
11:      Return T L i D A R ;
12:  end

2.5. Factor Graph Optimization

In this paper, factor graph optimization is used for data fusion, and the IMU pre-integration between consecutive keyframes is put into the factor graph. To reduce the redundant keyframes and improve the efficiency of the algorithm, a multiple keyframe selection strategy based on the Euclidean metric, rotation angle and overlap of the point cloud scan matching is introduced. The LiDAR odometry factor is constructed by the pose constraints between consecutive keyframes. The back-end optimization uses the GICP algorithm for matching. The loop closure factor is formed by the local loop closure and the global loop closure, which is introduced to construct the factor graph. Local loop closure is the intervisibility relationship between keyframes, which provides more constraints and improves the robustness of the system to LiDAR rotation motion when the mobile robot turns. The global loop closure means that the mobile robot returns to the revisit position and establishes a loop closure constraint to reduce the global error accumulation. A new robot pose node is added to the factor graph when the change of robot pose exceeds the user-defined threshold. The factor graph is optimized upon the insertion of a new node using incremental smoothing and mapping (iSAM2) with the Bayes tree [47]. A schematic diagram of factor graph optimization by the proposed method is shown in Figure 5.

3. Experimental Analysis

In order to verify the proposed method, qualitative and quantitative analysis experiments were undertaken with the data from an indoor corridor and an underground coal mine collected by a self-designed mobile robot platform and compared with other state-of-the-art LiDAR SLAM methods.

3.1. Qualitative Analysis

3.1.1. Qualitative Analysis with Indoor Corridor

The mobile robot ran in the indoor corridor according to the path A B C B A , and the total length of the trajectory was 0.15 km, as shown in Figure 6a, which shows the localization and mapping results by the proposed method. It can be seen that the point cloud map constructed by the proposed method directly and accurately reflects the actual situation of the indoor corridor environment. The AB and BC segments are typical degenerate scenes with equal width on both sides, which can easily cause matching failure. The robot mistakenly believed it did not move, resulting in it mapping a length smaller than that of the actual trajectory. LIO-SAM drifted in the X-axis direction of the AB and BC segments, as shown in the yellow box in Figure 6b, and there is a ghosting in the mapping. Figure 6c shows the localization and mapping results by LeGO-LOAM. It can be seen that LeGO-LOAM has significantly degenerated in the Y-axis direction of the AB and BC segments, and the trajectory is shortened in the forward direction of the robot. There were matching errors in the pose estimation process, resulting in large localization drift and mapping ghosting. The proposed method detected and compensated for the degeneration in all directions. The trajectory was consistent with the actual motion of the mobile robot and the deviation of the point cloud map was small.

3.1.2. Qualitative Analysis with the Underground Coal Mine

Next, experimental data were collected from the underground coal mine, and the length of motion trajectory was about 0.45 km. As shown in Figure 7, there are long and narrow coal mine roadways with different widths, which brings great challenges to the existing LiDAR SLAM methods. Figure 7a–d show the data collection environment of the underground coal mine, and Figure 7e is the motion trajectory (red points) of the mobile robot; the point cloud map of the coal mine was obtained by the proposed method. The statistical analysis showed that the threshold of the degeneration factor in the underground coal mine is 240, and the degeneration state accounts for 45.7% of the total measured data. To quantitatively evaluate the absolute localization accuracy of the proposed method, the coordinate values of 20 reference points (K1-K20, green points) on the motion trajectory were obtained by Total Station. The stopping time interval of the mobile robot was recorded at the corresponding reference point, and then, the average value of the pose estimation result at the reference point was taken as the measurement value of the proposed method. The localization trajectory of the proposed method is close to the reference trajectory points.
The localization and mapping results by the proposed method and the LeGO-LOAM and LIO-SAM methods are shown in Figure 8 and Figure 9. The trajectories of LeGO-LOAM and LIO-SAM degenerated significantly in the direction of rotation and translation at the preparation roadway. The degeneration of pitch angle by LeGO-LOAM led to a large trajectory drift and mapping ghosting in the gateway and main haulageway. LIO-SAM mistakenly performed loop closure due to the degeneration of the X-axis direction at the main haulageway, resulting in a large trajectory drift and mapping ghosting of the development roadway and the preparation roadway. In this paper, IMU pre-integration was used to compensate for degeneration, and factor graph optimization was introduced to reduce trajectory drift. The point cloud map constructed by the proposed method can intuitively reflect the actual condition of the roadway environment and had good robustness and accuracy in the underground coal mine. As shown in Figure 8, the trajectory of the proposed method was consistent with the reference trajectory points. LIO-SAM established a wrong loop closure constraint near the start point, which led to the drift of overall trajectory. LeGO-LOAM had a large drift in the Z-axis due to the interference of some similar point clouds and no loop closure constraint.
Figure 10 shows the localization and mapping results by the proposed method, LeGO-LOAM and LIO-SAM in the preparation roadway. Overall, 76.1% of the LiDAR data were in a degenerated state. The localization trajectory of the proposed method is close to the reference trajectory points. It can be seen that the mapping accuracy is high from the thickness of the vertical wall. As shown in Figure 10b, LIO-SAM degenerated in the Y-axis direction, and the localization trajectory has a large drift, resulting in point cloud mismatching and mapping ghosting. LeGO-LOAM degenerated in rotation and translation, and the whole map rotates incorrectly in the yaw direction, resulting in a large trajectory drift and mapping ghosting, as shown in Figure 10c, where there are many wrong walls built around. The proposed method performs degeneration detection and compensation by IMU pre-integration and obtained better localization and mapping results, as shown in Figure 10a.

3.2. Quantitative Evaluation

To quantitatively evaluate the accuracy of the proposed method, a laser range finder was used to measure the distance of AB and BC in the indoor corridor. The mean value measured by the laser range finder was the reference value. The quantitative comparisons of the proposed method, LIO-SAM and LeGO-LOAM are shown in Table 2, where L A _ A represents the distance from the start to the end of the mobile robot. The proposed method showed a higher localization accuracy in the degenerated corridor than the other two methods. The error percentage between the proposed method and the reference value is less than 0.6%, and the distance error of the trajectory closure is 0.07 m. Due to the large windows at the end of the BC segment, the laser passed through the window, resulting in the lack of laser scanning data, which reduced the constraints and made the accuracy of pose estimation in the BC segment slightly lower than that in the AB segment. The sidewalls in the corridor environment only provided lateral and insufficient vertical constraints. Some similar point clouds are prone to mismatching, and the robot cannot determine that it is moving forward. LIO-SAM and LeGO-LOAM were degenerated in the Y-axis direction and had poor localization performance.
Figure 11 shows the absolute localization error between the proposed method, LIO-SAM and LeGO-LOAM. The localization accuracy of LeGO-LOAM is poor compared with the proposed method and LIO-SAM, and the maximum translation error in the triaxial direction is more than 1 m. The mean and median of triaxial direction error by the proposed method are lower than those of LIO-SAM. Although the error in the Y-axis direction fluctuates, it is compensated for to a certain extent, and the maximum triaxial direction translation error is smaller than that of LIO-SAM. The localization accuracy of the proposed method in the X-axis and Z-axis directions is significantly higher than that in LIO-SAM. The mean translation error in the triaxial direction is less than 0.2 m.
In order to further verify the proposed method, the RMSE was used to evaluate theabsolute accuracy. The results are shown in Table 3. The degeneration of LeGO-LOAM led to the RMSE being more than 0.5 m in the triaxial direction between the measured data and the reference points, and its position RMSE was 0.976 m. The localization accuracy of the proposed method was high, especially in the Z-axis direction, which can reach 0.044 m. The RMSE of position error was 0.161 m. Base on the triaxial direction and position RMSE, the localization accuracy of the proposed method is higher than that of LIO-SAM.
The relative error of the trajectory by the three methods in the three-axis direction is shown in Figure 12. Compared with the proposed method, LeGO-LOAM and LIO-SAM had a certain drift in the X-axis and Y-axis directions, which was caused by the degeneration of the matching process. However, the relative error of the LIO-SAM trajectory was smaller than that of LeGO-LOAM. The main reason is that LIO-SAM fuses IMU information and adds loop closure to suppress cumulative errors. The two methods had great drift in the Z-axis direction, especially for the relative error by LeGO-LOAM, which was as high as 4 m. The mapping of LeGO-LOAM is shown in Figure 9c. There was a large cumulative error in the Z-axis direction after a period of operation. The proposed method detects and compensates for the degenerated scenes in the underground coal mine environment, adds loop closure detection and constructs a global constraint LiDAR SLAM based on factor graph optimization, which obtains a better localization result.

3.3. Time Performance

The time performance of the proposed method is shown in Figure 13, where P represents the preprocessing, DD represents degeneration detection, DC represents degeneration compensation, FGO represents factor graph optimization and M represents mapping. It can be seen that the preprocessing takes less time, and the degeneration detection consumes little time. The mean and median values of degeneration compensation are no more than 20 ms, and the maximum time for factor graph optimization is no more than 10 ms. The mapping thread takes the most time and can reach 165.75 ms, with an average of 109.97 ms. In the proposed method, the mapping thread receives the point cloud and updates the map at 5 Hz, which can run in real-time.

3.4. Discussion

(1) Pose estimation: To achieve accurate and robust pose estimation, the disturbance model was used to detect the direction and degree of degeneration caused by insufficient line and plane feature constraints for obtaining the factor and vector of degeneration. The IMU poses were projected onto plane features and fused into new LiDAR poses for local map matching to achieve two-step degeneration compensation. Moreover, loop closure and factor graph optimization were added to suppress the cumulative error of pose estimation. The comprehensive experiments showed that the proposed method was superior to state-of-the-art methods in the underground coal mine. However, the quantitative analysis was limited by the actual situation, as shown in Figure 7e. Only 20 reference points at key positions were selected for quantitative analysis.
(2) Mapping results: The proposed method had a good mapping effect due to the detection and compensation of degeneration states. However, the mapping results by LeGO-LOAM and LIO-SAM in the underground coal mine were relatively inaccurate as the walls were thick and not even aligned, as shown in Figure 9 and Figure 10. The disadvantage here is that it is difficult to make a quantitative comparison. The tight integration of LiDAR and IMU data may further improve SLAM accuracy.
(3) Time performance: It can be seen that the runtime of the proposed method is slightly higher due to the detection and compensation of degeneration, as shown in Figure 13. However, the runtime is still less than 0.1 s, which can run in realtime on low-power embedded systems.

4. Conclusions

To address the problem of LiDAR SLAM easily degenerating in the shotcrete surface and symmetrical roadway of underground coal mines, this paper proposed a robust LiDAR SLAM method which detects and compensates for degenerated scenes by integrating LiDAR and IMU data. The disturbance model is used to detect the direction and degree of degeneration for obtaining the factor and vector of degeneration. The pose obtained by IMU pre-integration is projected to plane features for the compensation of rotation state degeneration. The compensated rotation and IMU translation state are fused into a new pose from LiDAR when the translation direction degeneration is detected, which is then used for scan-to-submap matching to achieve two-step degenerated compensation. Lastly, globally consistent LiDAR SLAM is implemented based on factor graph optimization. It can reduce the global cumulated error and improve the trajectory accuracy and map consistency. Extensive experimental results show that the proposed method achieves better robustness than state-of-the-art LiDAR SLAM methods. In the indoor corridor, the error percentage of the proposed method did not exceed 0.6%, and the distance error of trajectory closure was only 0.07 m. In the underground coal mine, the accuracy of pose estimation by the proposed method was the highest in the Z-axis direction, its translation error was only 0.044 m and the absolute position RMSE was 0.161 m. The point cloud map constructed has an excellent performance in integrity and geometric structure authenticity, providing an important reference for underground autonomous navigation and positioning in intelligent mining and safety inspection. In the future, multi-sensor fusion localization and mapping will be further carried out in combination with the degenerated scene in the coal mine to improve the accuracy of localization and mapping.

Author Contributions

X.Y. wrote the manuscript. X.Y. designed the experiments and framework. X.L. supervised the entire process of the research and polished the manuscript. W.Y. and H.M. helped organize the paper. J.Z. and B.M. gave constructive advice for the preparation of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Nos. 42201484 and 51975468).

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

The research in this article was supported by the National Natural Science Foundation of China (Grant Nos. 42201484 and 51975468), which is deeply appreciated.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Barnewold, L.; Lottermoser, B.G. Identification of digital technologies and digitalisation trends in the mining industry. Int. J. Min. Sci. Technol. 2020, 30, 747–757. [Google Scholar] [CrossRef]
  2. Guofa, W. New technological progress of coal mine intelligence and its problems. Coal Sci. Technol. 2022, 50, 1–27. [Google Scholar]
  3. Vassena, G.; Clerici, A. Open pit mine 3D mapping by tls and digital photogrammetry: 3D model update thanks to a slam based approach. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 42, 1145–1148. [Google Scholar] [CrossRef] [Green Version]
  4. Ebadi, K.; Chang, Y.; Palieri, M.; Stephens, A.; Hatteland, A.; Heiden, E.; Thakur, A.; Funabiki, N.; Morrell, B.; Wood, S.; et al. LAMP: Large-Scale Autonomous Mapping and Positioning for Exploration of Perceptually-Degraded Subterranean Environments. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), virtually, 31 May–31 August 2020; pp. 80–86. [Google Scholar]
  5. Chang, Y.; Ebadi, K.; Denniston, C.E.; Ginting, M.F.; Rosinol, A.; Reinke, A.; Palieri, M.; Shi, J.; Chatterjee, A.; Morrell, B.; et al. LAMP 2.0: A Robust Multi-Robot SLAM System for Operation in Challenging Large-Scale Underground Environments. IEEE Robot. Autom. Lett. 2022, 7, 9175–9182. [Google Scholar] [CrossRef]
  6. Qin, T.; Li, P.; Shen, S. Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef] [Green Version]
  7. Zhao, S.; Zhang, H.; Wang, P.; Nogueira, L.; Scherer, S. Super odometry: IMU-centric LiDAR-visual-inertial estimator for challenging environments. In International Conference on Intelligent Robots and Systems (IROS); IEEE: Piscataway, NI, USA, 2021; pp. 8729–8736. [Google Scholar]
  8. Lin, J.; Zheng, C.; Xu, W.; Zhang, F. R2LIVE: A Robust, Real-Time, LiDAR-Inertial-Visual Tightly-Coupled State Estimator and Mapping. IEEE Robot. Autom. Lett. 2021, 6, 7469–7476. [Google Scholar] [CrossRef]
  9. Deren, L. Towards geospatial information technology in 5G/6G era. Acta Geod. Et Cartogr. Sin. 2019, 48, 1475–1481. [Google Scholar]
  10. Ruizhi, C.; Lei, W.; Deren, L.; Liang, C.; Wenju, F. A survey on the fusion of the navigation and the remote sensing techniques. Acta Geod. Et Cartogr. Sin. 2019, 48, 1507–1522. [Google Scholar]
  11. Yongjun, Z.; Zuxun, Z.; Jianya, G. Generalized photogrammetry of spaceborne, airborne and terrestrial multi-source remote sensing datasets. Acta Geod. Et Cartogr. Sin. 2021, 50, 1–11. [Google Scholar]
  12. Huang, K.; Shi, B.; Li, X.; Li, X.; Huang, S.; Li, Y. Multi-modal Sensor Fusion for Auto Driving Perception: A Survey. arXiv 2022, arXiv:2202.02703. [Google Scholar]
  13. Zhang, J.; Singh, S. LOAM: Lidar Odometry and Mapping in Real-time. Robot. Sci. Syst. 2014, 2, 1–9. [Google Scholar]
  14. Shan, T.; Englot, B. LeGO-LOAM: Lightweight and Ground-Optimized Lidar Odometry and Mapping on Variable Terrain. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4758–4765. [Google Scholar]
  15. Hess, W.; Kohler, D.; Rapp, H.; Andor, D. Real-time loop closure in 2D LIDAR SLAM. In Proceedings of the IEEE International Conference on Robotics and Automation(ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 1271–1278. [Google Scholar]
  16. Koide, K.; Yokozuka, M.; Oishi, S.; Banno, A. Globally Consistent and Tightly Coupled 3D LiDAR Inertial Mapping. arXiv 2022, arXiv:2202.00242. [Google Scholar]
  17. Xu, W.; Zhang, F. Fast-lio: A fast, robust lidar-inertial odometry package by tightly-coupled iterated kalman filter. IEEE Robot. Autom. Lett. 2021, 6, 3317–3324. [Google Scholar] [CrossRef]
  18. Ye, H.; Chen, Y.; Liu, M. Tightly Coupled 3D Lidar Inertial Odometry and Mapping. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 3144–3150. [Google Scholar]
  19. Qin, C.; Ye, H.; Pranata, C.E.; Han, J.; Liu, M. LINS: A Lidar-Inertial State Estimator for Robust and Efficient Navigation. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), virtually, 31 May–31 August 2020; pp. 8899–8906. [Google Scholar]
  20. Li, K.; Li, M.; Hanebeck, U.D. Towards High-Performance Solid-State-LiDAR-Inertial Odometry and Mapping. IEEE Robot. Autom. Lett. 2021, 6, 5167–5174. [Google Scholar] [CrossRef]
  21. Shan, T.; Englot, B.; Meyers, D.; Wang, W.; Rus, D. LIO-SAM: Tightly-coupled Lidar Inertial Odometry via Smoothing and Mapping. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 5135–5142. [Google Scholar]
  22. Kim, H.; Choi, Y. Comparison of three location estimation methods of an autonomous driving robot for underground mines. Appl. Sci. 2020, 10, 4831. [Google Scholar] [CrossRef]
  23. Miller, I.D.; Cladera, F.; Cowley, A.; Shivakumar, S.S.; Lee, E.S.; Jarin-Lipschitz, L.; Bhat, A.; Rodrigues, N.; Zhou, A.; Cohen, A.; et al. Mine tunnel exploration using multiple quadrupedal robots. IEEE Robot. Autom. Lett. 2020, 5, 2840–2847. [Google Scholar] [CrossRef] [Green Version]
  24. Mascarich, F.; Khattak, S.; Papachristos, C.; Alexis, K. A multi-modal mapping unit for autonomous exploration and mapping of underground tunnels. In Proceedings of the IEEE Aerospace Conference, Big Sky, MT, USA, 3–10 March 2018; pp. 1–7. [Google Scholar]
  25. Tian, R.; Zhang, Y.; Feng, Y.; Yang, L.; Cao, Z.; Coleman, S.; Kerr, D. Accurate and robust object SLAM with 3D quadric landmark reconstruction in outdoors. IEEE Robot. Autom. Lett. 2021, 7, 1534–1541. [Google Scholar] [CrossRef]
  26. Shen, B.; Chen, Y.; Yang, H.; Zhan, J.; Sun, Y.; Xiong, R.; Dai, S.; Wang, Y. Anti-degenerated UWB-LiDAR Localization for Automatic Road Roller in Tunnel. arXiv 2021, arXiv:2109.10513. [Google Scholar]
  27. Hongwei, M.; Yan, W.; Lin, Y. Research on depth vision based mobile robot autonomous navigation in underground coal mine. J. China Coal Soc. 2020, 45, 2193–2206. [Google Scholar]
  28. Johannsson, H.; Kaess, M.; Fallon, M.; Leonard, J.J. Temporally Scalable Visual SLAM using a Reduced Pose Graph. In Proceedings of the IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 54–61. [Google Scholar]
  29. Tardioli, D.; Villarroel, J.L. Odometry-less localization in tunnel-like environments. In Proceedings of the IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC), Espinho, Portugal, 14–15 May 2014; pp. 65–72. [Google Scholar]
  30. Pathak, K.; Birk, A.; Vaškevičius, N.; Poppinga, J. Fast Registration Based on Noisy Planes With Unknown Correspondences for 3-D Mapping. IEEE Trans. Robot. 2010, 26, 424–441. [Google Scholar] [CrossRef] [Green Version]
  31. Li, R.; Qiang, L.; Gui, J.; Gu, D.; Hu, H. A novel RGB-D SLAM algorithm based on points and plane-patches. In Proceedings of the IEEE International Conference on Automation Science and Engineering (CASE), Fort Worth, TX, USA, 21–25 October 2016; pp. 1348–1353. [Google Scholar]
  32. Proença, P.F.; Gao, Y. Probabilistic Combination of Noisy Points and Planes for RGB-D Odometry. Lect. Notes Comput. Sci. 2017, 10454, 340–350. [Google Scholar]
  33. Raposo, C.; Lourenco, M.; Antunes, M.; Barreto, J. Plane-based Odometry using an RGB-D Camera. BMVC 2013, 2, 6. [Google Scholar]
  34. Sun, Q.; Jing, Y.; Zhang, X.; Sun, F. RGB-D SLAM in Indoor Environments with STING-Based Plane Feature Extraction. IEEE/ASME Trans. Mechatron. 2017, 23, 1071–1082. [Google Scholar] [CrossRef]
  35. Taguchi, Y.; Jian, Y.D.; Ramalingam, S.; Feng, C. Point-plane SLAM for hand-held 3D sensors. In Proceedings of the IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 5182–5189. [Google Scholar]
  36. Aghili, F.; Su, C.-Y. Robust Relative Navigation by Integration of ICP and Adaptive Kalman Filter Using Laser Scanner and IMU. IEEE/ASME Trans. Mechatronics. 2016, 21, 2015–2026. [Google Scholar] [CrossRef]
  37. Chen, S.; Zhou, B.; Jiang, C.; Xue, W.; Li, Q. A LiDAR/Visual SLAM Backend with Loop Closure Detection and Graph Optimization. Remote Sens. 2021, 13, 2720. [Google Scholar] [CrossRef]
  38. Chan, S.H.; Wu, P.T.; Fu, L.C. Robust 2D Indoor Localization Through Laser SLAM and Visual SLAM Fusion. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics (SMC), Miyazaki, Japan, 7–10 October 2018; pp. 1263–1268. [Google Scholar]
  39. Ding, X.; Wang, Y.; Li, D.; Tang, L.; Yin, H.; Xiong, R. Laser map aided visual inertial localization in changing environment. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 4794–4801. [Google Scholar]
  40. Kim, H.; Choi, Y. Location estimation of autonomous driving robot and 3D tunnel mapping in underground mines using pattern matched LiDAR sequential images. Int. J. Min. Sci. Tech. 2021, 31, 779–788. [Google Scholar] [CrossRef]
  41. He, X.; Gao, W.; Sheng, C.; Zhang, Z.; Pan, S.; Duan, L.; Zhang, H.; Lu, X. LiDAR-Visual-Inertial Odometry Based on Optimized Visual Point-Line Features. Remote Sens. 2022, 14, 622. [Google Scholar] [CrossRef]
  42. Zhang, L.; Liu, Z.Y.; Cao, J.Y.; Shen, P.Y.; Jiang, D.Z.; Mei, L.; Zhu, G.-M.; Mian, Q.-G. Cartographer algorithm and system implementation based on enhanced pose fusion of sweeping robot. Ruan Jian Xue Bao/J. Software. 2020, 31, 2678–2690. [Google Scholar]
  43. Zhang, J.; Kaess, M.; Singh, S. On degeneracy of optimization-based state estimation problems. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 809–816. [Google Scholar]
  44. Lv, J.; Xu, J.; Hu, K.; Liu, Y.; Zuo, X. Targetless Calibration of LiDAR-IMU System Based on Continuous-time Batch Estimation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 9968–9975. [Google Scholar]
  45. Forster, C.; Carlone, L.; Dellaert, F.; Scaramuzza, D. On-Manifold Preintegration for Real-Time Visual--Inertial Odometry. IEEE Trans. Robot. 2016, 33, 1–21. [Google Scholar] [CrossRef] [Green Version]
  46. Zhang, J.; Singh, S. Low-drift and Real-time Lidar Odometry and Mapping. Auton. Robot. 2017, 41, 401–416. [Google Scholar] [CrossRef]
  47. Kaess, M.; Johannsson, H.; Roberts, R.; Ila, V. iSAM2: Incremental smoothing and mapping using the Bayes tree. Int. J. Robot. Res. 2012, 31, 216–235. [Google Scholar] [CrossRef]
Figure 1. Data acquisition platform and environment.
Figure 1. Data acquisition platform and environment.
Remotesensing 15 00186 g001
Figure 2. Schematic diagram of the proposed method.
Figure 2. Schematic diagram of the proposed method.
Remotesensing 15 00186 g002
Figure 3. The extraction results of line and plane feature points in the preparation roadway. The red, white and yellow arrows are constraints in the three-axis direction.
Figure 3. The extraction results of line and plane feature points in the preparation roadway. The red, white and yellow arrows are constraints in the three-axis direction.
Remotesensing 15 00186 g003
Figure 4. The statistical result of degeneration factor.
Figure 4. The statistical result of degeneration factor.
Remotesensing 15 00186 g004
Figure 5. Schematic diagram of factor graph optimization by the proposed method.
Figure 5. Schematic diagram of factor graph optimization by the proposed method.
Remotesensing 15 00186 g005
Figure 6. Data collection environment, trajectory and mapping results in the indoor corridor. (ac) Trajectories and mappings of the proposed method, the LIO-SAM method and the LeGO-LOAM method, respectively.
Figure 6. Data collection environment, trajectory and mapping results in the indoor corridor. (ac) Trajectories and mappings of the proposed method, the LIO-SAM method and the LeGO-LOAM method, respectively.
Remotesensing 15 00186 g006
Figure 7. Data collection environment and the mapping result in the underground coal mine. (ad) The corresponding data collection environment. (e) The mapping result by the proposed method and the layout of reference points.
Figure 7. Data collection environment and the mapping result in the underground coal mine. (ad) The corresponding data collection environment. (e) The mapping result by the proposed method and the layout of reference points.
Remotesensing 15 00186 g007
Figure 8. Comparison of the trajectories between the proposed method, LeGO-LOAM and LIO-SAM.
Figure 8. Comparison of the trajectories between the proposed method, LeGO-LOAM and LIO-SAM.
Remotesensing 15 00186 g008
Figure 9. Comparison of mapping results between the proposed method and other methods. (ac) Mapping results by the proposed method, LIO-SAM and LeGO-LOAM, respectively.
Figure 9. Comparison of mapping results between the proposed method and other methods. (ac) Mapping results by the proposed method, LIO-SAM and LeGO-LOAM, respectively.
Remotesensing 15 00186 g009
Figure 10. Trajectories and mapping results by the proposed method and other methods in the preparation roadway. (ac) Trajectories and mapping results by the proposed method, LIO-SAM and LeGO-LOAM, respectively.
Figure 10. Trajectories and mapping results by the proposed method and other methods in the preparation roadway. (ac) Trajectories and mapping results by the proposed method, LIO-SAM and LeGO-LOAM, respectively.
Remotesensing 15 00186 g010
Figure 11. Absolute localization errors.
Figure 11. Absolute localization errors.
Remotesensing 15 00186 g011
Figure 12. The relative error of trajectory in three-axis direction.
Figure 12. The relative error of trajectory in three-axis direction.
Remotesensing 15 00186 g012
Figure 13. Time performance of the proposed method.
Figure 13. Time performance of the proposed method.
Remotesensing 15 00186 g013
Table 1. Specifications of data acquisition platform.
Table 1. Specifications of data acquisition platform.
EquipmentTypeSpecifications
LiDARVLP-16Scanning frequency: 10 Hz
Operating range: 0.2~150 m
IMUEllipse2-NOutput frequency: 200 Hz
Error: Roll/Pitch ± 0.1°,Yaw ± 0.5°
ControllerAutolabor PCCPU: AMD Ryzen3 3200 G
Memory: DDR4 8 GB
RobotAutolabor Pro1Driving mode:4WD Speed: 0.8 m/s
Applicable terrain: All terrain
Table 2. Quantitative comparison of the proposed method, LIO-SAM and LeGO-LOAM (m).
Table 2. Quantitative comparison of the proposed method, LIO-SAM and LeGO-LOAM (m).
LengthReferenceLeGO-LOAMLIO-SAMProposedLeGO-LOAM PercentageLIO-SAM PercentageProposed Percentage
L A B 38.8734.9538.4138.6810.08%1.18%0.49%
L B C 36.2532.1635.7936.0411.28%1.27%0.58%
L A _ A 0.00.920.200.070.61%0.13%0.05%
Table 3. The RMSE of absolute localization errors(m).
Table 3. The RMSE of absolute localization errors(m).
MethodXYZPosition
LeGO-LOAM0.5040.5330.6070.952
LIO-SAM0.1970.2650.1720.372
Proposed0.0840.1300.0440.161
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, X.; Lin, X.; Yao, W.; Ma, H.; Zheng, J.; Ma, B. A Robust LiDAR SLAM Method for Underground Coal Mine Robot with Degenerated Scene Compensation. Remote Sens. 2023, 15, 186. https://doi.org/10.3390/rs15010186

AMA Style

Yang X, Lin X, Yao W, Ma H, Zheng J, Ma B. A Robust LiDAR SLAM Method for Underground Coal Mine Robot with Degenerated Scene Compensation. Remote Sensing. 2023; 15(1):186. https://doi.org/10.3390/rs15010186

Chicago/Turabian Style

Yang, Xin, Xiaohu Lin, Wanqiang Yao, Hongwei Ma, Junliang Zheng, and Bolin Ma. 2023. "A Robust LiDAR SLAM Method for Underground Coal Mine Robot with Degenerated Scene Compensation" Remote Sensing 15, no. 1: 186. https://doi.org/10.3390/rs15010186

APA Style

Yang, X., Lin, X., Yao, W., Ma, H., Zheng, J., & Ma, B. (2023). A Robust LiDAR SLAM Method for Underground Coal Mine Robot with Degenerated Scene Compensation. Remote Sensing, 15(1), 186. https://doi.org/10.3390/rs15010186

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop