Next Article in Journal
Study of Obstacle-Crossing and Pitch Control Characteristic of a Novel Jumping Robot
Next Article in Special Issue
A Strip Adjustment Method of UAV-Borne LiDAR Point Cloud Based on DEM Features for Mountainous Area
Previous Article in Journal
Time Series Forecasting of Univariate Agrometeorological Data: A Comparative Performance Evaluation via One-Step and Multi-Step Ahead Forecasting Strategies
Previous Article in Special Issue
Point Cloud Semantic Segmentation Network Based on Multi-Scale Feature Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Coarse-to-Fine Registration Scheme for Mobile Laser Scanner Point Clouds Using Multiscale Eigenvalue Statistic-Based Descriptor †

School of Geospatial Information, PLA Strategic Support Force Information Engineering University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in: Fu, Y.; Li, Z.; Wang, W.; He, H.; Xiong, F.; Deng, Y. A robust coarse-to-fine pairwise registration based on multiple eigenvalues statistic. In Proceedings of the Sixth National LiDAR Conference, Beijing, China, 20–22 November 2020.
Sensors 2021, 21(7), 2431; https://doi.org/10.3390/s21072431
Submission received: 12 January 2021 / Revised: 27 March 2021 / Accepted: 28 March 2021 / Published: 1 April 2021
(This article belongs to the Special Issue Selected Papers from The Sixth National LiDAR Conference)

Abstract

:
To overcome the drawbacks of pairwise registration for mobile laser scanner (MLS) point clouds, such as difficulty in searching the corresponding points and inaccuracy registration matrix, a robust coarse-to-fine registration method is proposed to align different frames of MLS point clouds into a common coordinate system. The method identifies the correct corresponding point pairs from the source and target point clouds, and then calculates the transform matrix. First, the performance of a multiscale eigenvalue statistic-based descriptor with different combinations of parameters is evaluated to identify the optimal combination. Second, based on the geometric distribution of points in the neighborhood of the keypoint, a weighted covariance matrix is constructed, by which the multiscale eigenvalues are calculated as the feature description language. Third, the corresponding points between the source and target point clouds are estimated in the feature space, and the incorrect ones are eliminated via a geometric consistency constraint. Finally, the estimated corresponding point pairs are used for coarse registration. The value of coarse registration is regarded as the initial value for the iterative closest point algorithm. Subsequently, the final fine registration result is obtained. The results of the registration experiments with Autonomous Systems Lab (ASL) Datasets show that the proposed method can accurately align MLS point clouds in different frames and outperform the comparative methods.

1. Introduction

Point clouds obtained with modern three-dimensional (3D) sensors, such as mobile laser scanner (MLS), have played an important role in civil and transportation engineering [1,2,3], forest structure monitoring [4,5], and spatial deformation monitoring [6,7]. However, due to errors in the calibration and positioning of sensors, MLS point clouds obtained from different frames or periods suffer deviations, several tens of centimeters and even to meters [8]. This impedes the application of MLS point clouds, such as in change detection and deformation monitoring. Therefore, the point clouds in multiple frames or periods must be registered before using them in the application of deformation monitoring, urban management, and similar processes.
Numerous studies have been carried out on point clouds registration [9,10,11], which can be divided into two groups comprising pairwise and multiview registration, depending on the amount of input point clouds. Most pairwise and multiview point cloud registration methods employ a coarse-to-fine strategy [12]. In particular, coarse registration algorithms can be further divided into four categories [13], hand-crafted feature-based registration methods, deep learning-based registration methods, four-points congruent set (4PCS)-based registration method, and probabilistic registration methods. And the fine registration methods mainly include iterative closest point (ICP) [14,15,16] and normal distribution transform (NDT)-based algorithms [17,18,19].
In the process of hand-crafted feature-based registration, coarse registration algorithms, such as random sample consensus (RANSAC) [20], are first applied to estimate the initial transformation between two adjacent point clouds. Next, fine registration algorithms, such as ICP, are utilized to refine the approximate rotation matrix and translation vector. The core of hand-crafted feature-based registration is the correspondence estimation in coarse registration, which is usually matched by the 3D surface feature descriptor. The traditional descriptors include 3D shape context (3DSC) [21], point feature histogram (PFH) [22], fast point feature histogram (FPFH) [23], signature of histogram of orientations (SHOT) [24], and binary shape context (BSC) [25]. Although these feature description languages are sufficiently descriptive, they also consume a lot of time during generation and matching owing to their higher dimensionality, i.e., 3DSC (1980 dimensions), SHOT (352 dimensions), and PFH (125 dimensions).
Compared with traditional descriptors, deep learning-based methods [26,27] can directly learn deep-level feature representations from a mass of data to achieve appropriate performance in terms of descriptiveness and robustness. This type of method has proven effective for the registration of indoor and small-scale point clouds; however, it is difficult to apply it to the registration of large-scale MLS and terrestrial laser scanner (TLS) point clouds because of the limitations related to the amount of data and complexity [13].
4PCS-based methods [28,29,30,31] achieve registration by repeating the following process to obtain the optimal solution: (1) randomly selecting geometrically consistent point pairs; (2) computing the registration matrix and the root mean square error (RMSE) between two-point clouds. The 4PCS scheme works well for datasets with small overlaps, and it requires no assumptions regarding the initial positions. However, the iterative procedure of matching the correspondences and rejecting the mismatched ones is time-consuming.
Coherent point drift (CPD)-based methods [32,33,34], which represents the probabilistic registration method, consider registration as a probability density estimation problem. They first use the Gaussian mixture models (GMM) centroids to describe the source point cloud, and then fit the GMM to the target point cloud by maximizing the likelihood of the objective function. These methods exhibit generality, accuracy, and robustness to noise and outliers. However, because the registration result depends on the sampling result, the method cannot simultaneously deal with large volume points.
The representative fine registration method is the ICP and the NDT algorithm. This type of method can achieve high-quality and high-precision registration by repeating point matching and transformation calculations. However, both ICP and NDT-based methods require a better initial transformation matrix to avoid congregation of points at a local optimum.
In this study, we defined a 3D surface feature descriptor using multiscale eigenvalue statistic to estimate the corresponding points between the MLS point clouds obtained from different frames. With these estimated correspondences, we performed coarse registration. Then, the result of the coarse registration was used as the initial value for the fine registration, such as ICP, to obtain a better result. The major contributions of this study are summarized as follows:
1.
A new 3D local descriptor with fewer dimensions (21 dimensions) was proposed to describe the keypoint under multiscale support radii.
2.
The proposed descriptor was further used to identify the corresponding points from the different frames of MLS point clouds.
The remainder of this manuscript is organized as follows: The proposed MEVS descriptor is defined in Section 2; Section 3 describes the coarse-to-fine registration scheme; and Section 4 introduces the experiments. The study is concluded in Section 5.

2. Multiscale Eigenvalues Statistic-Based Descriptor

In this section, we defined a novel 3D local feature descriptor, named the multiscale eigenvalue statistic (MEVS), to describe the keypoint using multiple eigenvalues obtained under multiscale support radii. The procedure has three main steps: (1) computing point-density and Euclidean-distance-related weights, (2) constructing the weighted covariance matrix, and (3) constructing the MEVS descriptor.
The keypoint and its neighbors are regarded as N = { q 0 , q 1 , q 2 , q m } , where q 0 represents the keypoint, q j ( j = 1 , 2 , m ) represents the j-th nearest neighbor, and m is the number of neighboring points. The covariance matrix constructed for q 0 using N is denoted by C, and the support radius for neighboring points searching is expressed as r scale .

2.1. Weight Assignment

First, we estimated the point-density-related weight w j density by Equation (1), the aim is to describe the surface shape of the point better
w j density = 1 # p n : p n q j < r density ,
where r density represents the radius for point-density estimation, p n is the neighboring point of q j , p n q j represents the 3D Euclidean distance between q j and p n , and # p n : p n q j < r density represents the number of points within r density , if it is equal to zero, then we just set w j density =1. By varying r density from 0.1 r scale to r scale in intervals of 0.1 r scale , we found that the MEVS descriptor performed best at an r density of 0.5 r scale by testing. Thus, we set r density = 0.5 r scale . The weight w j density was used to compensate for varying point density; thus, the points in regions with low point density contribute more than those in the dense regions.
Next, the Euclidean-distance-related weight w j distance was calculated by Equation (2)
w j distance = r scale q 0 q j r scale .
The weight w j distance is expected to improve the robustness of the MEVS descriptor, for the distant points contribute less to the overall covariance matrix.

2.2. Weighted Covariance Matrix

By using the point set N = { q 0 , q 1 , q 2 , q m } and the corresponding weights for q j , we calculated the weighted covariance matrix C as follows
C = 1 q j N w j density · w j distance q j N w j density · w j distance q 0 q j q 0 q j T .

2.3. MEVS Descriptor

Eigenvalues { λ 1 , λ 2 , λ 3 } in the decreasing order of magnitude were obtained by decomposing the weighted covariance matrix C and further normalized by Equation (4)
λ C i = λ i / i = 1 3 λ i .
It is assumed that the initial support radius is R, by varying R = R + j · m r ( j = 1 , 2 , k ) , where k represents the multiscale dimensions and mr represents the mean resolution of the point cloud (in this paper, its unit is meter), we calculated a set of weighted covariance matrices C j and consequently obtain various combinations of eigenvalues. Through normalization, we set E V j = λ 1 C j , λ 2 C j , λ 3 C j . Then, the MEVS descriptor was defined by Equation (5)
M E V S = E V 1 , E V 2 , E V k .
The MEVS descriptor is expected to be highly viewpoint invariant, because each covariance matrix C j is real and symmetric, so that its eigenvalues do not change when the point set N j is rotated, and each covariance matrix C j remains the same when the point set N j is translated. Because C j is computed from data points, which change and also the density of which changes when the same area is scanned from different viewpoints, the MEVS descriptor is not perfectly viewpoint invariant, but the density-based weighting can improve the viewpoint invariance.

2.4. MEVS Generation Parameters

The MEVS feature descriptor had two important parameters: (1) initial support radius R and (2) multiscale dimensions k. The performance of the MEVS descriptor under different settings of the two parameters was tested on the tuning datasets using a precision versus recall curve (PR curve) [35].

2.4.1. PR Curve Generation

Given a model point cloud, a scene point cloud, and the ground-truth transformation T between them, the PR curve was calculated as follows:
1.
A number of keypoints were detected from both the model and scene point clouds using the keypoint detector.
2.
The proposed MEVS feature descriptor for each keypoint was computed using the proposed method.
3.
The nearest neighbor distance ratio (NNDR) technique was used to match the feature descriptors.
Specifically, the nearest and second nearest neighbors for each MEVS descriptor MEVS p i model in the model point cloud were selected from the scene point cloud, which are denoted by MEVS p j scene and MEVS p j scene , respectively. Then, the ratio between the two distances was calculated as MEVS p i model MEVS p j scene / MEVS p i model MEVS p j scene . If the distance ratio was less than a threshold τ , the two feature descriptors MEVS p i model and MEVS p j scene were considered an estimated match, and their corresponding points p i model and p j scene were considered a corresponding point pair. Furthermore, the correspondence was assumed a correct match, if the distance p i model T · p j scene was less than a predefined threshold (i.e., half of the initial support radius of keypoint p i model in this study). Otherwise, it was assigned a false match. Finally, the precision of match assignment was calculated as the number of correct matches with respect to the total number of estimated matches, as in Equation (6)
P r e c i s i o n = The   number   of   correct   matches   The   number   of   estimated   mathces .
Recall was calculated as the number of correct matches with respect to the number of ground-truth corresponding points between the given scene and model point cloud, as in Equation (7)
R e c a l l = The   number   of   correct   matches   The   number   of   ground - truth   corresponding   points .
The PR curve can be obtained by varying the threshold τ from 0 to 1. Ideally, the PR curve should be located on the top-right corner of the precision-recall coordinate system, and the larger the area under the PR curve, the more descriptive the descriptor. We tested the performance of the descriptor by examining the different combinations of the two main parameters. By calculating the area under the PR curve, as shown in Table 1, we found that the optimal value of the initial support radius R was 12 mr, and the optimal value of the multiscale dimension k was 7.

2.4.2. Initial Support Radius

The initial support radius is an important parameter for the generation of the MEVS feature descriptor because it determines both the descriptiveness and robustness of the descriptor. We tested the performance of the MEVS descriptor with respect to varying initial support radii, with the multiscale dimension set at k = 7. Figure 1 illustrates the generated PR curves under the initial support radius R ranging from 5 mr to 15 mr.
As shown in Figure 1, the MEVS local feature descriptor generated with a small initial support radius R (e.g., R = 5 mr) cannot eliminate the effect of noise. Hence, the descriptor generated with a small R is less robust. Moreover, a small initial support radius allowed the generator to consider less local information, leading to its low descriptiveness. A large initial support radius R (e.g., R = 15 mr) is more sensitive to occlusion, thereby reducing the descriptiveness of the generated descriptor. We found that the MEVS generated with R = 12 mr can optimize local surface shape information and robustness. Therefore, in practice, we used R = 12 mr as the initial support radius.

2.4.3. Multiscale Dimension

The multiscale dimension, which determines the maximum search region and dimension of the MEVS descriptor, is another important parameter determining the robustness and descriptiveness of the descriptor. We tested the performance of the descriptor by varying the values of the multiscale dimension, with the initial support radius set to R = 12 mr. Figure 2 illustrates the PR curves generated under different multiscale dimensions k.
As shown in Figure 2, the descriptiveness and robustness of the generated MEVS feature descriptor first increase with an increase in the multiscale dimension k (e.g., from 3 to 7), and then weaken with the further increase of k (e.g., from 7 to 15). This phenomenon occurs because k determines the maximum range of the neighborhood search, which directly affects the descriptiveness and robustness of the local descriptor. To improve the descriptiveness, a smaller support radius should be used. Meanwhile, to enhance the robustness of the MEVS descriptor, the support radius should be increased appropriately, but not so much that it increases the sensitivity of the descriptor to occlusion and confusion. Therefore, we used k = 7 as the multiscale dimension.

3. Coarse-to-Fine Pairwise Registration

The proposed coarse-to-fine pairwise registration scheme primarily included the following steps: (1) correspondences estimation, (2) mismatches rejection, and (3) registration calculation. Figure 3 demonstrates the specific process of the proposed algorithm.

3.1. Correspondences Estimation

To improve the efficiency, some keypoints were extracted from the original point cloud using a keypoint detector, such as 2.5D SIFT [36,37], 3D Harris [38], NARF [39], and intrinsic shape signatures (ISS) [40]. As shown in [41] that the ISS detector performed the best, we adopted the ISS algorithm to detect the keypoints.
A bidirectional matching strategy [42] was applied in which the keypoints extracted from the source and target point clouds were regarded as K P S = p 1 s , p 2 s , p T s s and K P T = p 1 t , p 2 t , p T t t , respectively. The corresponding MEVS was M S = m 1 s , m 2 s , m T s s and M T = m 1 t , m 2 t , m T t t . For each m k s from M S , if there exists an element m h t in M T that satisfies the constraint as in Equation (8), then m k s and m h t together were considered matched descriptors. In other words, only when m h t is the nearest descriptor to m k s in M T , and m k s to m h t in M S , then m k s and m h t were the corresponding descriptors, and the corresponding points were regarded the matched point pair
h = arg min m k s m n t ( n = 1 , 2 , N t ) k = arg min m h t m n s ( n = 1 , 2 , N s ) ,
where the norm is the 21D Euclidean distance.
All the matched point pairs calculated by the above procedure were set as F C = c 1 , c m , c M F C , where c m = m m s , m m t represents the m-th corresponding point pair, and M F C is the length of F C .

3.2. Mismatches Rejection

In principle, after obtaining FC, the transformation from the source to target point clouds can be directly calculated; however, there were some mismatches in FC, which lowed the accuracy of registration, even leading to incorrect result. Therefore, we introduced geomatic consistency constraint [43] to remove the mismatches from FC. If two correspondences in FC, named c m = m m s , m m t and c n = m n s , m n t , satisfy Equation (9), then they are regarded as the right corresponding point pairs
a b s p m s p n s p m t p n t < ε ,
where p m s , p m t , p n s , and p n t are the points corresponding to the descriptors m m s , m m t , m n s , and m n t , respectively. ε is a threshold, which, in this study, was set as 5 mr. a b s · represents the absolute value and   is the 3D Euclidean distance between the two points.
For each c m = m m s , m m t in FC, we traverse FC to determine the correspondences that satisfy Equation (9), and the results combined with c m = m m s , m m t are regarded as a group. Finally, we obtained M F C groups. The more elements in a group, the more right corresponding point pairs it may contain. Thus, we chose the largest group as the final matched point pairs G C = c 1 , c m , c M G C .

3.3. Registration Calculation

The registration adopts a coarse-to-fine strategy. First, coarse registration is performed using the point set GC. Second, the transformation obtained by coarse registration is further refined by the ICP algorithm. The coarse registration uses the corresponding points to estimate roughly yet quickly the transformation between the source and target point clouds. By iteratively random sampling GC, we can get a series δ m according to Equation (10), when δ m is less than a threshold, the coarse registration matrix T s , t C is obtained
δ m = j = 1 L m T s , t C · p j s p j t 2 ,
where p j s and p j t are the j-th corresponding points in GC, T s , t C is the coarse registration matrix, L m is the length of a sample of GC, and δ m is the sum of squares of the Euclidean distance between the corresponding points after transformed to the same coordinate system.
The goal here is not pairwise coarse registration itself but to provide a robust, reliable initial transformation for fine registration [44]. Therefore, T s , t C was further refined by the ICP algorithm to obtain a better registration matrix T s , t F , which is regarded as the final transformation between source and target point clouds.

4. Experiments and Analysis

In this section, we test the performance of the proposed coarse-to-fine registration method by evaluating the effectiveness of the keypoint matching using MEVS for registration. Accordingly, we design Method1, in which we use the RANSAC algorithm to directly address the extracted keypoints to behave the coarse registration with the unit matrix as the initial value. This value initialized the ICP algorithm to address the source and target point clouds and obtain the fine registration matrix.
Test the advantages and disadvantages of keypoint matching using MEVS for registration, we design Method2 and Method3. In Method2, we used the sample consensus initial alignment (SAC-IA) [23] algorithm with the FPFH descriptor to address keypoints to obtain the coarse registration; In Method3, we used the corresponding points from the source and target point clouds identified by the SHOT descriptors to obtain the coarse registration with RANSAC algorithm; And the coarse registration was further refined using the ICP algorithm. As suggested by [12], we set the parameter of the support radius for FPFH and SHOT as 15 mr.
All the experiments were implemented on a ThinkPad X1 Extreme laptop with an Intel Core i7-9750h CUP @ 2.6 GHz clock speed and 16 GB RAM.

4.1. Data Description

Four datasets from the ASL Datasets Repository [45] were used to test and evaluate the performance of the proposed and existing methods. These data sets were collected to verify registration algorithms for point clouds obtained in specific environments and conditions. The different point clouds are characterized by diverse environments and geometric primitives. And the data were collected used a custom-made rotating scanner (Hokuyo UTM-30LX), and its precision is about ± 3 cm, and the ground-truth was obtained by TS15, a theodolite from Leica Geosystems, as shown in Figure 4. Figure 5 shows the original four datasets used in this study. Figure 6 summarizes the registration results by the ground-truth (the ground-truth are listed in Table 2), in which the different color represents the point clouds in different frames. Table 3 details the information about the four datasets.

4.2. Evaluation Criteria

The criteria followed to evaluate for the performance of the proposed and comparative methods are the measurements of rotation error, translation error, and registration efficiency, which are commonly used for the evaluation of point cloud registration [46,47].
Given a source point cloud P s , the transformation T s , t from P s to the target point cloud P t can be calculated using the proposed and comparative methods. The residual transformation is Δ T s , t , defined as
Δ T s , t = T s , t ( T s , t G ) 1 = Δ R s , t Δ t s , t 0 1 ,
where T s , t is the estimated transformation from P s to P t , and T s , t G is the corresponding ground-truth transformation.
Then, the rotation error e s , t r and translation error e s , t t form P s to P t were calculated based on their corresponding rotation component Δ R s , t and translation component Δ t s , t , as follows
e s , t r = arccos t r ( Δ R s , t ) 1 2 e s , t t = Δ t s , t ,
where t r ( Δ R s , t ) denotes the trace of Δ R s , t , and the rotation error e s , t r corresponds to the angle of rotation in the axis-angle representation.

4.3. Results and Discussion

4.3.1. Keypoints Processing

To improve the efficiency of registration, some keypoints were first extracted from the source and target point clouds using the ISS [40] algorithm. Figure 7 illustrates the results of the keypoint extraction. The matching correspondences estimated by the method proposed in Section 3 are shown in Figure 8. Further details on keypoint processing are listed in Table 4.
The comprehensive analyses of Figure 7 and Figure 8 and Table 4 reveal that (1) the ISS algorithm can extract some significant points from the original point cloud, and (2) MEVS can determine accurate correspondences from the keypoints of the source and target point clouds, and provide high-quality corresponding points for the subsequent rough registration. However, the method still has the following shortcomings: (1) As shown in Figure 7, the spatial distribution of keypoints extracted by the ISS algorithm is not appropriate. For example, the keypoints of the Apartment dataset are concentrated in a relatively dense house, whereas the corridor contains hardly any points; (2) As shown in Figure 8, a few mismatches remain; and (3) As listed in Table 4, the number of final matched keypoints is relatively small, accounting for only 30–40% of the total keypoints. These limitations should be addressed in future studies.

4.3.2. Coarse-to-Fine Registration

Figure 9 illustrates the results of registration for the four ASL datasets using the proposed and comparative methods.
As shown by the qualitative testing results in Figure 9, the proposed algorithm can obtain better registration results for the four challenging public datasets, which indicates that it is highly feasible for the registration of datasets obtained from different scenes.
Table 5 lists the rotation and translation errors of coarse and fine registration by different methods.
The following conclusions can be drawn from the analysis of Table 5: (1) The proposed method can achieve suitable registration results between the MLS point clouds in two frames regardless of the type of data [indoor data (Apartment), outdoor data (Wood in Summer and Wood in Autumn), or mixed data (Stairs)]. The rotation and translation have accuracies of more than 0.04 rad and 0.08 m, respectively. (2) The proposed algorithm has a smaller registration error than Method1, indicating that MEVS can effectively describe the local neighborhood information of keypoints. Moreover, it can accurately match keypoints extracted from the source and target point clouds with one another. (3) For the Wood in Autumn dataset, the final registration of the proposed method is slightly inferior to that of Method2 in terms of translation error, but better in terms of the rotation error, so the two methods perform similarly with this dataset. For the Apartment dataset, the final registration of the proposed method is slightly inferior to that of Method3 in terms of translation error, but better in terms of the rotation error, so they perform similarly with Apartment. For the other datasets, the proposed method registers better than Method2 and Method3, which indicates that the MEVS matches keypoints better than FPFH and SHOT. In other words, MEVS is more descriptive and makes less mismatches than FPFH and SHOT, the reason is that it uses a series of radii for the ball nearest neighbor searching, rather than just one radius, to calculate the descriptor, so it can distinguish the correspondences much better. (4) The proposed algorithm can accurately register the data of Wood in Summer and Wood in Autumn, indicating that it is not affected by the change of season. (5) With a rotation error of 0.0316 rad and a translation error of 0.078 m, the proposed method performed the worst with the Apartment dataset, which may have been caused by the dynamic changes during the scanning of the Apartment dataset. With a rotation error of 0.0090 rad and a translation error of 0.023 m, this method performed the best with the Stairs dataset, the reason may be the good spatial distribution of the keypoints.
All of the four methods use ICP for refinement, so it is necessary to consider the efficiency of registration, which is measured in terms of time consumption. Table 6 lists the time consumed by the four methods.
Table 6 reveals that the proposed method required the least number of iterations and computation time, which indicates that it can yield highly reliable coarse registration results as input for ICP or other iterative fine-registration algorithms.

5. Conclusions

Thus far, it has been difficult to determine accurate corresponding points and achieve reliable registration for different frames of MLS point cloud. To solve this problem, a new 3D local descriptor with fewer dimensions (21 dimensions) was proposed to describe the keypoint under multiscale support radii, which was further used to estimate the correspondences from the MLS point clouds in different frames. With these correspondences, we proposed a coarse-to-fine registration scheme for the MLS point cloud from the pairwise frames. First, coarse registration was conducted using the RANSAC algorithm with the correspondences. Next, the initial value calculated in coarse registration was refined by the ICP algorithm to obtain an optimal one. The proposed coarse-to-fine registration scheme achieved globally optimal registration for four experimental datasets, with maximum rotation and translation errors of 0.0316 rad and 0.078 m, respectively, and minimum rotation and translation errors of 0.009 rad and 0.023 m, respectively. Besides, it has good efficiency, coarsely aligning two-point clouds with 230,000 points (average) in each within 30 s, and refining them within 6 min. However, the ratio of keypoint matching with this method is slightly low, which should be the focus of future research.

Author Contributions

Conceptualization, Y.F. and Z.L.; methodology, Y.F.; software, Y.F.; validation, H.H., F.X. and W.W.; formal analysis, Z.L.; investigation, Y.D.; resources, Y.D.; data curation, Y.F.; writing—original draft preparation, Y.F.; writing—review and editing, Z.L.; visualization, F.X.; supervision, Z.L.; project administration, Y.D.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gehrung, J.; Hebel, M.; Arens, M.; Stilla, U. A framework for voxel-based global scale modeling of urban environments. Int. Arch. Photogram. Remote Sens. Spat. Info Sci. 2016, 42, 45–51. [Google Scholar] [CrossRef] [Green Version]
  2. Serna, A.; Marcotegui, B. Urban accessibility diagnosis from mobile laser scanning data. ISPRS J. Photogramm. Remote Sens. 2013, 84, 23–32. [Google Scholar] [CrossRef] [Green Version]
  3. Yang, B.; Dong, Z.; Liu, Y.; Liang, F.; Wang, Y. Computing multiple aggregation levels and contextual features for road facilities recognition using mobile laser scanning data. ISPRS J. Photogramm. Remote Sens. 2017, 126, 180–194. [Google Scholar] [CrossRef]
  4. Kelbe, D.; Van, A.J.; Romanczyk, P.; Van, L.M.; Cawse-Nicholson, K. Marker-free registration of forest terrestrial laser scanner data pairs with embedded confidence metrics. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4314–4330. [Google Scholar] [CrossRef]
  5. Polewski, P.; Yao, W.; Heurich, M.; Krzystek, P.; Stilla, U. Learning a constrained conditional random field for enhanced segmentation of fallen trees in ALS point clouds. ISPRS J. Photogramm. Remote Sens. 2018, 140, 33–44. [Google Scholar] [CrossRef]
  6. Prokop, A.; Panholzer, H. Assessing the capability of terrestrial laser scanning for monitoring slow moving landslides. Nat. Hazards Earth Syst. Sci. 2009, 9, 1921–1928. [Google Scholar] [CrossRef] [Green Version]
  7. Huang, R.; Ye, Z.; Boerner, R.; Yao, W.; Xu, Y.; Stilla, U. Fast pairwise coarse registration between point clouds of construction sites using 2D projection based phase correlation. Int. Arch. Photogram. Remote Sens. Spat. Info. Sci. 2019, XLII-2, 1015–1020. [Google Scholar] [CrossRef] [Green Version]
  8. Kukko, A.; Kaijaluoto, R.; Kaartinen, H.; Lehtola, V.; Jaakkola, A.; Hyyppä, J. Graph SLAM correction for single scanner MLS forest data under boreal forest canopy. ISPRS J. Photogramm. Remote Sens. 2017, 132, 199–209. [Google Scholar] [CrossRef]
  9. Liu, H.; Zhang, Y.; Lei, L.; Xie, H.; Li, Y.; Sun, S. Hierarchical Optimization of 3D Point Cloud Registration. Sensors 2020, 20, 6999. [Google Scholar] [CrossRef]
  10. Choi, O.; Park, M.-G.; Hwang, Y. Iterative K-Closest Point Algorithms for Colored Point Cloud Registration. Sensors 2020, 20, 5331. [Google Scholar] [CrossRef]
  11. Li, P.; Wang, R.; Wang, Y.; Gao, G. Fast Method of Registration for 3D RGB Point Cloud with Improved Four Initial Point Pairs Algorithm. Sensors 2020, 20, 138. [Google Scholar] [CrossRef] [Green Version]
  12. Guo, Y.; Sohel, F.; Bennamoun, M.; Lu, M.; Wan, J. Rotational projection statistics for 3D local surface description and object recognition. Int. J. Comput. Vision. 2013, 105, 63–86. [Google Scholar] [CrossRef] [Green Version]
  13. Dong, Z.; Liang, F.; Yang, B.; Xu, Y.; Zang, Y.; Li, J.; Wang, Y.; Dai, W.; Fan, H.; Hyyppä, J.; et al. Registration of large-scale terrestrial laser scanner point clouds: A review and benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 163, 327–342. [Google Scholar] [CrossRef]
  14. Besl, P.J.; McKay, N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar] [CrossRef]
  15. Ren, Z.; Wang, L.; Bi, L. Robust GICP-based 3D LiDAR SLAM for underground mining environment. Sensors 2019, 19, 2915. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Tazir, M.L.; Gokhool, T.; Checchin, P.; Malaterre, L.; Trassoudaine, L. Cluster ICP: Towards Sparse to Dense Registration. In Proceedings of the 15th International Conference on Intelligent Autonomous Systems, Baden-Baden, Germany, 11–15 June 2018; pp. 730–747. [Google Scholar]
  17. Takeuchi, E.; Tsubouchi, T. A 3-D scan matching using improved 3-D normal distributions transform for mobile robotic mapping. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–15 October 2006; pp. 3068–3073. [Google Scholar]
  18. Das, A.; Waslander, S.L. Scan registration with multi-scale k-means normal distributions transform. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012; pp. 2705–2710. [Google Scholar]
  19. Das, A.; Waslander, S.L. Scan registration using segmented region growing NDT. Int. J. Robot. Res. 2014, 33, 1645–1663. [Google Scholar] [CrossRef]
  20. Fischler, M.; Bolles, R. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1987, 24, 381–395. [Google Scholar] [CrossRef]
  21. Frome, A.; Huber, D.; Kolluri, R.; Bülow, T.; Malik, J. Recognizing objects in range data using regional point descriptors. In Proceedings of the 8th European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; pp. 224–237. [Google Scholar]
  22. Rusu, B.; Blodow, N.; Marton, Z.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
  23. Rusu, R.; Nico, B.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  24. Salti, S.; Tombari, F.; Stefano, L. SHOT: Unique signatures of histograms for surface and texture description. Comput. Vis. Image Underst. 2014, 125, 251–264. [Google Scholar] [CrossRef]
  25. Dong, Z.; Yang, B.; Liu, Y.; Liang, F.; Li, B.; Zang, Y. A novel binary shape context for 3D local surface description. ISPRS J. Photogramm. Remote Sens. 2017, 130, 431–452. [Google Scholar] [CrossRef]
  26. Li, W.; Wang, C.; Wen, C.; Zhang, Z.; Lin, C.; Li, J. Pairwise registration of TLS point clouds by deep multi-scale local features. Neurocomputing 2020, 386, 232–243. [Google Scholar] [CrossRef]
  27. Khazari, A.E.; Que, Y.; Sung, T.L.; Lee, H.J. Deep global features for point cloud alignment. Sensors 2020, 20, 4032. [Google Scholar] [CrossRef]
  28. Aiger, D.; Mitra, N.; Cohen-Or, D. Four-points congruent sets for robust surface registration. ACM Trans. Graph. 2008, 27, 1–10. [Google Scholar] [CrossRef] [Green Version]
  29. Mellado, N.; Aiger, D.; Mitra, N. Super 4PCS fast global point cloud registration via smart indexing. Comput. Graph. Forum. 2014, 3, 205–215. [Google Scholar] [CrossRef] [Green Version]
  30. Huang, J.; Kwok, T.-H.; Zhou, C. V4PCS: Volumetric 4PCS algorithm for global registration. J. Mech. Des. 2017, 139, 1–9. [Google Scholar] [CrossRef]
  31. Xu, Y.; Boerner, R.; Yao, W.; Hoegner, L.; Stilla, U. Pairwise coarse registration of point clouds in urban scenes using voxel-based 4-planes congruent sets. ISPRS J. Photogramm. Remote Sens. 2019, 151, 106–123. [Google Scholar] [CrossRef]
  32. Myronenko, A.; Song, X.; Carreira-Perpiñán, M.A. Non-rigid point set registration: Coherent point drift. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; pp. 1009–1016. [Google Scholar]
  33. Golyanik, V.; Taetz, B.; Reis, G.; Stricker, D. Extended coherent point drift algorithm with correspondence priors and optimal subsampling. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–9. [Google Scholar]
  34. Zang, Y.; Lindenbergh, R. An improved coherent point drift method for TLS point cloud registration of complex scenes. Int. Arch. Photogram. Remote Sens. Spat. Info. Sci. 2019, W13, 1169–1175. [Google Scholar] [CrossRef] [Green Version]
  35. Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J.; Kwok, N. A comprehensive performance evaluation of 3D local feature descriptors. Int. J. Comput. Vis. 2016, 116, 66–89. [Google Scholar] [CrossRef]
  36. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  37. Lo, T.; Siebert, J. Local feature extraction and matching on range images: 2.5D SIFT. Comput. Vis. Image Underst. 2009, 113, 1235–1250. [Google Scholar] [CrossRef]
  38. Sipiran, I.; Bustos, B. Harris 3D: A robust extension of the Harris operator for interest point detection on 3D meshes. Vis. Comput. 2011, 27, 963–976. [Google Scholar] [CrossRef]
  39. Steder, B.; Rusu, R.B.; Konolige, K.; Burgard, W. Point feature extraction on 3D range scans taking into account object boundaries. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Shanghai, China, 9–13 May 2011; pp. 2601–2608. [Google Scholar]
  40. Zhong, Y. Intrinsic shape signatures: A shape descriptor for 3D object recognition. In Proceedings of the IEEE 12th International Conference on Computer Vision Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 689–696. [Google Scholar]
  41. Zhou, R.; Li, X.; Jiang, W. 3D surface matching by a voxel-based buffer-weighted binary descriptor. IEEE Access 2020, 7, 86635–86650. [Google Scholar] [CrossRef]
  42. Yang, B.; Dong, Z.; Liang, F.; Liu, Y. Automatic registration of large-scale urban scene point clouds based on semantic feature points. ISPRS J. Photogramm. Remote Sens. 2016, 113, 43–58. [Google Scholar] [CrossRef]
  43. Chen, H.; Bhanu, B. 3D free-form object recognition in range images using local surface patches. Pattern Recognit. Lett. 2007, 28, 1252–1262. [Google Scholar] [CrossRef]
  44. Theiler, P.; Wegner, J.; Schindler, K. Globally consistent registration of terrestrial laser scans via graph optimization. ISPRS J. Photogramm. Remote Sens. 2015, 109, 126–138. [Google Scholar] [CrossRef]
  45. Pomerleau, F.; Liu, M.; Colas, F.; Siegwart, R. Challenging data sets for point cloud registration algorithms. Int. J. Robot. Res. 2012, 31, 1705–1711. [Google Scholar] [CrossRef] [Green Version]
  46. Dong, Z.; Yang, B.; Liang, F.; Huang, R.; Scherer, S. Hierarchical registration of unordered TLS point clouds based on binary shape context descriptor. ISPRS J. Photogramm. Remote Sens. 2018, 144, 61–79. [Google Scholar] [CrossRef]
  47. Petricek, T.; Svoboda, T. Point cloud registration from local feature correspondences—Evaluation on challenging datasets. PLoS ONE 2017, 12, e0187943. [Google Scholar] [CrossRef] [Green Version]
Figure 1. PR curves generated under different initial support radii R with the multiscale dimension k = 7.
Figure 1. PR curves generated under different initial support radii R with the multiscale dimension k = 7.
Sensors 21 02431 g001
Figure 2. PR curves generated under different multiscale dimension k with an initial support radius of R = 12 mr.
Figure 2. PR curves generated under different multiscale dimension k with an initial support radius of R = 12 mr.
Sensors 21 02431 g002
Figure 3. Flowchart of the proposed coarse-to-fine registration algorithm.
Figure 3. Flowchart of the proposed coarse-to-fine registration algorithm.
Sensors 21 02431 g003
Figure 4. The illustration of obtaining the ground-truth. (https://projects.asl.ethz.ch/datasets/doku.php?id=hardware:tiltinglaser, accessed on 8 March 2021).
Figure 4. The illustration of obtaining the ground-truth. (https://projects.asl.ethz.ch/datasets/doku.php?id=hardware:tiltinglaser, accessed on 8 March 2021).
Sensors 21 02431 g004
Figure 5. Original point clouds. (ad) is the dataset of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively. The red and green points represent the source and target point clouds, respectively.
Figure 5. Original point clouds. (ad) is the dataset of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively. The red and green points represent the source and target point clouds, respectively.
Sensors 21 02431 g005aSensors 21 02431 g005b
Figure 6. Results of point cloud registration by the ground-truth. (ad) is the registration results of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively. The red and green points represent the source and target point clouds, respectively.
Figure 6. Results of point cloud registration by the ground-truth. (ad) is the registration results of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively. The red and green points represent the source and target point clouds, respectively.
Sensors 21 02431 g006aSensors 21 02431 g006b
Figure 7. Results of keypoints extraction. The blue points represent the keypoints extracted by the intrinsic shape signatures (ISS) algorithm. The red and green points represent the source and target point clouds, respectively. (ad) present the datasets of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively.
Figure 7. Results of keypoints extraction. The blue points represent the keypoints extracted by the intrinsic shape signatures (ISS) algorithm. The red and green points represent the source and target point clouds, respectively. (ad) present the datasets of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively.
Sensors 21 02431 g007
Figure 8. Results of keypoints matching by multiscale eigenvalue statistic (MEVS). The blue points represent the keypoints. The red and green points represent the source and target point clouds, respectively. The line means the corresponding. (ad) present the matching results of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively.
Figure 8. Results of keypoints matching by multiscale eigenvalue statistic (MEVS). The blue points represent the keypoints. The red and green points represent the source and target point clouds, respectively. The line means the corresponding. (ad) present the matching results of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively.
Sensors 21 02431 g008
Figure 9. Results of registration. (ad) are the registration results of Method1, Method2, Method3, and our method, respectively. From left to right: registration results of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively. The red and green points represent the source and target point clouds, respectively.
Figure 9. Results of registration. (ad) are the registration results of Method1, Method2, Method3, and our method, respectively. From left to right: registration results of Apartment, Stairs, Wood in Summer, and Wood in Autumn, respectively. The red and green points represent the source and target point clouds, respectively.
Sensors 21 02431 g009
Table 1. The area under the precision versus recall (PR) curve generated by different combinations of R and k.
Table 1. The area under the precision versus recall (PR) curve generated by different combinations of R and k.
R = 5 mrR = 8 mrR = 10 mrR = 12 mrR = 15 mr
k = 30.5160.5870.5830.5830.560
k = 50.6320.7260.6960.7580.685
k = 70.7140.7380.7140.7780.630
k = 90.7360.7500.6700.7620.599
k = 110.7760.7230.6250.7010.449
k = 130.7650.6910.5120.6920.288
k = 150.7560.6600.3770.6370.156
Table 2. The ground-truth of the datasets used in this study.
Table 2. The ground-truth of the datasets used in this study.
Ground-TruthApartmentStairsWood in SummerWood in Autumn
Rotation matrix0.9931−0.11640.00580.98830.15240.00430.9843−0.1727−0.03610.9895−0.1438−0.1438
0.11640.99310.0034−0.15220.9878−0.03010.17260.9849−0.00350.14400.9891−0.0280
−0.0062−0.00270.9999−0.00880.02910.99950.0362−0.00270.9993−0.00490.02900.9995
Translation vector0.6148−0.01420.00850.4232−0.05080.06700.60570.04070.02690.49460.04960.0150
Table 3. Details of the four datasets.
Table 3. Details of the four datasets.
DatasetsFramesMean Point NumberSceneSpatial Scale (m)Dynamics
Apartment45365,000Indoors17 × 10 × 3Furniture moved between scans
Stairs31191,000Mixed21 × 111 × 27Non
Wood in Summer37182,000Outdoors30 × 53 × 20Seasonal changes
Wood in Autumn32178,000Outdoors36 × 60 × 22Seasonal changes
Table 4. Details of keypoints processing.
Table 4. Details of keypoints processing.
DatasetsOriginal Points’ NumberKeypoints’ NumberMatching Keypoints’ Number
Apartment370,260/370,2779140/10,4363102
Stairs185,850/181,0776829/70672280
Wood in Summer192,802/183,1794556/47211746
Wood in Autumn190,499/189,4414811/51301914
Table 5. Quantitative evaluation of the registration accuracy.
Table 5. Quantitative evaluation of the registration accuracy.
DatasetsMethodsCoarse RegistrationFine Registration
Rotation Error (rad)Translation Error (m)Rotation Error (rad)Translation Error (m)
ApartmentMethod10.07890.3360.05300.165
Method20.82441.0940.23770.601
Method30.23360.3650.03990.071
Our method0.04830.1450.03160.078
StairsMethod10.05030.1390.04910.139
Method20.22460.7260.06120.150
Method30.06020.1220.02140.149
Our method0.01240.0840.00900.023
Wood in SummerMethod10.05920.3300.02780.092
Method20.24761.2250.09570.640
Method30.25590.2960.02380.109
Our method0.06820.1320.02200.039
Wood in AutumnMethod10.06340.3040.05030.240
Method20.18610.5450.02800.050
Method30.31180.5050.05110.209
Our method0.05030.2400.01250.078
Table 6. Time consumption of the proposed and comparative methods.
Table 6. Time consumption of the proposed and comparative methods.
DatasetsMethodsCoarse Registration
Time (min)
Fine RegistrationTotal Time (min)
Time (min)Iterations
ApartmentMethod10.1017.711717.81
Method26.5135.053141.56
Method30.1229.682629.80
Our method0.2615.251515.21
StairsMethod10.056.39146.44
Method24.857.621712.47
Method30.063.51103.57
Our method0.952.0452.99
Wood in SummerMethod10.044.21144.25
Method22.889.953312.83
Method30.069.92339.98
Our method0.083.72133.80
Wood in AutumnMethod10.045.02155.06
Method22.1016.044818.14
Method30.058.49248.54
Our method0.103.73113.83
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fu, Y.; Li, Z.; Wang, W.; He, H.; Xiong, F.; Deng, Y. Robust Coarse-to-Fine Registration Scheme for Mobile Laser Scanner Point Clouds Using Multiscale Eigenvalue Statistic-Based Descriptor. Sensors 2021, 21, 2431. https://doi.org/10.3390/s21072431

AMA Style

Fu Y, Li Z, Wang W, He H, Xiong F, Deng Y. Robust Coarse-to-Fine Registration Scheme for Mobile Laser Scanner Point Clouds Using Multiscale Eigenvalue Statistic-Based Descriptor. Sensors. 2021; 21(7):2431. https://doi.org/10.3390/s21072431

Chicago/Turabian Style

Fu, Yongjian, Zongchun Li, Wenqi Wang, Hua He, Feng Xiong, and Yong Deng. 2021. "Robust Coarse-to-Fine Registration Scheme for Mobile Laser Scanner Point Clouds Using Multiscale Eigenvalue Statistic-Based Descriptor" Sensors 21, no. 7: 2431. https://doi.org/10.3390/s21072431

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop