Next Article in Journal
Remotely Powered Two-Wire Cooperative Sensors for Biopotential Imaging Wearables
Previous Article in Journal
Motion Analysis of the Extensor Carpi Ulnaris in Triangular Fibrocartilage Complex Injury Using Ultrasonography Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DOPNet: Achieving Accurate and Efficient Point Cloud Registration Based on Deep Learning and Multi-Level Features

School of Physical Science and Technology, Southwest Jiaotong University, Chengdu 622731, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(21), 8217; https://doi.org/10.3390/s22218217
Submission received: 8 August 2022 / Revised: 21 October 2022 / Accepted: 23 October 2022 / Published: 27 October 2022
(This article belongs to the Section Remote Sensors)

Abstract

:
Point cloud registration aims to find a rigid spatial transformation to align two given point clouds; it is widely deployed in many areas of computer vision, such as target detection, 3D localization, and so on. In order to achieve the desired results, registration error, robustness, and efficiency should be comprehensively considered. We propose a deep learning-based point cloud registration method, called DOPNet. DOPNet extracts global features of point clouds with a dynamic graph convolutional neural network (DGCNN) and cascading offset-attention modules, and the transformation is predicted by a multilayer perceptron (MLP). To enhance the information interaction between the two branches, the feature interaction module is inserted into the feature extraction pipeline to implement early data association. We compared DOPNet with the traditional method of using the iterative closest point (ICP) algorithm along with four learning-based registration methods on the Modelnet40 data set. In the experiments, the source and target point clouds were generated by sampling the original point cloud twice independently; we also conducted additional experiments with asymmetric objects. Further evaluation experiments were conducted with point cloud models from Stanford University. The results demonstrated that our DOPNet method outperforms these comparative methods in general, achieving more accurate and efficient point cloud registration.

1. Introduction

3D point clouds contain abundant spatial information. This kind of 3D data is significant in many areas, such as 3D target detection [1], 3D localization [2], and robotics [3]. Position mismatching is inevitable due to the inconsistency of acquisition angles and reference frames during the data acquisition process [4]. Therefore, point cloud registration is a fundamental task of point cloud data processing, and it is crucial for point cloud data applications to achieve highly accurate registration.
Current registration methods can be generally divided into two branches: traditional methods and learning-based ones. The classic traditional method is the iterative closest point (ICP) algorithm [5], which is simple in theory and can achieve suitable registration accuracy with a small initial spatial difference. However, ICP has non-negligible limitations, such as its sensitivity to the initial position of the point clouds, frequently falling into local minima, disappointing efficiency, and so on [6]. Researchers have successively proposed Go-ICP [7], normal iterative closest point (NICP) [8], generalized ICP (GICP) [9], and many other ICP-style algorithms to solve these problems, which do perform better than ICP, but these variants still fail to eliminate the inherent defects of ICP.
Benefiting from parameter-learning ability, learning-based registration algorithms are generally more efficient and effective than traditional algorithms. PointNet [10] and its improved version PointNet++ [11] can directly analyze point cloud data. PCRNet [12] constructs an iterative registration network based on PointNet, with simple structural complexity and satisfactory registration error. PointNetLK [13] embeds the point clouds into a higher-dimensional space using PointNet, while adopting the improved Lucas–Kanade algorithm [14] instead of the common T-net-style network in PointNet to align the point clouds. Dynamic graph convolutional neural network (DGCNN) [15] designs the EdgeConv module to extract local features. Deep closest point (DCP) [16] combines DGCNN with Transformer [17] to generate soft registration, and then approximates the ground truth by iterating. It has achieved significant improvement compared with previous registration algorithms.
However, these algorithms have various defects in feature extraction, calculation efficiency, or registration accuracy, most of which are generally sensitive to noise. Considering these factors, we propose a learning-based end-to-end point cloud registration method: DOPNet. DOPNet mainly consists of two parts: the feature extraction layer and the transformation prediction layer. The feature extraction layer is made up of three parts: (1) Using DGCNN to extract initial features and embed the point clouds into a higher-dimensional space; (2) cascaded offset-attention modules are utilized [18] for further feature extraction; (3) the feature interaction module is inserted to optimize the registration performance. In the transformation prediction layer, seven parameters indicating the spatial transformation are predicted by a multilayer perceptron (MLP). For some methods, experiments were conducted with two identical point clouds (e.g., DCP). We obtained the source and target point clouds by sampling the original point cloud twice independently to improve the precision.
The main contributions of this paper can be summarized as follows:
  • We propose an end-to-end registration network that can efficiently extract multiple features of point clouds using DGCNN and an attention mechanism. A feature interaction module is inserted into the feature extraction pipeline to achieve high-quality registration.
  • The source and target point clouds were generated by independent sampling, and additional experiments on asymmetric objects from Modelnet40 were conducted, which to some degree achieved a more rational setting for the registration experiments.
  • The proposed method was compared with five other registration methods. Adequate experiments verify that DOPNet is capable of extracting more salient hybrid information, as it achieves better registration accuracy and robustness than other methods.

2. Related Work

2.1. Traditional Methods

ICP [5] and its variants are representative methods of traditional registration algorithms. ICP constructs a rigid transformation matrix by searching the nearest point sets between two point clouds, and then iterating until the accuracy reaches the threshold. ICP is an effective optimal registration method without sophisticated procedures like segmentation or feature extraction. However, the iterative computation and the process of searching the nearest point sets still cause extensive computation expenditure, and it is prone to local minima and sensitive to initial positions, as ICP only considers point-to-point distances. Generalized ICP (GICP) [9] combines point-to-point, point-to-plane, and plane-to-plane strategies to improve registration accuracy and robustness. Normal iterative closest point (NICP) [8] takes into consideration the normal vectors and the local curvature, and thus it enhances the utilization of local structure information. Variants of ICP do make improvements to some extent, but there are still limitations in terms of calculation efficiency, accuracy, and robustness.

2.2. Learning-Based Methods

Learning-based methods tend to possess better robustness and generalization than traditional ones because of their parameter-learning ability. PointNet [10] is the first neural network which can directly process 3D point clouds; it utilizes an MLP to achieve feature mapping without additional transformation. Its improved version, PointNet++ 11, enhances the utilization of local features with extensive computation cost because of its iterative feature extraction. Since then, various methods for processing point cloud data have been proposed, such as DGCNN [15], PointCNN [19], FCGF [20], and so on.
In terms of point cloud registration, PointNetLK [13] leverages the Lucas–Kanade (LK) algorithm, which is generally applied in image registration to solve the point cloud registration challenge and successfully improves registration accuracy, but its iterative computation style renders the calculation efficiency inferior. CorsNet [21] extracts global features using PointNet, in which the high-dimensional features are concatenated with the 64-dimensional features extracted by PointNet to enhance the utilization of local features. However, such stitching is not rigorous due to PointNet’s inherent deficiency in extracting local features [11]. DeepGMR [22] constructs a registration network based on the Gaussian mixture model (GMM), and it formulates registration as the process of minimizing the KL divergence [23] between two probability distributions modeled as mixtures of Gaussians. The 3DRegNet [24] network utilizes pointwise correspondence to align two point clouds. A key point detector in DeepVCP [25] is designed to enhance registration accuracy via learning searches of qualified key point sets. FMR [26] constrains the global feature distance of the inputs with an extra decoder.
DCP [16] adopts DGCNN [15], which is more capable of extracting local features compared with PointNet, to map point clouds into the higher-dimensional space, and it approximates soft matching through transformer and pointer layers. DCP is an effective migration application of an attention mechanism from natural language processing (NLP) to point cloud registration, but it relies on special local features of point clouds to generate effective feature-point correspondences. Two identical point clouds were employed for experiments, which makes the evaluation results less rigorous.

2.3. Attention Mechanism

The attention mechanism is proposed in [17], which is widely applied in the NLP and computer vision (CV) fields [27]. In terms of point clouds, point cloud transformer (PCT) [18], point transformer (PT) [28], and DCP, etc. use an attention mechanism to extract features or to improve their feature-extraction ability. PREDATOR [29] applies cross-attention to construct an overlap-attention module and capture global pictures. Lepard [30] directly predicts the transformation matrix with cascaded self-attention and cross-attention modules. DeepUME [31] achieves accurate registration based on the combination of deep learning and a universal manifold embedding algorithm, while the transformer structure is leveraged to extract features.
PCT proposes offset attention (OA), which is an improvement based on self-attention. OA can be stacked in the point cloud processing pipeline to construct a multi-level feature extraction network, which can be used for different tasks such as semantic segmentation and classification of point clouds. This modified attention module is an excellent migration application of the transformer structure from NLP to point cloud data, and it is also more applicable to the data structure of point clouds, which inspired us to obtain abundant feature information with cascaded OA modules.

3. Problem Statement

In terms of the two given point clouds PS and PT with the same geometry and different spatial position, where P s = { s 1 , s 2 , , s n } R 3 and P t = { t 1 , t 2 , , t n } R 3 denote the source point cloud and the target point cloud, respectively, point cloud registration is aimed at finding a rigid spatial transformation ST that can align PS and PT by acting ST on PS. S T S E ( 3 ) is defined as:
S T = [ R T 0 1 ]
Here, R S O ( 3 ) and T R 3 denote the rotation matrix and the translation vector, respectively, and the rotation matrix is transformed from a quaternion q = w + x i + y j + z k by Equation (2):
R = [ 1 2 y 2 2 z 2 2 x y 2 w z 2 x z + 2 w y 2 x y + 2 w z 1 2 x 2 2 z 2 2 y z 2 w x 2 x z + 2 w y 2 y z + 2 w x 1 2 x 2 2 y 2 ]
The translation vector is defined as Equation (3), where tx, ty, and tz are three translation parameters.
T = [ t x t y t z ]
Thus, the ultimate goal of our algorithm can be formulated as solving seven parameters [ t x , t y , t z , w , x , y , z ] T , and then the spatial transformation can be derived.

4. Methodology

4.1. Network Architecture

Figure 1 shows the structure of DOPNet. The name DOPNet denotes the DGCNN, OA, and MLP, which are the main structural parts of the network. The model is mainly composed of the feature extraction layer and the transformation prediction layer. In summary, we utilize a Siamese network that consists of DGCNN, OA modules, and feature interaction modules as the feature extraction layer. An MLP in the form of PointNet plays the role of the transformation prediction layer, which finally outputs seven parameters to derive the predicted transformation.

4.2. Feature Extraction Layer

The first stage of the pipeline is embedding the original inputs PS and PT into a higher-dimensional space using DGCNN. For the i-th point pi of the point clouds, DGCNN constructs a local neighbor graph by searching its k nearest neighbors with the KNN (k-nearest neighbor) algorithm, and the graph can characterize the local geometric structure of the point clouds. Then the EdgeConv module is applied to extract corresponding edge features F on the edges connected between the nodes and their neighbors. The forward mechanism is formulated as:
F = h θ [ K θ ( p i , k ) , i N ]
where N denotes the number of points of the corresponding point cloud, K θ denotes the KNN algorithm, and h θ denotes EdgeConv. Compared with PointNet, DGCNN is able to extract more abundant structural information from the point sets by dynamically updating the graph structure between different layers, which enables DGCNN to outperform PointNet in capturing local features. This claim will be proven in the ablation experiments in Section 5.5.
DGCNN can extract a wealth of geometric information, but the semantic information in the feature space is insufficient. Thus, four OA modules were employed for further multi-layer feature extraction after DGCNN to excavate more semantic feature information, and Figure 2 shows the structure of the OA module. The l-th OA takes the feature Fl−1 of dimension (N,d) extracted by the previous network as input. First, the feature is re-weighted by the self-attention module without dimensional variation, then the subtraction is made between the weighted feature and the original feature. The output Fl is finally generated by adding the original feature to the output of the LBR (linear + BatchNorm + ReLU) layer. Note that the OA uses softmax + l1Norm instead of the original scale+softmax [19] to calculate attention scores in the self-attention module; such alteration can enhance the attention weights of meaningful features and weaken the effects of noise. The mechanism is shown in Equation (5).
F l = F l 1 + [ F l 1 S E l ( F l 1 ) ]
Here, the superscript letters denote the number of OAs. The OA is designed based on self-attention, it enhances the robustness against noise, and the utilization of the feature offsets enriches the semantic information contained in the output features, which also enhances the utilization of previous features. Its residual structure minimizes the adverse effects of training deeper neural networks, which makes it suitable for constructing a deep stacked feature extraction network. Thus, four cascaded OA modules are leveraged as the multilevel feature extraction layer to construct more comprehensive features.
In order to achieve higher registration accuracy, the global features ought to be comprehensive representations of the two point clouds’ structural and feature information. The combination of DGCNN and OA can extract satisfactory features in one branch, but the features of the two branches are uncorrelated, which means the final global features might be not applicable. Considering that case, we inserted the feature interaction module (FI) into the feature extraction pipeline to establish information connections. Such connections make the two feature extraction branches gain information from each other early on, before constructing global features to mine more desired features. Equation (6) shows the mechanism of FI for the source point cloud branch.
F s i = C o n v ( c a t { F s , R [ M a x ( F t ) , N s ] } )
Here, Fs and Ft represent the low-dimensional local features of PS and PT, respectively, extracted by the previous network, and Ns denotes the number of points of PS, while R denotes repeating the features Ns times in the spatial dimension. For another branch, this is still applicable by changing the number of points and swapping the position of Fs and Ft. FI is inserted into the pipeline after the second OA module to make the features interact early on. Furthermore, the dimension of features will be doubled by FI, which ought to make the subsequent OA modules extract more latent features in the higher-dimensional space.
Finally, the outputs of FI and OA are concatenated through skip connection to conduct multi-level feature extraction. The multi-level features contain multi-dimensional structural and semantic information, which comprehensively characterizes the point clouds. Then the features are processed by max pooling to obtain global features.

4.3. Transformation Prediction Layer

The total features are generated by stitching the global features of PS and PT, and then the total features are concatenated with the spatial coordinates of PT as the input of the transformation prediction layer, which aims at aligning the global features to the coordinates of each point of PT. Finally, the hybrid features contain multi-dimensional features and the Euclidean spatial information of PT, which further enhances the feature linkage between PS and PT. Such composite features enable the transformation prediction layer to predict more accurate transformation.
An MLP is employed as the transformation prediction layer. The parameter-learning-capability of the MLP makes it able to process the composite features more efficiently, and it can mine the complex latent information to directly predict the transformation matrix. The output of the transformation prediction layer is a 1 × 7 vector [ t x , t y , t z , w , x , y , z ] T . The first three parameters denote the translation vector T, and the last four parameters denote the quaternion q.

4.4. Loss Function

The goal is aligning the given point clouds by predicting accurate transformation, so the predicted transformation and the true transformation are chosen as the variables to construct the loss function, which is defined as Equation (7).
L o s s = Q p r e Q g t + λ t p r e t g t 2
where t and Q denote the translation vector and the rotation matrix, respectively, while the subscripts pre and gt denote the predicted results and the ground truth respectively. λ is the weight coefficient, which adjusts the model’s sensitivity to rotation and translation, it is empirically set to 1.5.

5. Experiments

5.1. Implementation Details

We used Modelnet40 [32], which covers 12,311 3D CAD models from 40 categories, and the models from the Stanford 3D Scanning Repository [33] as the experimental data sets. However, it was not reasonable for some symmetrical objects from Modelnet40 (bottle, bowl, cone, cup, flowerpot, lamp, tent, vase, etc.) to evaluate the registration accuracy by using the present metrics [34]. Thus, dual experiments were conducted on total objects (TO) and asymmetric objects (AO) by removing those symmetrical objects. We used cosine annealing to adjust the learning rate with an initial learning rate of 0.0001 and adopted the Adam optimizer [35] to optimize the network parameters for 350 epochs of training. All experiments were conducted with an AMD 5600X CPU and an NVIDIA GeForce RTX 3060 GPU.

5.2. Comparison and Evaluation Metrics

DOPNet was compared with ICP [5] and four learning-based methods, including DCP [16], PointNetLK [13], DeepGMR [22], and CorsNet [21]. DOPNet was temporarily designed without considering partial-to-partial registration. Our first motivation was choosing algorithms that had similar application scenarios. In addition, those methods which were similar in construction or method of manipulating features also could be ideal choices. Thus, we chose five methods as comparisons to validate the registration ability and methodology of DOPNet.
For all training and testing, unless otherwise specified, PS and PT were generated by randomly sampling 1024 points twice from the original point cloud; the rotation angles were randomly generated in [0°, 45°] and the initial translation distances were randomly selected in [−0.5, 0.5] as the true transformation. We measured the root mean square error (RMSE) between the predicted value and the ground truth for anisotropic error evaluation. The metrics proposed in RPMNet [36] were employed for isotropic error evaluation, which are given by Equations (8) and (9).
E r r o r ( R ) = arccos [ t r ( R g t T R p r e ) 1 2 ]
E r r o r ( t ) = t g t t p r e 2
where Rpre, tpre, Rgt, and tgt denote the predicted transformation and the true transformation, respectively, and tr(·) is the trace of the matrix.

5.3. Modelnet40

5.3.1. Unseen Shapes

In this experiment, ModelNet40 was randomly divided into the training set and the testing set. We used 9843 models as the training set and 2688 models as the testing set. Table 1 shows the results. Our method achieved excellent accuracy, while it was slightly inferior to DeepGMR in translation error. Such slight deficiency might be created by DeepGMR’s ability to capture special position relationships between point sets, while DOPNet learns more comprehensive features. That means that DeepGMR is more sensitive to special geometric structures, which might explain its minor superiority in unseen shapes and inferiority in unseen categories. Figure 3 shows the registration results. It can be seen that the proposed method can still achieve good registration results for highly symmetric structures, repetitive structures, and complex structures, which frequently cause falling into local optima or unsuccessful registration. Furthermore, the flowerpot (one of the special symmetric categories) was well aligned by DOPNet, but the corresponding metrics in AO were still much better than those in TO. Such variation demonstrates that registration errors for these special symmetrical objects cannot be fairly evaluated with the metrics.

5.3.2. Unseen Categories

To evaluate the generalization of these methods, we used the first 20 categories as the training set and the remaining 20 categories as the testing set; note that ICP was evaluated on the testing set. The results are shown in Table 2. The ICP error rate slightly decreased, while the errors of learning-based methods increased due to the differences in features of various categories of point clouds; CorsNet especially almost failed registration. Benefiting from the hybrid features which contain abundant spatial structure information and latent semantic information, DOPNet is less sensitive to the variation of feature styles, it shows the best generalization with the lowest errors.
Furthermore, in order to evaluate the sensitivity of these methods to the initial rotation angle, ICP and the trained models were tested on the total model data (TO) of the testing set with the range of [0°, 90°] by taking 10° as the step. For brevity, Error(R) and Error(t) were chosen as the bases of evaluation. Figure 4 shows the results. The red folded line denotes DOPNet. DOPNet achieved a similar error rate to other methods in the range of [0°, 50°], and it was slightly inferior in translation error when the rotation angle was bigger. However, on the whole it still showed stronger robustness toward rotation angles. As different types of model data were used in training and testing, the results also further reflect the good generalization of DOPNet. Table 3 and Table 4 show the corresponding quantitative results.

5.3.3. Robustness

Noise is an inevitable disturbance factor affecting registration accuracy. We tested robustness against noise for these methods with the data set and trained models in Section 5.3.1. The noise was sampled from N(0, 0.012) and clipped to [−0.5, 0.5]. Table 5 shows the results. As the noise blurred the key features and disturbed the correspondences between PS and PT, almost all the metrics of these compared methods became slightly or significantly worse. As similar datasets in Section 5.3.1 were employed to conduct this experiment, DeepGMR still achieved the best performance in translation error for the potential characteristics described in Section 5.3.1. The OA carried out consequential suppression of noise, and our method achieved relatively superior registration results in both TO and AO.

5.3.4. Calculation Efficiency

Calculation efficiency is an important factor in evaluating a registration algorithm. We tested the efficiency of these methods under the same settings as in Section 5.3.1, and the average cost on the TO was used as the time cost of the corresponding method. The results are shown in Table 6. Our method was slightly less efficient than DeepGMR, as DeepGMR has fewer learnable parameters, and it is superior to other registration methods in calculation efficiency. Our registration network was designed in a non-iterative style, and we also took into consideration the latent variation of features during registration in detail. Therefore, DOPNet can not only achieve excellent registration accuracy, it also possesses fairly good calculation efficiency.

5.4. Stanford Point Clouds

The Stanford 3D Scanning Repository [33] contains many point clouds obtained by scanning real objects. These models are more complex compared with those in Modelnet40. Thus, the Stanford bunny model and armadillo model were utilized to further evaluate the generalization and registration ability of DOPNet. The trained network model and evaluation metrics in Section 5.3.2 were employed to conduct this experiment. Note that 4096 points were sampled from the original models to retain more significant structural information, as the models are much more structurally complicated. The results are shown in Figure 5 and Table 7.
Error(R) increased slightly compared with the corresponding metric of AO in Section 5.3.2, while other metrics improved. Figure 5 shows that DOPNet can still achieve fine registration with this different data set without retraining. The qualitative and quantitative results further prove that DOPNet possesses good generalization and registration.

5.5. Ablation Studies

We conducted several ablation experiments on Modelnet40 to dissect DOPNet and understand the value of these modules. All experiments were conducted with the same settings as in Section 5.3.1. The experiments consisted of three parts: replacing DGCNN with PointNet, removing the OA module, and removing the FI module. Error(R) and Error(t) were used as the metrics. The results are shown in Table 8.
The last row denotes DOPNet. It can be seen that the metrics would analogously be slightly worse due to replacing DGCNN with PointNet and removing the FI module. Such variation demonstrates that DGCNN is more capable of extracting representative features than PointNet, and it also verifies the necessity of making features interact between two feature extraction branches for high-quality registration. Not surprisingly, removing OA modules terribly weakens the performance of DOPNet, which establishes that OA modules can effectively extract comprehensive semantic feature information for subsequent procedures. The results support our motivation to construct the network with OA modules.

6. Conclusions

We introduced a point cloud registration method based on deep learning, taking into account the calculation efficiency and accuracy of the point cloud registration, and achieving fast and accurate registration. DGCNN was leveraged for embedding point clouds into high-dimensional space, the cascaded OA modules were used for extracting comprehensive features, and the feature interaction module was inserted into the feature extraction pipeline to enhance information connection and registration accuracy. The predicted transformation is given by the MLP. On ModelNet40 and Stanford point cloud datasets, we carried out comprehensive comparison experiments with ICP and some mainstream learning-based methods, while the point clouds were generated by sampling the original models twice independently. The results adequately demonstrate that DOPNet possesses excellent registration ability and efficiency compared with these algorithms. In the future, we will work on solving the network limitations regarding partial-to-partial registration and improving accuracy and applicability based on the idea of predicting overlap masks.

Author Contributions

Conceptualization, R.Y. and J.L.; methodology, R.Y., J.L. and L.L.; software, R.Y.; validation, J.L., L.L., Y.Z., X.G. and J.G.; formal analysis, R.Y.; investigation, R.Y.; resources, L.L., Y.Z. and X.G.; data curation, R.Y.; writing—original draft preparation, R.Y.; writing—review and editing, R.Y., J.L., Y.Z., X.G. and J.G.; visualization, R.Y.; supervision, J.L. and L.L.; project administration, R.Y. and J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61960206010).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in [32,33].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, K.; Ma, L.; Meng, Y.; Liu, L.; Wang, J.; Junior, J.M.; Gonçalves, W.N.; Li, J. 3D Vehicle Detection Using Multi-Level Fusion from Point Clouds and Images. IEEE Trans. Intell. Transp. Syst. 2022, 23, 15146–15154. [Google Scholar] [CrossRef]
  2. Baek, J.; Park, J.; Cho, S.; Lee, C. 3D Global Localization in the Underground Mine Environment Using Mobile LiDAR Mapping and Point Cloud Registration. Sensors 2022, 22, 2873. [Google Scholar] [CrossRef] [PubMed]
  3. Mu, Z.; Cai, Z.; Zeng, C.; Li, Z.; Liang, X.; Yang, F.; Chen, T.; Dong, S.; Deng, C.; Niu, S. A point cloud registration-based calibration algorithm for robot offline programming automatic loading in aero-grinding applications. Ind. Robot Int. J. Robot. Res. Appl. 2022, 49, 1218–1228. [Google Scholar] [CrossRef]
  4. Li, H.; Hartley, R. The 3D-3D registration problem revisited. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Rio de Janeiro, Brazil, 14–20 October 2007; pp. 1–8. [Google Scholar]
  5. Besl, P.; McKay, N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar] [CrossRef] [Green Version]
  6. Zhang, X.; Li, Z.; Wang, X.; Li, P.; Wang, Y. Research of 3D point cloud data registration algorithms based on feature points and improved ICP. Transducer Microsyst. Technol. 2012, 31, 116–122. [Google Scholar]
  7. Yang, J.; Li, H.; Campbell, D.; Jia, Y. Go-ICP: A globally optimal solution to 3D ICP point-set registration. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 2241–2254. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Serafin, J.; Grisetti, G. NICP: Dense normal based point cloud registration. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–3 October 2015; pp. 742–749. [Google Scholar]
  9. Segal, A.; Haehnel, D.; Thrun, S. Generalized-icp. In Proceedings of the Robotics: Science and Systems, Seattle, WA, USA, 28 June–1 July 2009; Volume 2, p. 435. [Google Scholar]
  10. Qi, C.R.; Su, H.; Mo, K.; Guibas, L. PointNet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  11. Qi, C.R.; Yi, L.; Su, H.; Guibas, L. PointNet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  12. Sarode, V.; Li, X.; Goforth, H.; Aoki, Y.; Srivatsan, R.; Lucey, S.; Choset, H. PCRNet: Point cloud registration network using PointNet encoding. arXiv 2019, arXiv:1908.07906. [Google Scholar]
  13. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. PointNetLK: Robust & efficient point cloud registration using PointNet. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7156–7165. [Google Scholar]
  14. Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Vancouver, BC, Canada, 24–28 August 1981. [Google Scholar]
  15. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
  16. Wang, Y.; Solomon, J. Deep closest point: Learning representations for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 3522–3531. [Google Scholar]
  17. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  18. Guo, M.-H.; Cai, J.-X.; Liu, Z.-N.; Mu, T.-J.; Martin, R.R.; Hu, S.-M. PCT: Point cloud transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
  19. Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution on x-transformed points. Adv. Neural Inf. Process. Syst. 2018, 31, 820–830. [Google Scholar]
  20. Choy, C.; Park, J.; Koltun, V. Fully Convolutional Geometric Features. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 8957–8965. [Google Scholar]
  21. Kurobe, A.; Sekikawa, Y.; Ishikawa, K.; Saito, H. Corsnet: 3d point cloud registration by deep neural network. IEEE Robot. Autom. Lett. 2020, 5, 3960–3966. [Google Scholar] [CrossRef]
  22. Yuan, W.; Eckart, B.; Kim, K.; Jampani, V.; Fox, D.; Kautz, J. DeepGMR: Learning Latent Gaussian Mixture Models for Registration. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 733–750. [Google Scholar]
  23. Eckart, B.; Kim, K.; Kautz, J. HGMR: Hierarchical gaussian mixtures for adaptive 3d registration. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 705–721. [Google Scholar]
  24. Pais, G.D.; Ramalingam, S.; Govindu, V.M.; Nascimento, J.C.; Chellappa, R.; Miraldo, P. 3DRegNet: A Deep Neural Network for 3D Point Registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 7191–7201. [Google Scholar]
  25. Lu, W.; Wan, G.; Zhou, Y.; Fu, X.; Yuan, P.; Song, S. Deepvcp: An end-to-end deep neural network for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 12–21. [Google Scholar]
  26. Huang, X.; Mei, G.; Zhang, J. Feature-metric registration: A fast semi-supervised approach for robust point cloud registration without correspondences. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11366–11374. [Google Scholar]
  27. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  28. Zhao, H.; Jiang, L.; Jia, J.; Torr, P.H.; Koltun, V. Point transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 16259–16268. [Google Scholar]
  29. Huang, S.; Gojcic, Z.; Usvyatsov, M.; Wieser, A.; Schindler, K. Predator: Registration of 3d point clouds with low overlap. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4267–4276. [Google Scholar]
  30. Li, Y.; Harada, T. Lepard: Learning partial point cloud matching in rigid and deformable scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–23 June 2022; pp. 5554–5564. [Google Scholar]
  31. Lang, N.; Francos, J.M. DeepUME: Learning the Universal Manifold Embedding for Robust Point Cloud Registration. arXiv 2021, arXiv:2112.09938. [Google Scholar]
  32. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  33. Curless, B.; Levoy, M. A volumetric method for building complex models from range images. In Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA, 4–9 August 1996; pp. 303–312. [Google Scholar]
  34. Xu, H.; Liu, S.; Wang, G.; Liu, G.; Zeng, B. OMNet: Learning overlapping mask for partial-to-partial point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 3132–3141. [Google Scholar]
  35. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference for Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 45–53. [Google Scholar]
  36. Yan, Z.; Hu, R.; Yan, X.; Chen, L.; Kaick, O.V.; Zhang, H.; Huang, H. RPM-Net: Recurrent prediction of motion and parts from point cloud. ACM Trans. Graph. 2019, 38, 240. [Google Scholar] [CrossRef]
Figure 1. Network architecture of DOPNet, including DGCNN, cascading OA, feature interaction module, and MLP.
Figure 1. Network architecture of DOPNet, including DGCNN, cascading OA, feature interaction module, and MLP.
Sensors 22 08217 g001
Figure 2. Network architecture of offset-attention, including features, self-attention, and LBR.
Figure 2. Network architecture of offset-attention, including features, self-attention, and LBR.
Sensors 22 08217 g002
Figure 3. Registration results (red: target point cloud; green: source point cloud; blue: predicted point cloud). The first column shows the initial position of the point clouds, while other columns show the registration results.
Figure 3. Registration results (red: target point cloud; green: source point cloud; blue: predicted point cloud). The first column shows the initial position of the point clouds, while other columns show the registration results.
Sensors 22 08217 g003
Figure 4. Generalization to the initial rotation angle. (a) denotes Error(R), (b) denotes Error(t).
Figure 4. Generalization to the initial rotation angle. (a) denotes Error(R), (b) denotes Error(t).
Sensors 22 08217 g004
Figure 5. Visual registration results. Red, green, and blue denote the target point cloud, source point cloud, and predicted point cloud, respectively, and (a1,b1) indicate the initial positions, while (a2,b2) denote corresponding registration results.
Figure 5. Visual registration results. Red, green, and blue denote the target point cloud, source point cloud, and predicted point cloud, respectively, and (a1,b1) indicate the initial positions, while (a2,b2) denote corresponding registration results.
Sensors 22 08217 g005
Table 1. Test on unseen shapes.
Table 1. Test on unseen shapes.
Total Objects (TO)Asymmetrical Objects (AO)
Error(R)RMSE(R)Error(t)RMSE(t)Error(R)RMSE(R)Error(t)RMSE(t)
ICP9.268011.63440.02490.02837.592410.62520.02130.0257
DCP8.87766.58960.03860.02538.20906.51520.03830.0252
PointNetLK4.10946.81720.03250.04422.66266.48580.02400.0408
DeepGMR6.04365.21490.01800.01194.68203.81500.01800.0118
CorsNet17.683411.26900.11940.080517.557711.18100.11850.0793
DOPNet3.26523.09540.02020.01302.37903.09540.01890.0130
Table 2. Test on unseen categories.
Table 2. Test on unseen categories.
TOAO
Error(R)RMSE(R)Error(t)RMSE(t)Error(R)RMSE(R)Error(t)RMSE(t)
ICP8.930210.58380.02360.02647.765610.36870.02000.0236
DCP9.76606.70740.03970.02598.79166.49180.03820.0245
PointNetLK4.63918.15760.03570.05233.03856.38260.02650.0435
DeepGMR6.76315.89110.01710.01115.64364.91850.01690.0109
CorsNet18.267311.52480.12340.082518.385111.65430.12590.0843
DOPNet3.45263.04810.01480.01092.91522.54190.01410.0101
Table 3. Results of Error(R).
Table 3. Results of Error(R).
AngleICPDCPPointNetLKDeepGMRCorsNetDOPNet
00.90847.58241.72481.461433.44830.0180
101.25875.59091.94392.726326.24061.9824
202.04305.63392.25753.979220.10142.1643
303.35966.78532.74295.018816.42992.8950
406.15028.59043.71876.089616.78913.0301
5013.392311.18126.34217.716820.37853.8056
6026.158615.579012.742911.578226.03977.2468
7039.491323.460125.128320.818133.274514.8409
8057.313635.336542.607436.549241.889327.1315
9073.358349.989163.154653.963351.639742.0658
Table 4. Results of Error(t).
Table 4. Results of Error(t).
AngleICPDCPPointNetLKDeepGMRCorsNetDOPNet
00.00740.03910.01780.01610.21300.0124
100.00790.03820.01930.01620.17020.0152
200.00950.03780.02130.01640.13370.0140
300.01210.03820.02460.01660.11100.0139
400.01760.03890.03080.01690.11350.0143
500.03200.03980.04480.01730.13700.0156
600.05670.04050.07940.01770.17310.0222
700.07890.04120.14470.01830.21790.0379
800.10630.04230.22800.01900.26780.0617
900.13000.04370.32370.01960.32260.0881
Table 5. Robustness against noise.
Table 5. Robustness against noise.
TOAO
Error(R)RMSE(R)Error(t)RMSE(t)Error(R)RMSE(R)Error(t)RMSE(t)
ICP9.626711.31170.02630.02838.130911.07790.02260.0253
DCP9.07856.70740.03820.02518.41296.63570.03010.0251
PointNetLK4.69187.30420.03750.04713.05786.76350.02790.0399
DeepGMR6.07835.23920.01800.01194.73003.85950.01800.0118
CorsNet17.664511.22460.11990.080717.733811.25750.11960.0801
DOPNet3.32553.11030.02020.01312.44761.92770.01890.0117
Table 6. Calculation efficiency.
Table 6. Calculation efficiency.
MethodICPDCPPointNetLKDeepGMRCorsNetDOPNet
Time (ms)36.2114212158018
Table 7. Quantitative registration results.
Table 7. Quantitative registration results.
ModelsError(R)RMSE(R)Error(t)RMSE(t)
Armadillo3.33201.83800.00360.0021
Bunny3.31351.82160.00310.0018
Table 8. Ablation studies.
Table 8. Ablation studies.
DGCNNOAFITOAO
Error(R)Error(t)Error(R)Error(t)
4.28690.05643.48030.0477
5.77530.07655.14830.0729
4.27480.05233.54020.0480
3.26520.02022.37900.0189
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yi, R.; Li, J.; Luo, L.; Zhang, Y.; Gao, X.; Guo, J. DOPNet: Achieving Accurate and Efficient Point Cloud Registration Based on Deep Learning and Multi-Level Features. Sensors 2022, 22, 8217. https://doi.org/10.3390/s22218217

AMA Style

Yi R, Li J, Luo L, Zhang Y, Gao X, Guo J. DOPNet: Achieving Accurate and Efficient Point Cloud Registration Based on Deep Learning and Multi-Level Features. Sensors. 2022; 22(21):8217. https://doi.org/10.3390/s22218217

Chicago/Turabian Style

Yi, Rongbin, Jinlong Li, Lin Luo, Yu Zhang, Xiaorong Gao, and Jianqiang Guo. 2022. "DOPNet: Achieving Accurate and Efficient Point Cloud Registration Based on Deep Learning and Multi-Level Features" Sensors 22, no. 21: 8217. https://doi.org/10.3390/s22218217

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop