Next Article in Journal
Automatic Identification of Earth Rock Embankment Piping Hazards in Small and Medium Rivers Based on UAV Thermal Infrared and Visible Images
Next Article in Special Issue
IPCONV: Convolution with Multiple Different Kernels for Point Cloud Semantic Segmentation
Previous Article in Journal
Integrating Topographic Skeleton into Deep Learning for Terrain Reconstruction from GDEM and Google Earth Image
Previous Article in Special Issue
An Object-Based Ground Filtering of Airborne LiDAR Data for Large-Area DTM Generation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Semi-Supervised Point Cloud Registration via Latent GMM-Based Correspondence

1
School of Mechanical Engineering and Automation, Harbin Institute of Technology, Shenzhen 518000, China
2
Faculty of Applied Sciences, Macao Polytechnic University, Macao 999078, China
3
Department of Medical Physics and Biomedical Engineering, University College London, London WC1E 6BT, UK
4
Department of Electronic Engineering, The Chinese University of Hong Kong, Hong Kong 999077, China
5
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing (LIESMARS), Wuhan University, Wuhan 430079, China
6
Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(18), 4493; https://doi.org/10.3390/rs15184493
Submission received: 24 July 2023 / Revised: 24 August 2023 / Accepted: 25 August 2023 / Published: 12 September 2023

Abstract

:
Due to the fact that point clouds are always corrupted by significant noise and large transformations, aligning two point clouds by deep neural networks is still challenging. This paper presents a semi-supervised point cloud registration (PCR) method for accurately estimating point correspondences and handling large transformations using limited prior datasets. Firstly, a modified autoencoder is introduced as the feature extraction module to extract the distinctive and robust features for the downstream registration task. Unlike optimization-based pairwise PCR strategies, the proposed method treats two point clouds as two implementations of a Gaussian mixture model (GMM), which we call latent GMM. Based on the above assumption, two point clouds can be regarded as two probability distributions. Hence, the PCR of two point clouds can be approached by minimizing the KL divergence between these two probability distributions. Then, the correspondence between the point clouds and the latent GMM components is estimated using the augmented regression network. Finally, the parameters of GMM can be updated by the correspondence and the transformation matrix can be computed by employing the weighted singular value decomposition (SVD) method. Extensive experiments conducted on both synthetic and real-world data validate the superior performance of the proposed method compared to state-of-the-art registration methods. These experiments also highlight the method’s superiority in terms of accuracy, robustness, and generalization.

1. Introduction

Point cloud registration (PCR) is a critical task in computer vision [1,2], robotic perception [3,4], medical image analysis [5], computer-integrated intervention [6,7], and aircraft measurement [8]. It is a problem to estimate the rigid transformation matrix to align two point clouds. Solving this problem involves two steps: obtaining the optimal mapping from a point in the source point cloud to its corresponding point in the target point cloud, and then computing the optimal transformation matrix based on the point correspondence. Various methods utilize the optimization or deep neural network to find robust and accurate point correspondence. In this paper, we focus on using learned embeddings to find the optimal point-to-GMM correspondence matrix; we also use the correspondence matrix and latent GMM components to estimate the rigid transformation.
In our previous PCR methods [9,10,11,12], the registration problem was cast into a maximum likelihood estimation problem via the hybrid mixture model (HMM), where positional and orientation features are incorporated to estimate the correspondence by Bayes’ theorem. Like other optimization-based PCR methods, our previous methods were highly sensitive to initialization and failed to handle large transformations well. In real-world applications, there are often large transformations and noise between point clouds, so it is a challenge to make the point cloud registration algorithm converge even in the case of noise and large transformations [13,14]. In recent years, several deep learning architectures utilizing point clouds have demonstrated remarkable performance in various 3D vision tasks, including classification, segmentation, and detection. Various deep learning-based architectures for PCR have also been widely studied. The focus of most existing point cloud registration methods mainly revolves around descriptor extraction [15,16] or supervised learning [17,18,19,20], which heavily rely on large annotated datasets for training. Consequently, it is essential to introduce semi-supervised or unsupervised learning approaches to reduce the dependency of point cloud registration networks on datasets.
This paper presents a semi-supervised method for solving the PCR problem by minimizing the Kullback-Leibler divergence (KL-divergence) between two probability distributions generated by a shared latent GMM. The proposed architecture consists of three main modules: the feature extraction module, the latent GMM module, and the transformation estimation module. The feature extraction module utilizes a modified autoencoder module to learn how to extract distinctive and robust pose-attentive features from point clouds. The latent GMM module estimates the correspondences between point clouds and latent GMM components and subsequently updates the parameters of the latent GMM components. The transformation estimation module employs the weighted singular value decomposition (SVD) to compute the rigid transformation between two point clouds. The overview of the proposed method is shown in Figure 1. The main contributions can be summarized as follows.
(1)
A new semi-supervised method for point cloud registration is proposed. The point cloud-based autoencoder is modified to serve as the feature extraction module, which is conducive to extracting distinctive and pose-attentive features from the limited data.
(2)
An augmented regression network is proposed for generating the correspondence between the point cloud and latent GMM components.
The rest of this paper is organized as follows: Section 1 summarizes the related point cloud registration methods and reviews the traditional GMM-based PCR method. Section 2 establishes the formulation of the probabilistic point cloud registration task and illustrates the architecture of the proposed method in detail. Section 3 presents and discusses the experimental results. Section 4 analyzes the advantages of the proposed method compared to other methods based on the experimental results. Section 5 concludes this paper.

1.1. Related Work

1.1.1. ICP Method and Its Variants

The iterative closest point (ICP) method [21] and its variants [22,23,24] are the most popular optimization-based registration methods in the field. These methods conduct one-to-one correspondence estimation and transformation updating iteratively. Because ICP adopts a hard alignment strategy, this method depends on a good initial transformation and easily falls into local optima. Moreover, these methods are vulnerable to outliers and noise. Numerous variants of the ICP method have been proposed to address the aforementioned limitations. Agamennoni et al. [25] proposed an improvement, where each point in one point cloud is associated with a set of points in the other point cloud, and these associations are weighted to form a probability distribution. To accelerate the convergence of ICP, Low [26] introduced the linear approximation method to solve the point-to-plane error metric. Chetverikov et al. [27] proposed the trimmed ICP (TrICP) to improve the robustness, where the least trimmed squares (LTS) approach was consistently used in all phases of the operation. Yang et al. [28] introduced the branch-and-bound theory, which searches the entire SE(3) space to obtain the globally optimal solution to PCR.

1.1.2. Probabilistic Registration Methods

The probabilistic registration method is another popular kind of optimization-based method to solve PCR. In this category, the one-to-many correspondence confidences are represented by probabilities, and the PCR is considered the maximum likelihood estimation problem. GMM and HMM are commonly employed in estimating the correspondence between point clouds. Subsequently, the expectation maximization (EM) algorithm is utilized to iteratively update the parameters of the mixture model and estimate the rigid transformation. The coherent point drift (CPD) method, proposed by Myronenko et al. [29], treats one point cloud as a probability distribution generated from the other, representing a GMM. The expectation conditional maximization for point registration (ECMPR) method, proposed by Horaud et al. [30], replaces the traditional M-step with a sequence of conditional maximization steps and introduces a general covariance matrix for PCR with anisotropic covariance. Evangelidis et al. [31] proposed the joint registration multiple PC (JRMPC) method, considering multiple point clouds as realizations of a latent GMM and transforming PCR into a clustering problem. Compared with GMM-based methods, HMM-based methods [9,10,11,12] used von Mises–Fisher (VMF) to model the orientational data. Ravikumar et al. [32] combined VMF and Student’s t distributions to model the orientation and position of the point cloud. Min et al. [9] proposed a rigid HMM-based registration method for computer-assisted orthopedic surgery, where one point cloud was considered as an HMM composed of GMM and VMF distributions. Then, this framework was extended to rigid and nonrigid PCR with anisotropic noise [10,11]. Considering that the estimation of normal vectors may introduce errors to the preoperative and intraoperative stages, we recently proposed the reliable HMM [12], where curvatures were introduced as indexes to evaluate the reliability of normal vectors. Due to the soft alignment strategy, the probabilistic method achieved better robustness to noise and outliers. However, the probabilistic method cannot handle a PCR task with large transformations.

1.1.3. End-to-End Learning-Based Registration Methods

The key idea of the end-to-end learning-based registration method is to input two point clouds into a neural network and estimate the transformation matrix by optimization or regression. PointNetLK [33] used PointNet [34] to extract global features from input point clouds and estimated the rigid transformation matrix by minimizing the differences in features. FMR [35] further enhanced PointNetLK by incorporating an autoencoder framework and a Chamfer distance loss, thereby reducing the reliance on labels. DCP [36] employed DGCNN [37] to extract features from input point clouds and utilized Transformer [38] to estimate the correspondence matrix between two point clouds. Subsequently, the rigid transformation was computed using SVD. However, both FMR and DCP attempt to extract feature differences or compute correspondences on a point-to-point basis. However, it is infeasible to apply these methods when the point cloud is sparse and contains noise and outliers, as such elements negatively affect the point-to-point relationship. Recently, DeepGMR [39] first proposed an end-to-end learning-based probabilistic registration that integrated neural networks into GMM-based registration. However, the neural network only extracted the global features by PointNet, and the combined features only contained the features of each isolated point; the topological features of point clouds were ignored. Hence, the efficiency and generalization of the feature extraction module will decrease when the prior labeled dataset is insufficient.

1.2. Recap of Traditional GMM-Based Point Cloud Registration

Given the source point cloud X R 3 × N and the target point cloud Y R 3 × M , ideally, the source point cloud X can be obtained by applying rotation transformation R and translation t to the target point cloud Y . The objective of rigid PCR is to estimate the rigid transformation R , t to align Y to X , as shown below:
R ^ , t ^ = arg min R , t X RY + t 2
where R ^ S O 3 and t ^ R 3 denote the optimal rotation matrix and translation vector.
We first review a traditional GMM-based method, like CPD, to solve the PCR problem, where the iterative EM algorithm is adopted. In CPD, the points in Y are assumed as GMM centroids, while the points in X are considered as points generated by the GMM. Then, given the mth target point y m , the probability density function (PDF) of the nth source point x n can be computed as follows:
p x n z n = m ; θ = 1 2 π σ 2 3 2 e x n y m 2 σ 2
where 1 n N , 1 m M and z n = m denote correspondence between the nth source point and the mth target point. Therefore, given the model parameters, the PDF of the nth point can be described as follows:
p x n θ = m = 1 M + 1 P ( z n = m ) p ( x n z n = m ; θ )
where m = M + 1 represents the outliers. Moreover, the additional uniform distribution p ( x n z n = M + 1 ) = 1 1 N N is introduced to account for outliers. P z n = m = 1 1 N N are assumed for residual target points m = 1 , , M . The GMM centroids are represented by a set of parameters θ = R , t , σ 2 and estimated by minimizing the negative log-likelihood function as follows:
Q θ = n = 1 N log m = 1 M + 1 P z n = m p x n z n = m
Then the above optimization problem can be solved by the EM framework. In the E-step, given initialized or previous parameters, the correspondence between the nth source point x n and the mth target point y m can be estimated by Bayes’ theorem as follows:
p z n = m x n ; θ o l d = P ( z n = m ) p ( x n z n = m ; θ o l d ) p ( x n θ o l d )
Then the minimization of the negative log-likelihood function (4) can be rewritten as follows:
Q θ = 1 2 σ 2 n = 1 N m = 1 M p m n x n R y m + t 2 + 3 N p 2 log σ 2
In the M-step, the new posteriors and previous model parameters are used to update the transformation matrix R , t and new model parameters by minimizing the objective function in (6). Then the above-mentioned E-step and M-step will iterate alternatively until convergence.

2. Methods

2.1. Problem Statement

In this paper, two point clouds are assumed as two probability distributions generated by the same latent GMM. Therefore, the PCR problem can be solved by minimizing the KL divergence between two probability distributions. The latent source and target GMM distributions are denoted as Θ = π , μ , σ 2 and Θ ¯ = π ¯ , μ ¯ , σ ¯ 2 respectively, where π , μ , σ 2 denote the mixing coefficients, and the means and covariances of GMM components. Then, according to the definition of KL-divergence [40], (1) can be transformed as follows:
R ^ , t ^ = arg min R , t K L ( ( R Θ ¯ + t ) Θ ) = arg max R , t Y p ( y ( R Θ ¯ + t ) ) ln p ( y Θ ) d y
For the target point cloud, the PDF of the j-th Gaussian distribution can be written as follows:
p ( y m Θ ¯ j ) = α j 2 π σ ¯ j 2 e 1 2 σ ¯ j 2 y m μ ¯ j 2
where 1 j J , and j = 1 J α j = 1 denotes the weight of each Gaussian distribution. Then, we minimize the KL divergence between R Θ ¯ + t , and Θ is converted to maximize the expected log-likelihood of data that follow the distribution of R Θ ¯ + t under Θ . Under the assumption that the transformed target point cloud RY + t = R y m + t m = 1 M is independently and identically sampled by the probability distribution p y R Θ ¯ + t , (7) can be expanded as follows by using the law of large numbers:
R ^ , t ^ = arg max R , t 1 M m = 1 M ln p R y m + t Θ = arg max R , t m = 1 M ln j = 1 J p R y m + t Θ j
In this paper, we first introduce the semi-supervised method into the optimization of (9). The detailed components and the training framework are presented in the next section.

2.2. Feature Extraction Module

The aim of the feature extraction module is to generate distinctive and efficient features of both the source and target point clouds. The modified autoencoder module is used as the feature extraction module, which is inspired by the dynamic graph CNN (DGCNN) [37] but with a modified encoder and three fully connected layers as the decoder. The proposed method can extract local and topological features to represent both two point clouds; it is different from the feature extraction module (PointNet) used in DeepGMR [39]. Compared to PointNet++ [41], DGCNN constructs a dynamic graph for each point cloud and uses the EdgeConv operation to compute features, allowing the network to adapt better to the geometric structures and capture more effective feature representations. Therefore, we chose the modified DGCNN as the encoder of the proposed feature extraction module. Following [42,43], the principle of designing the feature extraction module is to minimize the learning space and generate features containing the pose difference. Hence, we normalize the two point clouds by subtracting the means of X and Y before feeding them into the network. The encoder generates an augmented latent representation, and the desired output of the decoder is a reconstructed point cloud, reflecting the rigid transformation difference between the two point clouds.
Figure 2 shows the architecture of the proposed feature extraction module inspired by DGCNN. The shared feature extraction module is used for two point clouds. Hence, we denote N u m = m a x M , N to fix the point number of the input and output of the autoencoder. When N M , we can upsample the point cloud with less point numbers to meet the above assumption. For the target point cloud, Y = y m R p m = 1 , , M represents the 3D space or the arbitrary feature space. For each y m point, a k-nearest neighbor graph is constructed. In this graph, the k-nearest points y m i i = 1 , , k are used to calculate the edge features y m , y m y m i .
The EdgeConv operation contains an edge function and a vertex-wise aggregation. The shared multi-layer perceptron (MLP) is used as the edge function, and the max-pooling is used as the vertex-wise aggregation. We concatenate edge features of three layers and use an MLP to generate the common feature. Then the max-pooling and average-pooling are both applied to the common feature to augment the global feature.
The encoder outputs two global representations with a dimension of 1024. Then we concatenate two global representations as augmented representations for the source or target point cloud. The decoder contains three fully connected layers with dimensions of 2048, 1024, and Num × 3, with the activation being LeakyReLU. The feature extraction module can be trained in an unsupervised manner. This training framework can help the encoder generate the pose-attentive features from the source or target point clouds, which can improve the success of training an efficient feature extractor for PCR when prior datasets are limited.

2.3. Latent GMM Module

The principle of designing the latent GMM module is to replace the original E-step in the EM algorithm by utilizing the features of the source and target point cloud. The latent GMM module contains the augmented regression network and the parameter estimation module. The augmented regression network contains the four-layer MLP with output dimensions of 1024, 256, 128, and J to generate the point-to-GMM correspondence, where J denotes the number of Gaussian distributions. The augmented global features with a dimension of 2048 are concatenated by max-pooling and average-pooling. The local features with the dimension of N u m × 1024 are generated by the MLP in the modified encoder. Then, we can estimate the correspondence γ ¯ R N u m × J between the target point cloud and target latent GMM, where j = 1 J γ i j = 1 and 1 i N u m . Similarly, γ R N u m × J denotes the correspondence between the source point cloud and source latent GMM. According to [44], the effective number of points corresponds to the j-th Gaussian distribution and the mixing coefficient for the j-th component can be computed as follows:
γ j = i = 1 N u m γ i j , π j = γ j N u m γ ¯ j = i = 1 N u m γ ¯ i j , π ¯ j = γ ¯ j N u m
Then the derivatives of the objective function in (9) with respect to the means μ j of the latent Gaussian components are set to zero:
m = 1 M p R y m + t Θ j j = 1 J p R y m + t Θ j · R y m + t μ j σ j 2 = 0
We replace the posterior probabilities with the correspondence matrix γ estimated by the neural network. Ideally, the transformed target point cloud R Y + t is equivalent to the source point cloud X . Then the means μ j of the latent source GMM can be computed as follows:
μ j = 1 γ ¯ j i = 1 N u m γ ¯ i j R y i + t 1 γ ¯ j i = 1 N u m γ ¯ i j x i
Then, by setting the derivatives of the objective function in (9) with respect to the covariances σ j 2 of the latent source GMM to zero, we can obtain:
σ j 2 = 1 γ ¯ j i = 1 N u m γ ¯ i j R y i + t μ j R y i + t μ j T 1 γ ¯ j i = 1 N u m γ ¯ i j x i μ j x i μ j T
Similarly, the means μ ¯ j and the covariances σ ¯ j 2 of the latent target GMM can be computed as follows:
μ ¯ j 1 γ j i = 1 N u m γ i j y i σ ¯ j 2 1 γ j i = 1 N u m γ i j y i μ ¯ j y i μ ¯ j T
After the above derivation, the correspondence between point clouds and latent GMMs, as well as the parameters of the latent GMMs, can be directly estimated through the proposed neural network without the need for multiple optimization iterations.

2.4. Weighted SVD Module

Given the correspondence matrix and GMM components, we can compute the lower bound over (9) as follows:
R ^ , t ^ = arg max R , t i = 1 N u m j = 1 J γ ¯ i j ln p R y i + t Θ j = arg max R , t i = 1 N u m j = 1 J γ ¯ i j ln α j 2 π σ j 2 e 1 2 σ j 2 R y i + t μ j 2 = arg min R , t i = 1 N u m j = 1 J γ ¯ i j σ j 2 R y i + t μ j 2
We assume the point cloud transformation to be a form of the GMM component transformation, where each GMM component varies with respect to the transformation. This simplification saves the complexity of the optimization and neural networks. Then the objective function in (15) can be written as follows:
R ^ , t ^ = arg min R , t j = 1 J γ ¯ j σ j 2 R μ ¯ j + t μ j 2
The weighted SVD method is used to compute the transformation matrix in closed-form. The centers of two latent GMM components are computed as follows:
μ c = j = 1 J π j μ j , μ ¯ c = j = 1 J π ¯ j μ ¯ j
Then the cross-covariance matrix is defined as follows:
H = j = 1 J π j σ ¯ j 2 μ j μ c μ ¯ j μ ¯ c T
The rigid transformation can be computed according to SVD:
R ^ = U d i a g 1 , 1 , det U V T V T t ^ = μ c R ^ μ ¯ c
where U and V can be computed by the SVD of H = U S V T .

2.5. Loss Definition

The primary goal of the training process is to ensure that the modified autoencoder module and the augmented correspondence regression effectively extract distinctive features and establish accurate correspondences. Our proposed method employs two loss functions, specifically Chamfer loss and transformation loss, as part of a semi-supervised approach.

2.5.1. Chamfer Loss

The objective of the Chamfer loss function is to train the modified autoencoder module in an unsupervised manner. The Chamfer distance is utilized to quantify the dissimilarity between input point clouds X , Y and reconstructed point clouds X r e c , Y r e c . Specifically, the Chamfer loss between X and X r e c can be written as follows:
χ C D s r c = 1 N p X min q X r e c p q 2 2 + 1 N q X r e c min p X p q 2 2
where p and q indicate points in X and X r e c .

2.5.2. Transformation Loss

The objective of PCR is to estimate the transformation matrix that aligns the target point cloud X with the source point cloud Y . Therefore, the transformation loss is employed to minimize the discrepancy between the ground-truth transformation and the estimated transformation:
χ R = R ^ T R g I 2 2 χ t = t ^ t g 2 2
where the ground-truth transformation matrix is composed of R g and t g .

2.5.3. Total Loss

The total loss function used for training the autoencoder module and the augmented feature regression module involves the combination of the Chamfer loss and the transformation loss:
χ t o t a l = α χ C D s r c + χ C D t g t + β χ t + χ R
where α , β are scalar parameters and χ C D t g t denotes the Chamfer loss between Y and Y r e c .

3. Results

In this section, we train and evaluate our method using the synthetic ModelNet40 dataset [45] and the real-world 7Scene dataset [46]. First, we present the evaluation metrics and provide implementation details. Next, we compare the accuracy and robustness of our method with state-of-the-art approaches, including ICP [21], CPD [29], FGR [47], DeepGMR [39], PointNetLK [33], DCP [36], and FMR [35].

3.1. Evaluation Metrics

The registration performance is evaluated by the rotational and translational errors as follows:
ε R = arccos t r R ^ T R g 1 2 ε t = t ^ t g 2
where rotational and translational errors are measured in degrees and meters, respectively. The root means squared error (RMSE) and the mean absolute error (MAE) of registration results are reported for all the methods mentioned above.

3.2. Implementation Details

We trained DCP, PointNetLK, DeepGMR, FMR, and our method on the synthetic ModelNet40 dataset. The feature extraction module utilized the modified DGCNN as the encoder, with a k value set to 10 in the dynamic KNN graph. In the augmented regression network, the number of Gaussian components was set as 20. The learning-based methods mentioned above were conducted by using their official implementation. The other optimization-based PCR methods (CPD, ICP, FMR) were conducted by the Probreg library [48] and the Open3D implementation [49]. For all experiments, the proposed method was trained for 250 epochs and Adam [50] was utilized as the optimizer of the neural network. The learning rate was initially set to 0.001 and decreased by a factor of 10 at epochs 75, 150, and 200, resulting in a total training duration of 250 epochs. All experiments were conducted in PyTorch and trained on two NVIDIA Titan XP GPUs.

3.3. Registration Results in the ModelNet40 Dataset

In this experiment, we evaluated the registration method performances on the ModelNet40 dataset, composed of 3D CAD models from 40 different categories. To assess the generalization ability of various methods, we divided the dataset into training and testing sets, ensuring an equal distribution of categories. We trained learning-based methods using the first 20 categories for training and evaluated them on the last 20 categories for testing. The optimization-based methods were also evaluated using the last 20 categories. In total, 5088 pairs of point clouds were used to train the network, while 1266 pairs of point clouds were used for testing. During the experiment, each model’s surface was randomly sampled to obtain 1024 points, which served as the source point cloud. The point cloud was then normalized and adjusted to fit within a unit box at the origin [ 0 , 1 ] 3 . The target point cloud was generated by randomly applying rigid transformations to the source point cloud. Throughout the training and testing processes, a rotation matrix was randomly generated using three Euler rotation angles in the range of [ 0 , 45 ] , and the translation vector ranged from 0 to 0.5 m on each axis.

3.3.1. Robustness to Noise

The results of various algorithms on clean data are presented in Table 1. The comparison results demonstrate that our algorithm achieves state-of-the-art performance on the clean dataset. In particular, our algorithm achieves the highest accuracy in terms of translation. In the absence of noise, DCP can accurately compute the point-to-point correspondence between two point clouds, thus obtaining more accurate rotation estimation results. After multiple iterations, CPD can also obtain accurate point-to-point correspondences in noise-free experiments, so that it can iteratively optimize and estimate the rotation matrix more accurately. In addition, CPD uses an iterative optimization method with well-defined mathematical constraints, which gives it better generalization for unknown scenes.
To assess the noise robustness of the proposed method, experiments were conducted using a dataset containing noise. We adopted the trained model in the clean dataset and tested it on four levels of noise. The noises were sampled from N 0 , σ 2 and clipped to 0.05 , 0.05 on each axis, where σ 2 ranges from 0.01 to 0.04. Note that the Gaussian noise is independently injected into each point of the target point cloud, which eliminates the point-to-point correspondence. Figure 3 shows the comparison results under different levels of noise. Compared with PointNetLK, FMR, and DeepGMR, the proposed method achieves more robust performance on the noisy dataset, even when the noise is very large ( σ 2 = 0.04). DCP is more sensitive to noise because the noise breaks the one-to-one correspondence. As an extension of the learning-based probabilistic registration method, the proposed method aligns more accurately than DeepGMR due to the efficiency of the latent feature and augmented regression network. The qualitative results on the noisy dataset are presented in Figure 4, where the target point cloud is corrupted by Gaussian noise with a ratio of 0.04.

3.3.2. Robustness to Sparse Data

To demonstrate that the proposed method has significant robustness to sparse data, we downsampled the clean source point cloud at different density levels and evaluated the registration performance against other methods. Note that the density level of the source point cloud ranges from 50 % to 100 % , while the point number of the target point cloud is still 1024. Figure 5 summarizes the comparison results on different density levels. Based on the analysis presented in Figure 5, it is observed that the proposed method exhibits robustness to sparse data, even when the density level of the source point cloud is as low as 50%. Notably, DCP struggles to handle the registration when only 90% of the source point cloud is available. Additionally, both FMR and our method demonstrate superior performance compared to other methods. This finding supports the effectiveness of the semi-supervised learning approach in extracting features that capture the pose differences. The qualitative results are shown in Figure 6, where the density levels of the source point cloud are 50% and 70%.
To better compare the robustness of different algorithms to the sparse point cloud, the source and target point cloud were both downsampled with 512 points. In these experiments, the covariance of the noise σ 2 applied to the target point cloud was 0.04. The experimental results are shown in Table 2. When the point cloud becomes sparse, the feature extraction modules of PointNetLK and DeepGMR also have difficulty extracting effective point cloud features, which can lead to poorer registration accuracy compared to FMR and the proposed method. Due to the autoencoder-based feature extraction module and augmented point cloud global features, our algorithm maintains good registration accuracy in the case of sparse point clouds. Our algorithm achieves accurate registration even with sparse point clouds by incorporating an autoencoder-based feature extraction module and augmented point cloud global features.

3.3.3. Robustness to Different Transformation

To test the robustness of each method to different transformations, three rotation angles and the range of translation vectors were expanded to 0 , 60 and 0.5 , 0.5 m. In these experiments, the covariance of Gaussian noise applied to the target point was 0.01 σ 2 = 0.01 . The comparison results are presented in Table 3. As the spatial transformation range increases, the accuracy of the registration algorithm decreases, especially for learning-based registration algorithms. This is because different rotational transformations make it difficult for the feature extraction module to accurately extract point cloud features in a new space. Based on the findings displayed in Table 3, it can be concluded that the proposed algorithm effectively tackles registration challenges in this particular scenario. We believe that there are two reasons that can explain this advantage. Firstly, as a pre-processing step, point cloud normalization can effectively reduce the learning space of the neural network, thus ensuring the effectiveness of the feature extraction module. In addition, an autoencoder-based feature extraction module was employed to enhance the neural network’s capability to extract point cloud features across various learning domains. Moreover, augmented point cloud features were utilized to describe the point cloud.

3.3.4. Robustness to Partial Data

In order to evaluate the robustness of each method to partial overlap, a portion of each target point cloud was randomly removed. Consequently, in each registration experiment, only 90% of two point clouds exhibited overlap. The registration task involving the partial target point cloud and the full source point cloud is referred to as partial-to-full registration. It is important to note that these experiments were conducted using the model that was previously trained, without any retraining specifically for this registration scenario. Table 4 summarizes the error results of each algorithm in the partial-to-full registration experiment. For FMR and PointNetLK, their test models are obtained based on full-to-full registration. In the partial-to-full registration testing scenario, the absence of point cloud features derived from missing point cloud structures results in errors when calculating the difference between the features of the two point clouds. In contrast, registration methods based on the Gaussian mixture model demonstrate more robustness in this particular registration scenario. The proposed method can accurately estimate the rotation transformation of point clouds in comparison to DeepGMR, owing to the advantages of the feature extraction module and the latent GMM module. The proposed method also achieves similar accuracy to other state-of-the-art methods in estimating translation transformation. Figure 7 qualitatively demonstrates the registration results of CPD, DeepGMR, and the proposed algorithm in this registration scenario. In Figure 7, the point numbers of the source and target point clouds are 1024 and 921, respectively. According to Figure 7, CPD has poor accuracy in predicting the translation vector, and DeepGMR has poor accuracy in predicting the rotation matrix. The proposed method achieves better prediction accuracy in both rotation and translation.

3.4. Registration Results in the 7Scene Dataset

7Scene is a collection of the real-world indoor dataset, which contains seven scenes recorded by an RGB-D camera. In this experiment, we demonstrated the generalization ability and accuracy of our algorithm to handle the real-world scenes generated by different sensors. We adopted the trained model on the clean ModelNet40 dataset and tested it on the real-world 7Scene dataset. It means that we used 355 real-world scenes to validate the trained model. The initial rotation and translation were the same as in Section 3.3. In the real-world application, the point cloud obtained by LIDAR or RGB-D was always sparse. To test the performance on sparse data, we randomly sampled the source point cloud at different levels. The number of sampled source point clouds ranges from 1 k to 10 k.

3.4.1. Robustness to Sparse Data

Table 5 provides a summary of the comparison results from various methods on the real-world dataset. In the experiment described in Table 5, both point clouds consist of 1K points. The point cloud utilized in this experiment is the sparsest in comparison to other experiments conducted on the real-world dataset. The proposed method demonstrates superior performance compared to both classic optimization-based methods and more recent learning-based methods. The observation can be concluded that the autoencoder-based feature extraction module can extract distinctive and efficient features to represent the point cloud after training. Furthermore, compared with the point-to-point alignment strategies (ICP, DCP) and the feature correspondence strategies (FGR, PointNetLK, FMR), our point-to-GMM method achieves more accurate performance. Due to the difficulty in obtaining more accurate point cloud representations through the feature extraction module for sparse point clouds in large scenes, there are errors in the point-to-GMM correspondences. Therefore, the FMR algorithm, which directly compares the global feature differences between the two point clouds, achieved a smaller MAE result. The qualitative registration results of two scenes are shown in Figure 8, where the number of points is 1 k. According to Figure 8, the ICP, PointNetLK, and FMR algorithms fail to converge in both scenarios, rendering the registration algorithms ineffective. Compared to other algorithms, our algorithm exhibits lower rotation and translation errors in these two scenarios.
Figure 9 presents the registration results of different density levels in the real-world dataset. The results of the semi-supervised methods (FMR, ours) outperform supervised methods (PointNetLK, DeepGMR) in terms of rotational errors. This implies that the semi-supervised method can extract rotation-attentive features in two point clouds, effectively reflecting the rotation difference in the transformation estimation. Furthermore, the GMM-based methods (DeepGMR, ours) achieve more accurate translational results than other methods. We can observe that our algorithm obtains better accuracy and robustness under different numbers of input points.

3.4.2. Robustness to Partial Overlap

Similar to the experiments in Section 3.3.4, we conducted partial-to-full registration experiments on the 7Scene dataset. The point numbers of two point clouds were 5 k and 4.5 k, respectively. Table 6 summarizes the results of different registration methods. The proposed method shows better accuracy than other methods. As the point cloud density increases, the registration accuracy of DCP correspondingly increases, thanks to the use of transformer-based point-to-point correspondence estimation. However, this requires significant computational memory and time consumption. The qualitative comparison results are shown in Figure 10. In Figure 10, the point numbers of two point clouds are 5 k and 4.5 k, respectively. The DeepGMR algorithm exhibits the highest rotation error in this point cloud scene compared to CPD and our method. While the CPD algorithm has better registration accuracy than the DeepGMR algorithm in this scene, it still has rotation errors. In contrast, our algorithm achieves the best registration accuracy. Therefore, our algorithm demonstrates better robustness to overlap data.

3.5. Ablation Study

Feature extraction is the key module for the PCR. Several ablation studies were conducted to discuss the advantages of feature extraction modules in different methods. Table 7 shows quantitative results under the noisy ModelNet40 dataset for different methods. The first row shows the results of the original DeepGMR. In the Our-V1 experiment, we replaced PointNet with modified DGCNN as the feature extraction module. As shown in the second row, this method improves the rotational accuracy but fails to improve the translational accuracy. The Ours-V1 method achieves better results than the original DeepGMR in terms of rotational errors, which highlights the advantage of our modified DGCNN encoder. In the Our-V2 experiment, we add the decoder to guarantee the modified encoder to extract distinctive and efficient features, which contain pose-attentive information. As shown in the third row, adding the decoder can improve the efficiency of the feature extraction module, which drastically improves the registration performance. Note that the proposed method with the modified autoencoder as the feature extraction module achieves better performance in terms of rotational and translational accuracy.

3.6. Runtime Analysis

Table 8 compares the average runtime for each registration in different methods. In these experiments, the spatial transformation and noise applied to the point clouds in the registration experiments were the same. In the experiment on ModelNet40, the point numbers of the source and target point cloud were 1024. Although the proposed method adopts a more complicated neural network, the proposed method does not consume much runtime compared to DeepGMR in each registration experiment. Moreover, the proposed algorithm has some advantages in terms of runtime efficiency compared to algorithms such as PointNetLK and FMR. In order to test the influences of different sparsity levels of point clouds on the runtime efficiency of the registration algorithm, we conducted three comparative experiments on the 7Scene dataset. In these three experiments, the point numbers in the point clouds were 1 k, 5 k, and 10 k, respectively. Due to the more complex point cloud structures in real datasets, more runtime is required for each registration. Although the proposed method is slightly more time-consuming compared to other learning-based methods, considering the improvement in registration accuracy, we believe the slight increase in runtime is reasonable.
Due to the introduction of an autoencoder as a feature extraction module and enhanced point cloud features as input to the regression network, our method does indeed increase memory consumption. However, since the proposed method does not require computing correspondences between points, it does not result in significant memory usage. During the training process, even with a point cloud containing up to 5 k points, the GPU memory usage is still below 10G, whereas the DCP algorithm already exceeds 12G of memory usage. In scenarios involving large-scale point cloud registration, we suggest downsampling or extracting key points fromthe point cloud and using themas input to the registration network.

4. Discussion

This paper introduces a semi-supervised framework for solving the probabilistic PCR problem. This framework incorporates modified feature extraction and augmented GMM component estimation modules. These modules are particularly suitable for registering point clouds that are sparse and corrupted by significant noise. Extensive experimental results demonstrate that our algorithm achieves superior accuracy and robustness to both noise and sparsity when compared with state-of-the-art methods. The benefits of incorporating semi-supervised learning and a correspondence strategy are evident when comparing the results of the aforementioned learning-based PCR methods in various experimental settings. It is important to emphasize that the registration performance significantly benefits from both the semi-supervised learning framework and the augmented point-to-GMM correspondence module under various experimental conditions. To better evaluate the superiority of our algorithm compared to other learning-based PCR methods, we provide a comparison of the learning framework and correspondence strategy.

4.1. Supervised Learning or Semi-Supervised Learning

In the above-mentioned PCR methods, FMR and our method adopt the semi-supervised learning framework, while other methods adopt a supervised learning framework. According to the comparison results in Section 3.4, PCR methods with a semi-supervised learning framework achieve more accurate registration results. Moreover, FMR and our method have better robustness to sparse data and noise. These experimental results show that the semi-supervised learning framework can have better generalization from synthetic data to real-world data, which can have a wide application when the prior datasets are absent. We can conclude that the semi-supervised learning framework can extract the distinctive, efficient, and pose-attentive features from the limited dataset. These features help to estimate the correspondence matrix or feature matrix to achieve better final registration performance.

4.2. Correspondence Strategy

Among the learning-based PCR methods in the experiments, there are three different correspondence strategies. PointNetLK and FMR solve the PCR by minimizing the difference between the features of two point clouds. DCP adopts the transformer module to estimate the point-to-point correspondence. DeepGMR and our method both estimate the point-to-GMM correspondence to align two input point clouds. According to the comparison results in Figure 3, point-to-point correspondence may be broken by noise. Thus, DCP is sensitive to noise and unable to handle registration tasks with large noise. Compared with PointNetLK and FMR, DeepGMR and our method achieve more robust registration performance under different levels of noise. Moreover, given the latent GMM module with augmented features, our method has better robustness to different noise and sparse data. It can be concluded that accurate and robust correspondence is beneficial for PCR and applicable to handling registration tasks with noise and sparse data.

5. Conclusions

In this paper, we propose a semi-supervised method to solve the PCR problem. The proposed algorithm reduces dependence on the labeled dataset and is robust to the high level of noise. The proposed algorithm uses a modified DGCNN to extract features of point clouds and adopts the autoencoder framework to conduct a semi-supervised training method, which improves the efficiency and distinctiveness of the feature extraction module. Moreover, the augmented regression network is adopted to estimate the correspondences between point clouds and GMM components. Utilizing the point-to-GMM correspondence, the GMM parameters and rigid transformation are computed alternately. Comparison results on the ModelNet40 dataset validate that our algorithm is more robust to different levels of noise and sparsity. Experiments on the 7Scene dataset demonstrate the better generalization ability of our algorithm. In the ablation study, the advantages of different feature extraction modules are discussed. Considering the challenges of labeling and the high production cost associated with the medical point cloud dataset, there are very few public datasets that can be used to train the deep learning model. In the future, due to the good generalization performance and accuracy of the proposed algorithm, we will extend this approach to medical applications to solve the above problems.

Author Contributions

Conceptualization, Z.Z., E.L. and Z.M.; Methodology, Z.Z.; Software, Z.Z.; Validation, Z.Z.; Formal analysis, A.Z.; Supervision, M.Q.-H.M.; Project administration, Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wei, H.; Qiao, Z.; Liu, Z.; Suo, C.; Yin, P.; Shen, Y.; Li, H.; Wang, H. End-to-End 3D Point Cloud Learning for Registration Task Using Virtual Correspondences. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2020; pp. 2678–2683. [Google Scholar] [CrossRef]
  2. Li, L.; Yang, M.; Wang, C.; Wang, B. Robust Point Set Registration Using Signature Quadratic Form Distance. IEEE Trans. Cybern. 2020, 50, 2097–2109. [Google Scholar] [CrossRef] [PubMed]
  3. Hitchcox, T.; Forbes, J.R. A Point Cloud Registration Pipeline using Gaussian Process Regression for Bathymetric SLAM. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2020; pp. 4615–4622. [Google Scholar] [CrossRef]
  4. Lyu, E.; Liu, T.; Wang, J.; Song, S.; Meng, M.Q.H. Motion Planning of Manipulator by Points-Guided Sampling Network. IEEE Trans. Autom. Sci. Eng. 2022, 20, 821–831. [Google Scholar] [CrossRef]
  5. Hansen, L.; Heinrich, M.P. Deep learning based geometric registration for medical images: How accurate can we get without visual features? In Proceedings of the Information Processing in Medical Imaging: 27th International Conference, IPMI 2021, Virtual Event, 28–30 June 2021; Proceedings 27. Springer: Berlin/Heidelberg, Germany, 2021; pp. 18–30. [Google Scholar]
  6. Fu, Y.; Lei, Y.; Wang, T.; Patel, P.; Jani, A.B.; Mao, H.; Curran, W.J.; Liu, T.; Yang, X. Biomechanically constrained non-rigid MR-TRUS prostate registration using deep learning based 3D point cloud matching. Med. Image Anal. 2021, 67, 101845. [Google Scholar]
  7. Baum, Z.M.; Hu, Y.; Barratt, D.C. Real-time multimodal image registration with partial intraoperative point-set data. Med. Image Anal. 2021, 74, 102231. [Google Scholar] [PubMed]
  8. Si, H.; Qiu, J.; Li, Y. A review of point cloud registration algorithms for laser scanners: Applications in large-scale aircraft measurement. Appl. Sci. 2022, 12, 10247. [Google Scholar]
  9. Min, Z.; Wang, J.; Meng, M.Q.H. Robust Generalized Point Cloud Registration With Orientational Data Based on Expectation Maximization. IEEE Trans. Autom. Sci. Eng. 2020, 17, 207–221. [Google Scholar] [CrossRef]
  10. Min, Z.; Wang, J.; Pan, J.; Meng, M.Q.H. Generalized 3-D Point Set Registration With Hybrid Mixture Models for Computer-Assisted Orthopedic Surgery: From Isotropic to Anisotropic Positional Error. IEEE Trans. Autom. Sci. Eng. 2021, 18, 1679–1691. [Google Scholar] [CrossRef]
  11. Min, Z.; Zhu, D.; Ren, H.; Meng, M.Q.H. Feature-Guided Nonrigid 3-D Point Set Registration Framework for Image-Guided Liver Surgery: From Isotropic Positional Noise to Anisotropic Positional Noise. IEEE Trans. Autom. Sci. Eng. 2021, 18, 471–483. [Google Scholar] [CrossRef]
  12. Zhang, Z.; Min, Z.; Zhang, A.; Wang, J.; Song, S.; Meng, M.Q.H. Reliable Hybrid Mixture Model for Generalized Point Set Registration. IEEE Trans. Instrum. Meas. 2021, 70, 2516110. [Google Scholar] [CrossRef]
  13. Žagar, B.L.; Yurtsever, E.; Peters, A.; Knoll, A.C. Point Cloud Registration With Object-Centric Alignment. IEEE Access 2022, 10, 76586–76595. [Google Scholar] [CrossRef]
  14. Zováthi, Ö.; Nagy, B.; Benedek, C. Point cloud registration and change detection in urban environment using an onboard Lidar sensor and MLS reference data. Int. J. Appl. Earth Obs. Geoinf. 2022, 110, 102767. [Google Scholar]
  15. You, B.; Chen, H.; Li, J.; Li, C.; Chen, H. Fast point cloud registration algorithm based on 3DNPFH descriptor. Photonics 2022, 9, 414. [Google Scholar]
  16. Qin, Z.; Yu, H.; Wang, C.; Guo, Y.; Peng, Y.; Xu, K. Geometric Transformer for Fast and Robust Point Cloud Registration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11133–11142. [Google Scholar] [CrossRef]
  17. Zheng, Y.; Li, Y.; Yang, S.; Lu, H. Global-PBNet: A Novel Point Cloud Registration for Autonomous Driving. IEEE Trans. Intell. Transp. Syst. 2022, 23, 22312–22319. [Google Scholar] [CrossRef]
  18. Chen, Z.; Sun, K.; Yang, F.; Tao, W. SC2-PCR: A Second Order Spatial Compatibility for Efficient and Robust Point Cloud Registration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13211–13221. [Google Scholar] [CrossRef]
  19. Wu, Y.; Yao, Q.; Fan, X.; Gong, M.; Ma, W.; Miao, Q. PANet: A Point-Attention Based Multi-Scale Feature Fusion Network for Point Cloud Registration. IEEE Trans. Instrum. Meas. 2023, 72, 2512913. [Google Scholar] [CrossRef]
  20. Chen, Z.; Yang, F.; Tao, W. Detarnet: Decoupling translation and rotation by siamese network for point cloud registration. Proc. AAAI Conf. Artif. Intell. 2022, 36, 401–409. [Google Scholar]
  21. Besl, P.; McKay, N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar] [CrossRef]
  22. Segal, A.; Haehnel, D.; Thrun, S. Generalized-icp. Proc. Robot. Sci. Syst. 2009, 2, 435. [Google Scholar]
  23. Bouaziz, S.; Tagliasacchi, A.; Pauly, M. Sparse iterative closest point. Proc. Comput. Graph. Forum 2013, 32, 113–123. [Google Scholar]
  24. Rusinkiewicz, S. A symmetric objective function for ICP. ACM Trans. Graph. 2019, 38, 1–7. [Google Scholar]
  25. Agamennoni, G.; Fontana, S.; Siegwart, R.Y.; Sorrenti, D.G. Point Clouds Registration with Probabilistic Data Association. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 4092–4098. [Google Scholar] [CrossRef]
  26. Low, K.L. Linear Least-Squares Optimization for Point-to-Plane ICP Surface Registration; University of North Carolina: Chapel Hill, NC, USA, 2004; Volume 4, pp. 1–3. [Google Scholar]
  27. Chetverikov, D.; Stepanov, D.; Krsek, P. Robust Euclidean alignment of 3D point sets: The trimmed iterative closest point algorithm. Image Vis. Comput. 2005, 23, 299–309. [Google Scholar]
  28. Yang, J.; Li, H.; Jia, Y. Go-ICP: Solving 3D Registration Efficiently and Globally Optimally. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 1457–1464. [Google Scholar] [CrossRef]
  29. Myronenko, A.; Song, X. Point Set Registration: Coherent Point Drift. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 2262–2275. [Google Scholar] [CrossRef]
  30. Horaud, R.; Forbes, F.; Yguel, M.; Dewaele, G.; Zhang, J. Rigid and Articulated Point Registration with Expectation Conditional Maximization. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 587–602. [Google Scholar] [CrossRef] [PubMed]
  31. Evangelidis, G.D.; Kounades-Bastian, D.; Horaud, R.; Psarakis, E.Z. A generative model for the joint registration of multiple point sets. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 109–122. [Google Scholar]
  32. Ravikumar, N.; Gooya, A.; Frangi, A.F.; Taylor, Z.A. Generalised coherent point drift for group-wise registration of multi-dimensional point sets. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 309–316. [Google Scholar]
  33. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. PointNetLK: Robust amp; Efficient Point Cloud Registration Using PointNet. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7156–7165. [Google Scholar] [CrossRef]
  34. Charles, R.Q.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar] [CrossRef]
  35. Huang, X.; Mei, G.; Zhang, J. Feature-Metric Registration: A Fast Semi-Supervised Approach for Robust Point Cloud Registration Without Correspondences. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11363–11371. [Google Scholar] [CrossRef]
  36. Wang, Y.; Solomon, J. Deep Closest Point: Learning Representations for Point Cloud Registration. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3522–3531. [Google Scholar] [CrossRef]
  37. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  39. Yuan, W.; Eckart, B.; Kim, K.; Jampani, V.; Fox, D.; Kautz, J. Deepgmr: Learning latent gaussian mixture models for registration. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 733–750. [Google Scholar]
  40. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  41. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in Neural Information Processing Systems; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  42. Gao, G.; Lauri, M.; Hu, X.; Zhang, J.; Frintrop, S. CloudAAE: Learning 6D Object Pose Regression with On-line Data Synthesis on Point Clouds. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 11081–11087. [Google Scholar] [CrossRef]
  43. Liu, X.; Han, Z.; Wen, X.; Liu, Y.S.; Zwicker, M. L2g auto-encoder: Understanding point clouds by local-to-global reconstruction with hierarchical self-attention. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 989–997. [Google Scholar]
  44. Svensén, M.; Bishop, C.M. Pattern recognition and machine learning. Technometrics 2007, 49, 366. [Google Scholar]
  45. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A deep representation for volumetric shapes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar] [CrossRef]
  46. Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3DMatch: Learning Local Geometric Descriptors from RGB-D Reconstructions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 199–208. [Google Scholar] [CrossRef]
  47. Zhou, Q.Y.; Park, J.; Koltun, V. Fast global registration. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 766–782. [Google Scholar]
  48. Kenta-Tanaka et al. Probreg. Available online: https://probreg.readthedocs.io/en/latest/ (accessed on 23 July 2023).
  49. Zhou, Q.Y.; Park, J.; Koltun, V. Open3D: A Modern Library for 3D Data Processing. arXiv 2018, arXiv:1801.09847. [Google Scholar] [CrossRef]
  50. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
Figure 1. The overview of the proposed PCR method. The equations denoted by the green dashed line are used to estimate the latent source GMM component, while the equations denoted by the orange dashed line in the lower right corner are used in the weighted SVD module.
Figure 1. The overview of the proposed PCR method. The equations denoted by the green dashed line are used to estimate the latent source GMM component, while the equations denoted by the orange dashed line in the lower right corner are used in the weighted SVD module.
Remotesensing 15 04493 g001
Figure 2. The architecture of the proposed autoencoder module.
Figure 2. The architecture of the proposed autoencoder module.
Remotesensing 15 04493 g002
Figure 3. Comparison results on the noisy dataset. Even when noise is relative large ( σ 2 = 0.04), the proposed method achieves accurate and robust performances.
Figure 3. Comparison results on the noisy dataset. Even when noise is relative large ( σ 2 = 0.04), the proposed method achieves accurate and robust performances.
Remotesensing 15 04493 g003
Figure 4. Qualitative registration results on the noisy ModelNet40 dataset ( σ 2 = 0.04 ). The figures in three rows show the registration results of three classes of objects, including person, stool, and toilet. The source and target point cloud are shown in green and blue, respectively.
Figure 4. Qualitative registration results on the noisy ModelNet40 dataset ( σ 2 = 0.04 ). The figures in three rows show the registration results of three classes of objects, including person, stool, and toilet. The source and target point cloud are shown in green and blue, respectively.
Remotesensing 15 04493 g004
Figure 5. Comparison results on different density levels. The results confirm that our algorithm consistently exhibits accuracy and robustness in sparse PCR experiments.
Figure 5. Comparison results on different density levels. The results confirm that our algorithm consistently exhibits accuracy and robustness in sparse PCR experiments.
Remotesensing 15 04493 g005
Figure 6. Qualitative registration results on noisy ModelNet40 dataset ( σ 2 = 0.04 ). Figures in two rows show the registration results of two classes of objects, including wardrobe and table. The results show that our algorithm has better robustness to the used sparse data.
Figure 6. Qualitative registration results on noisy ModelNet40 dataset ( σ 2 = 0.04 ). Figures in two rows show the registration results of two classes of objects, including wardrobe and table. The results show that our algorithm has better robustness to the used sparse data.
Remotesensing 15 04493 g006
Figure 7. Qualitative registration results on partial data. The figures show the registration results of CPD, DeepGMR, and ours.
Figure 7. Qualitative registration results on partial data. The figures show the registration results of CPD, DeepGMR, and ours.
Remotesensing 15 04493 g007
Figure 8. Qualitative registration results on the real-world dataset. Figures in two rows show the registration results of two classes of scenes, including the kitchen and office.
Figure 8. Qualitative registration results on the real-world dataset. Figures in two rows show the registration results of two classes of scenes, including the kitchen and office.
Remotesensing 15 04493 g008
Figure 9. Comparison results of different density levels on the 7Scene dataset. The point numbers of the source and target point clouds range from 1 k to 10 k. In all registration experiments, our method maintains accuracy and robustness.
Figure 9. Comparison results of different density levels on the 7Scene dataset. The point numbers of the source and target point clouds range from 1 k to 10 k. In all registration experiments, our method maintains accuracy and robustness.
Remotesensing 15 04493 g009
Figure 10. Qualitative registration results on partial data in the 7Scene dataset. The figures show the registration results of CPD, DeepGMR, and ours.
Figure 10. Qualitative registration results on partial data in the 7Scene dataset. The figures show the registration results of CPD, DeepGMR, and ours.
Remotesensing 15 04493 g010
Table 1. ModelNet40: Comparison results of the unseen categories without Gaussian noise. The best results are displayed in bold, while the second-best results are underlined.
Table 1. ModelNet40: Comparison results of the unseen categories without Gaussian noise. The best results are displayed in bold, while the second-best results are underlined.
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
ICP30.35140823.8793470.2912030.250332
FGR25.92291522.4601840.0065550.004499
CPD4.7865440.7629300.0031450.000309
PointNetLK9.0471261.7364420.0395450.006354
DeepGMR4.9021932.5412450.0030740.002126
DCP2.6827131.8027360.0050240.003697
FMR5.9399451.2349020.0203270.003700
Ours3.0640521.6098430.0013250.000938
Table 2. ModelNet40: Comparison results of the test on ModelNet40. The point numbers of two point clouds are both 512. The best results are displayed in bold, while the second-best results are underlined.
Table 2. ModelNet40: Comparison results of the test on ModelNet40. The point numbers of two point clouds are both 512. The best results are displayed in bold, while the second-best results are underlined.
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
ICP39.79975936.8631480.2881500.250225
FGR16.4737265.0501440.0409780.015393
CPD9.0016744.4578610.0036200.028238
PointNetLK19.2436514.8519670.0291230.005638
DeepGMR4.7701062.5638120.0030020.002057
DCP38.30571021.3167100.0092960.006652
FMR10.1199063.5696910.0195050.007553
Ours3.3733271.8800490.0015260.001117
Table 3. ModelNet40: Comparison results of the test on the expanded transformation. The best results are displayed in bold, while the second-best results are underlined.
Table 3. ModelNet40: Comparison results of the test on the expanded transformation. The best results are displayed in bold, while the second-best results are underlined.
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
ICP48.18373144.3498800.2881500.250225
FGR26.29852110.5526040.0515240.022009
CPD19.9176549.4090870.0416130.030905
PointNetLK38.74749416.4342450.0479380.016704
DeepGMR9.3281544.0623770.0039800.002622
DCP11.7185255.8869140.0030530.002211
FMR26.88746212.4224950.0391900.018913
Ours6.2209093.2034930.0016160.001138
Table 4. ModelNet40: Comparison results of the test on partial data. The point numbers of two point clouds are 1024 and 921, respectively.
Table 4. ModelNet40: Comparison results of the test on partial data. The point numbers of two point clouds are 1024 and 921, respectively.
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
ICP25.92291522.4601860.4944280.410775
FGR27.22924610.6125480.0539870.027252
CPD11.5537485.6243170.0428790.032712
PointNetLK28.03674711.7870260.0668980.040418
DeepGMR24.42221111.9742140.0413030.036290
DCP31.40671318.3138480.0446680.038792
FMR28.32853714.1105570.0450460.029118
Ours8.3400024.3528210.0439250.038332
Table 5. 7Scene: Comparison results of the test on real-world scenes. The best results are displayed in bold, while the second-best results are underlined.
Table 5. 7Scene: Comparison results of the test on real-world scenes. The best results are displayed in bold, while the second-best results are underlined.
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
ICP25.62171022.0873550.2858770.246103
FGR32.46610317.4003280.0269840.012873
CPD13.1852467.7728480.0372700.028007
PointNetLK15.2882352.8121650.0131350.002467
DeepGMR13.4887776.8922670.0026480.001648
DCP5.8592643.3147870.0055330.003899
FMR6.9392861.4008400.0087160.001800
Ours5.2953273.9096640.0012430.000864
Table 6. 7Scene: Comparison results of the test on partial real-world scenes.
Table 6. 7Scene: Comparison results of the test on partial real-world scenes.
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
ICP25.92291322.4601840.3560670.292270
FGR16.8243229.7798900.0330030.021518
CPD17.64369810.2072560.0474800.034908
PointNetLK13.7176976.5523530.0345060.021542
DeepGMR28.63736317.1253180.0255910.021899
DCP9.9640907.0423270.0262800.022408
FMR43.03485930.9419800.0549100.039119
Ours7.4649275.5612170.0257300.021929
Table 7. Comparison results of the ablation study on the noisy ModelNet40 dataset ( σ 2 = 0.01 ).
Table 7. Comparison results of the ablation study on the noisy ModelNet40 dataset ( σ 2 = 0.01 ).
Rotation Error ( )Translation Error (m)
RMSEMAERMSEMAE
DeepGMR6.4165412.7749740.0030530.002131
Ours-V13.1460871.5359210.0031670.002299
Ours-V23.1031101.6270200.0013010.000921
Table 8. Comparison of registration runtime results for different algorithms.
Table 8. Comparison of registration runtime results for different algorithms.
Registration Runtime (ms)
ModelNet407Scene
10241 k5 k10 k
ICP1.16486.48506.50535.04
FGR26.80507.22587.40896.00
CPD6.36490.74493.65509.09
PointNetLK64.32156.96261.15454.83
DeepGMR3.72485.04480.81494.96
DCP11.95493.921828.992431.65
FMR22.23479.49479.18478.65
Ours9.20486.61535.36646.58
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Z.; Lyu, E.; Min, Z.; Zhang, A.; Yu, Y.; Meng, M.Q.-H. Robust Semi-Supervised Point Cloud Registration via Latent GMM-Based Correspondence. Remote Sens. 2023, 15, 4493. https://doi.org/10.3390/rs15184493

AMA Style

Zhang Z, Lyu E, Min Z, Zhang A, Yu Y, Meng MQ-H. Robust Semi-Supervised Point Cloud Registration via Latent GMM-Based Correspondence. Remote Sensing. 2023; 15(18):4493. https://doi.org/10.3390/rs15184493

Chicago/Turabian Style

Zhang, Zhengyan, Erli Lyu, Zhe Min, Ang Zhang, Yue Yu, and Max Q.-H. Meng. 2023. "Robust Semi-Supervised Point Cloud Registration via Latent GMM-Based Correspondence" Remote Sensing 15, no. 18: 4493. https://doi.org/10.3390/rs15184493

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop