Next Article in Journal
Investigating the Neural Mechanisms of Self-Controlled and Externally Controlled Movement with a Flexible Exoskeleton Using EEG Source Localization
Previous Article in Journal
Secure Fusion with Labeled Multi-Bernoulli Filter for Multisensor Multitarget Tracking Against False Data Injection Attacks
Previous Article in Special Issue
Robust Human Tracking Using a 3D LiDAR and Point Cloud Projection for Human-Following Robots
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Elastic Fine-Tuning Dual Recurrent Framework for Non-Rigid Point Cloud Registration

1
Hefei Institute of Physical Science, Chinese Academy of Sciences, Hefei 230031, China
2
Zhongke Technology Achievement Transfer and Transformation Center of Henan Province, Zhengzhou 450046, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3525; https://doi.org/10.3390/s25113525
Submission received: 25 April 2025 / Revised: 30 May 2025 / Accepted: 2 June 2025 / Published: 3 June 2025

Abstract

:
Non-rigid transformation is based on rigid transformation by adding distortions to form a more complex but more consistent common scene. Many advanced non-rigid alignment models are implemented using supervised learning; however, the large number of labels required for the training process makes their application difficult. Here, an elastic fine-tuning dual recurrent computation for unsupervised non-rigid registration is proposed. At first, we transform a non-rigid transformation into a series of combinations of rigid transformations using an outer recurrent computational network. Then, the inner loop layer computes elastic-controlled rigid incremental transformations by controlling the threshold to obtain a finely coherent rigid transformation. Finally, we design and implement loss functions that constrain deformations and keep transformations as rigid as possible. Extensive experiments validate that the proposed method achieves state-of-the-art performance with 0.01219 earth mover’s distances (EMDs) and 0.0153 root mean square error (RMSE) in non-rigid and rigid scenes, respectively.

1. Introduction

Motivated by the development of the meta-universe industry, 3D models demonstrate enhanced applicability in small-scale indoor entertainment environments, such as indoor outfit change. The 3D point cloud obtained through consumer-level depth sensing devices tends to become the main carrier of information exchange between the real world and parallel virtual worlds [1,2,3,4]. The small-scale indoor 3D point data manifests two characteristics: (1) the low-cost depth-sensing device usually obtains partial geometric representations due to constrained viewpoint sampling; (2) the short-range measurement configuration captures detailed geometric features and demonstrates sensitivity to morphological deformations. Non-rigid registration technology emilites point cloud global transformations and local distortions to align source and target objects, which is the fundamental step of 3D data application. Although much pioneering research of point cloud registration has been proposed, the non-rigid registration problem is still a difficult problem that many researchers continue to study [5,6,7,8].
In many traditional optimization-based studies, non-rigid registration usually converts shape distortions’ (may differ by bending and stretching) alignment to minimize the distance between surfaces with regular expressions, such as Tukey, Welsch, and Geman–McCLure, which corrects the registration process in terms of smoothness, point location, local topology, etc. [9,10,11]. However, the discrete nature of the point cloud prevents the use of micro-miniaturized mathematical properties in computation. Furthermore, many studies have transformed the non-rigid alignment problem into a nonlinear and non-convex mathematical optimization problem. As a representative, coherent point drift (CPD) based on optimal probability distribution is the most widely used non-rigid registration method [12]. Registration based on iterative computation has the problems of low convergence speed and a poor smoothing effect, which makes the non-rigid alignment focus on solving the dense real correspondence point relationship [13,14,15]. This model-driven point correspondence estimation is obtained by calculating the correlation between manual feature points and descriptors, which is not deformation-resistant and is easily affected by noise, density, and overlap rate variations [11,16,17,18].
In recent years, the learning-based framework has demonstrated the capability to establish data-driven point cloud correspondences through feature-based matching or shape-matching techniques [5,19,20,21,22]. The problems of these methods conclude that a large number of labeled data are required during training processes or predetermined internal structures are set, such as lattice representations, in the input data, which significantly increases the difficulty of acquiring training datasets and restricts broader applications. Currently, many non-rigid registration models based on unsupervised learning have been proposed and demonstrated state-of-the-art performance [6,23,24,25,26]. However, existing similarity-constrained loss functions have substantial deficiencies in effectively driving the network to learn optimal solutions.
To address these challenges, this study investigates unsupervised learning for non-rigid registration by leveraging complementary advantages of feature and shape matching while considering the characteristics of small-scale scene point clouds. Combining the recurrent neural network representation of the LK algorithm, we propose a dual recurrent registration framework, NGRLK, that decomposes non-rigid transformations into a series of elastic fine-tuning rigid transforms. To be specific, based on the computation process with iteratively solving rigid transform increments, the point pairs’ correspondence can be continuously computed and updated. Through updating the weights of rigid transformation through an attention mechanism to eliminate the influence of invalid transformations, the latest transformed source point cloud is concatenated from the combination of rigid transformation with corresponding weight in each round, which is fed into another recurrent neural network to output the aligned results in an end-to-end manner. Moreover, to tackle the shape similarity metric problem, we further design a loss function of local spatial consistency measure to constrain deformation similarity, which is complementary to the two-dimensional manifold representation of the point cloud for deformation shape similarity measurement. To the best of our knowledge, extensive experiments have validated that the NGRLK proposed in this paper has advanced performance compared with other SOTA non-rigid registration methods on both synthetic and real scanning datasets.
Thus, the main contributions of this paper are as follows:
(1)
We present that a point cloud non-rigid transform can comprise some rigid transforms with different weights.
(2)
We achieve fine-tuning incremental changes between successive rigid transformations, which increases the coherence and edge smoothness from rigid to non-rigid transformations.
(3)
We propose a local spatial consistency metric loss function to compute similarities between the transformed neighborhoods of corresponding points, which makes the rigid incremental transformations sufficiently small.
(4)
Extensive experiments have been conducted on various datasets for the point cloud non-rigid and rigid registration, which shows that the method proposed in this paper has SOTA performance.

2. Related Work

Rigid registration based on the PointNet framework. As a milestone in point cloud feature learning, PointNet is the first end-to-end point cloud solution that can directly perform feature learning on points [27]. Many excellent point cloud feature extraction methods based on the PointNet framework have been proposed. To improve the expression abilities of the point features extracted from the PointNet, the method mentioned in [28] treats the PointNet features as global features that are joined with local features. For the problem of information loss of maximum pooling in PointNet, PointNetX conducts compensation by stratifying and fusing global features from different layers [29]. Specifically for point cloud registration, inspired by PointNet, PPFNet embeds global information with local feature descriptors to enhance the expressiveness of features, which facilitates matching accuracy [30]. PPF-FoldNet utilizes PointNet as an encoder–decoder for point features and optimizes the learning model based on the Chamfer loss function to obtain correspondences [31]. PointNetLK utilizes PointNet to extract point cloud data global features and computes the transformation matrix with an inverse synthesis algorithm [32]. PCRNet [33], inspired by PointNetLK, employs PointNet to encode location and shape information into feature vectors for alignment, which effectively improves the robustness to noise. Inspired by these methods, we have designed an incrementally controllable rigid transformation estimation based on the recurrent neural network representation of the LK algorithm. Different from previous studies, our approach is not only applicable to rigid registration but can also be extended to non-rigid registration.
Non-rigid registration based on supervised learning. Different from registration models with iterative optimization computation affected by variations in noise, density, and overlap to degrade performance, the convolutional neural network deeply extracts point information to obtain distinctive feature expressions, which determines highly accurate correspondence [5]. By learning the surface representation of discrete surfaces, DiffusionNet computes the correspondence between a pair of shapes to accomplish non-rigid transformation [34]. For the DispVoxNets, the non-rigid registration result can be directly obtained through the learning of displacement fields and parameter representations. By performing a voxel mesh representation of point clouds, the convolutional neural network is utilized to compute the voxel displacement field to accomplish overall and refine deformation [35]. The SMPL model (Skinned Multi-Person Linear Model) is widely used to characterize human morphology, and the model parameters are continuously trained and optimized for the non-rigid registration of human 3D morphology [20,21,22]. Among many learning-based methods, the correspondence performance with the transform mechanism stands out. Trappolini et al. [36] designed a surface attention mechanism to obtain point correspondences, which was insensitive to point cloud density. Boscaini et al. [37] trained the high-accuracy shape feature correspondences by anisotropy fused with a convolutional neural network in different weights. To improve the correspondence accuracy with missing data, Attaiki et al. [19] integrated the shape features with a cross-attention module, which assigned the non-overlapping features a lower weight to preserve the overlapping regions as much as possible. All these supervised learning methods minimize errors and require labeled correspondence, which is usually unavailable for registration tasks. Incorporating shape feature correspondences with attention mechanisms, this paper focuses on non-rigid registration estimation through unsupervised learning.
Non-rigid registration based on unsupervised learning. Through registration error constraints, such as Gaussian loss and Chamfer distance, in conjunction with optimization regularity terms, many non-rigid registration methods that directly model the data have emerged. The PR-Net models the point cloud to be aligned based on thin plate spline templates, which utilize mesh assistance for shape feature learning and constrain the alignment accuracy by GMM losses [23]. The CPD-Net learns geometric transformations of varying complexity through a neural network and uses Chamfer distances to constrain the alignment direction, which is appropriate for scenarios with anomalous and incomplete point correspondences [24]. Halimi et al. [25] designed an unsupervised learning method for approximating the preservation of surface metric structures due to deformations, which conducted non-rigid registration by constraining the minimum deformation correspondence. The RMA-Net designs a continuously differentiable end-to-end cyclic framework to iteratively solve small rigid transformations and approximates global non-rigid transformations by solving multi-view projected shape similarity errors [6]. The CorrNet3D designs and implements a symmetric deformer in conjunction with a 2D flow shape and point correspondence, which constrains the learning process with point-to-point loss [7]. The GP-Aligner designs a model-free descriptor to characterize shape correlations among groups of point sets, which implements a novel unsupervised method [38]. Currently, unsupervised non-rigid registration methods have demonstrated state-of-the-art (SOTA) performance. However, most existing methods predominantly depend on per-point features or shape constraints. In contrast, the approach proposed in this paper integrates the complementary strengths of feature matching and shape constraints, without any shape templates, achieving a dual recurrent non-rigid registration framework. Furthermore, a novel loss function is designed to enforce spatial similarity constraints.

3. Problem Formulation

Inspired by the RMANet implementation, we approximately convert the non-rigid transforms into a series of rigid transforms of different weights by recurrent neural networks. Moreover, we incorporate the recurrent rigid transformation computation similar to PointNetLK to obtain rigid incremental transformations. Ultimately, we convert the non-rigid alignment process into a two-layer recurrent point cloud feature transform fitting computation and propose our non-rigid registration framework NGRLK, as shown in Figure 1.
Let f denote the point feature extraction R N × 3 R N + 1 × K and N be the point number. The N K-dimensional vectors embedded with global and local features are computed with f for a point cloud set. When two point clouds are closer to each other, the features are more similar. The point cloud registration problem can be converted into solving the transformation of point cloud features, and the non-rigid point cloud registration can be defined in Equation (1):
f T f r · S
where T denotes the target point cloud, S is the source point cloud, and r represents a rigid transformation.
Based on the derivation of the Inverse Compositional (IC) formulation in reference [39], Equation (1) can be transformed into Equation (2):
f S f T + ε f r 1 · t ε
where t is an exponential mapping containing the distortion parameter ε and ε f r 1 · t denotes the Jacobi determinant J.
The rigid incremental transformation r is obtained by the LK algorithm in pointnetLK [32] for each column in J and expressed as Equation (3):
J c o l = f exp ε t · T f T ε ε = J + f S f T r 1 = exp i ε i t
where ε is the infinitesimal perturbation of ε and usually empirically set as some small constant value over all iterations, J + is the Moore–Penrose generalized inverse of the Jacobi determinant, and i is the number of columns in J .
For the approximate equality sign instead of the equality sign = in Equations (1) and (2), the above equations express a single rigid accumulation transformation. For each new rigid transformation r = r 1 · r 2 · r 3 · r t h , the number of cycles t h is determined by adjusting the t h r e s h o l d value and able to control the amplitude of the rigid incremental transformation, and the corresponding rigidity cumulative transformations R = r should be transformed. For f r · S , S is the most recently transformed source point cloud instead of the initial source point cloud. Through the attention mechanism, each transformed source point cloud is assigned a weighted value ω S , and the non-rigid transform should be expressed as an accumulation of transformed source point clouds with different weights, as shown in Equation (4):
f T = i = 1 k w i f i r · S
where the weights w should satisfy the restriction i = 1 k w i = 1 to eliminate jumps like the joints of structural components.

4. Methodology

The non-rigid registration modeling framework NGRLK proposed in this paper mainly contains feature extraction, approximation transformation module, and loss function constraints modules. The key to this achievement is the combination of point cloud feature computation, point cloud feature correspondence estimation, elastic rigid transformation computation, and non-rigid transformation constraints. Although the outer-layer recurrent rigid computation framework is similar to the RMANet, the rigid transformation is elastically controllable by the inner-layer recurrent computation. Moreover, different from PointNetLK, the registration ability of NGRLK has been extended from rigid to non-rigid scenes.

4.1. Feature Extraction

The feature acquisition module in the NGRLK is independently pluggable and can be adapted to many feature extraction methods, such as PointNet [27], DGCNN [40], DIRGMR [41], etc. In PointNetLK, the point cloud feature is extracted by the PointNet, which fails to portray localized information and makes the characterization incomplete. As an improvement, we utilize an advanced feature extraction method published by DGCNN, which represents the point cloud as a graph structure and considers the topology of neighboring points to form a strong representation. The target point cloud features are used to compute the Jacobi J and jointly performed with the source point cloud features to obtain the similarity estimation, which can be transformed into a point correspondence matrix. In the whole process, the feature extraction of the target point cloud only needs to be performed once, and the features of the source point cloud and the point correspondence estimation between the two point cloud sets need to be computed several times.

4.2. Approximation Transformation Module

The key implementation of NGRLK lies in the double-loop computation of the fine rigid transformation, which is superimposed to form the final non-rigid transformation. First, the NGRLK simulates the overall process of approximating rigid transformation to non-rigid transformation through the first outer iterative computation, which is essentially a recurrent neural network. Inspired by the implementation of RAFT and RMA-NET [6], the NGRLK learns the weight coefficients of r changes by utilizing a gated loop unit GRU update incorporated with an attention mechanism. From the initial state of the GRU unit, each iterative computation generates an update direction. The transformed features, correlations, and hidden states serve as inputs to the GRU, which outputs the updated hidden state and the rigid transformation. Before performing the first computation, the feature f S 0 extracted from DGCNN and zero-initialized hidden state h s 0 are employed as the initial feature and hidden state to ensure unbiased attention allocation in the model. Furthermore, the adaptive optimizer Adam is adopted to ensure computational convergence and precision. In the n-th computation, the input of GRU contains the hidden state h s n 1 and the feature f S n 1 . Referring to [6], the GRU updates the hidden state as h s n . The dense layers in GRU are replaced with MLPs, which update the W list formed by weight recalculation for the n-th target point cloud rigid transformation with the attention mechanism. The magnitude value indicates the influence of the corresponding rigid transformation. Then, referring to the LK algorithm in PointNetLK [32], the NGRLK computes the source point cloud rigid transformation increment r in the second inner layers, which increases the coherence of the transformation by controlling the threshold to change the r computation, as shown in Figure 2.
For the n-th inner-layer computation, let the incrementally transformed source point cloud be denoted as S n . At the n-th outer-layer computation, the transformed source point cloud is denoted as S n . The incremental transformation follows Equation (5):
S n = r ( r 1 · r 2 · r 3 · r t h ) S n 1
For computational simplicity, this paper sets the threshold high for only one inner-layer incremental rigid transformation computation in each round of outer-layer recurrent computation, and the incremental transformation can be simplified as Equation (6):
S n = r ( r 1 ) S n 1
Then, the similarity of f T and f S n is compared to obtain the scoring list s i m i l a r i t y   s c o r e n , and the transformation weights are updated according to the normalization of the s i m i l a r i t y   s c o r e n and composed into a new weight vector w 1 , w 2 , , w n . The hidden state is updated and the source point cloud S is updated according to Equation (7):
S n = w 1 S 1 + w 2 S 2 + + w n S n
where the normalization of the score list ensures that the accumulative sum of w 1 to w n is 1. The computation details are presented in Algorithm 1.
Algorithm 1 Inner-layer computation
1input: point cloud feature f T , point cloud feature f S n 1 , Jacobi matrix J of T
2output: the n-th stage point cloud S n , weight vector w 1 , w 2 , , w n
3begin
4      attention compute f T as f T
5      compute r i with f T , f S n 1 , and J
6for  r i > threshold value
7      update S n = r i S n 1
8       i = i + 1 and compute r i with f T , f S n , and J
9 end for
10      obtain incremental transformation r = r 1 · r 2 · r 3 · r i
11      compute f S n with incremental transformation r and f S n 1
12      attention compute f S n as f S n
13      update s i m i l a r i t y   s c o r e n to form new weight vector w 1 , w 2 , , w n
14end

4.3. Loss Function Constraints

For unsupervised learning, the loss function directly affects the learning results. For the non-rigid change, we can constrain the shape change by taking the chamfer distance (CD) as the loss function, which can be expressed as Equation (8):
L C D = | | S T | | 2 + | | T S | | 2
However, for severe surface deformation or training overfitting, L C D fails in deforming S to T effectively. During each iterative process, the correspondence matrix between the features of the deformed S and T is recomputed. According to the AtlasNet surface generation method [42], the correspondence matrix can be regarded as a continuous smooth 2D manifold structure, and the multiplication of S with the correspondence matrix, denoted as S ~ , can realize the mapping of S to a 2D manifold. Referring to the reconstructed point cloud method used by CorrNet3D [7], the similarity between f S n and f T is computed and the point correspondence can be viewed as continuously smoothed 2D manifolds·. The deformed source point cloud S ~ is obtained by S n cross-multiplying with the correspondence matrix, which preserves the features of S n . The S ~ is concatenated with the extension term of f G T to form a combination [ S ~ , e x p a n d f T ], and the reconstruction from S ~ to the target point cloud T ^ is realized by the MLP computation at once, which is an important output of the outer-layer and shown in Figure 3.
According to Equation (9), the difference between T and T ^ can be minimized.
L s u r f = | | T T ^ | | 2
For the point correspondence matrix C, the constraint term expressed in Equation (10) is utilized to minimize the one-to-many point matching.
L c o r r = | | C C T I | | 2
where I denotes the unit matrix.
The spatial distribution of a local point neighborhood can be effectively characterized by a Gaussian model, of which the probability density function can be expressed in Equation (11).
f x = 1 2 π σ e x p x μ 2 2 σ 2
where μ and σ 2 are denoted as the mean and variance. For the multidimensional data, the Gaussian probability density function (PDF) can be expressed in the form of Equation (12).
P x θ = 1 2 π D 2 1 2 e x p X μ T 1 X μ 2
where D denotes the dimensionality, μ represents the mean vector across all dimensions, and signifies the covariance matrix. Based on Equation (12), the 3D point cloud Gaussian distribution can be represented in Equation (13).
g a u s s P = 1 2 π 3 2 ( P ) 1 2 e x p ψ P
where ( P ) denotes the covariance matrix of point cloud P , ψ P represents 1 2 P x y z μ P T 1 P x y z μ P , and μ P is the component-wise mean values of P . The degree of divergence between probability distributions can be effectively measured by the Kullback–Leibler (KL) divergence. Particularly advantageous is the existence of a closed-form expression for the KL divergence between Gaussian distributions, which substantially simplifies computational implementation. The amount of change in the relative position of the points in local space caused by each transformation of the source point cloud should be sufficiently small. We designed a loss function based on the KL to constrain the spatial distribution of the points as close as possible, as in Equation (14).
L s p a c e = p , q E D K L ( p | | q )
where E is the k-nn neighborhood of the same point (p and q) before and after the incremental transformation of the source point set and D K L denotes the relative entropy of the different distributions. Equation (14) essentially computes the mean log difference between g a u s s p and g a u s s q concerning the former, and as the two distributions get closer together, the value of KL becomes smaller, and vice versa the value will keep increasing. For the KL divergence computation in Equation (14), the detailed mathematical derivation can be expanded as shown in Equation (15).
D K L ( p | | q ) = g u a s s p x l o g g u a s s p x g u a s s q x d x
Based on Equation (13), the detailed expression of A can be derived as presented in Equation (16).
g u a s s p x g u a s s q x = q 1 / 2 p 1 / 2 e x p   ( ψ q ψ p )
By applying logarithmic operations and evaluating each quadratic term in Equation (15) with the expectation characteristics of Gaussian distributions, the final computation formulation is presented in Equation (17).
D K L ( p , q ) = 1 2 l o g q p + T r q 1 p 3 + ϕ
where T r is the matrix trace and ϕ is ( μ q μ p ) T q ( μ q μ p ) .
By observing the distribution of points in the neighborhood space before and after the transformation of the point clouds, as shown in Figure 4, this constraint can efficiently keep the change in the local space point cloud distribution small to make the transformation as rigid as possible.
The final loss function is formed as shown in Equation (18):
L o s s = λ 1 L C D + L s u r f + λ 2 L c o r r + λ 3 L s p a c e
This section represents the core component of this research. Although an efficient recurrent neural network, such as GRU, is employed, the iterative computation process and loss function evaluation involve quite computation, and efficiency is not an advantage when processing large-scale or complex point clouds. Thanks to the stable sampling environment of small indoor scenes, a uniformly distributed 3D model can usually be obtained. By employing down-sampling algorithms, the computational load can be effectively reduced, making it particularly suitable for non-real-time human–computer interaction experience scenarios.

5. Experimentation and Analysis

We implemented the proposed NGRLK model in Python v3.7. The experiment environment is configured as a cluster with the SIMT accelerator made in China. The cluster includes many nodes, each containing one CPU and four accelerators. The CPU has four NUMA nodes, and each NUMA node has 8 × 86-based processors. The accelerator, made in China, adopts a GPU-like architecture consisting of a 16 GB HBM2 device memory and many compute units. For accelerators connected to the CPU with PCI-E, the peak bandwidth of the data transmission between main memory and device memory is 16 GB/s.

5.1. Dataset Processing

Non-rigid dataset. Surreal is a standard dataset for large-scale personnel generated from RGB videos, which is 230 K and contains depth, body part, optical flow, 2D/3D human pose, and surface normal information. In this experiment, 115 K datasets are randomly generated for non-rigid alignment training. Moreover, the SHREC is a very challenging dataset for the stronger pose distortion and incomplete shapes, which contains 430 datasets and is only used for non-rigid registration testing. For the two datasets, we use a down-sampling algorithm to select 1024 points as experiment objects.
Rigid dataset. ModelNet40 [43] is an object-level point cloud dataset consisting of 12,311 models from 40 categories. For each category, 80% of models are taken for parameter learning and the rest, 20%, are utilized for validation. For each model, we use a down-sampling algorithm to select 1024 points as experiment objects. Then, the random rotation and translation are performed for each object in the dataset to form the target point cloud datasets, where the rotation is restricted to [0, 45°] and the translation is constrained to [−0.5, 0.5] for X, Y, and Z axes, respectively.
The real sampled data. Self-sampled data is captured from the environment shown in Figure 5 and the red arrows of Figure 5b represent x, y and z axes. The source and target point data are captured from four identical RealSense cameras, which are placed on the four brackets labeled #1, #2, #3, and #4 in Figure 5a for the same scene. To ensure the effective coverage of different-sized objects, the design structure of the camera mount is shown in Figure 5b, which can realize the precise adjustment of the camera in horizontal, vertical, forward, and backward and pitching directions according to different experimental scenes. During the experiment, Hbody01-02 and Fbody01-02 are the sampling data with noise and deformation for non-rigid registration testing, and Face01-02 and Body01-02 are utilized as the representative data for rigid registration validation. The all-sampled data are pre-processed with our previous methodology published in [44].
Since the most complex non-rigid transform object in small indoor scenes is human, the non-rigid experiments focus on human subjects for registration performance. Conversely, the rigid experiments mainly employ object-level targets for evaluation.

5.2. Experiment Evaluation

The comparison methods. In order to validate the performance of NGRLK, some methods are selected for comparison in non-rigid and rigid scenes. CPD (2007) [12] is the benchmarking algorithm and is most widely used in the non-rigid registration field. F3D (2019) [45] is a novel deep neural network for non-rigid registration, which directly learns stream embeddings and hierarchical features from point clouds to represent point motion. As a deformation-to-reconstruction non-rigid registration framework, CorrNet3D (2021) [7] is a novel dense correspondence learning model. RMANet (2021) [6] is an advanced non-rigid registration method based on a series of rigid transformations. Lepard (2022) [5] is an excellent model for non-rigid registration based on a deep learning framework to extract point cloud features from space and position perspectives in the last three years. RPM-Net (2020) [46], as an upgraded version of ICP with soft correspondence and simulated annealing, is an excellent rigid registration method. We compare NGRLK to CPD, F3D, CorrNet3D, RMANet, and Lepard on non-rigid correspondence, public datasets, and real sampling data scenes. Moreover, we compare NGRLK to CorrNet3D, RMANet, and the representative rigid registration method RPMNet on public datasets and real sampling data scenes. During the various experiments, these representative registration methods are re-trained. For our NGRLK, the data batch size is 10, the number of loops will be increased from 1 up to 8, the learning rate is 1 × 10−4, and the empirical values λ1, λ2, and λ3 of the loss function are set to 0.5.
The evaluation metrics. In the non-rigid experimental session, the overall effect of non-rigid registration is measured by analyzing the Corr (%) defined in CorrNet3D with different error tolerances and the earth mover’s distance (EMD) between the aligned and target datasets for different non-rigid registration algorithms. In the rigid phase, the performance of rigid registration is validated by comparing the Corr (%) and the root mean square error (RMSE) values between the aligned and target objects with different registration algorithms.

5.3. Non-Rigid Experiment

For the experiment analysis of non-rigid alignment, CPD, F3D, CorrNet3D, RMANet, and Lepard algorithms are selected for comparison. As the CPD and the RMANet do not manifest correspondences, only F3D, CorrNet3D, Lepard, and NGRLK algorithms compute and record the Corr (%) in Figure 6 and Figure 7. The F3D and Lepard mainly perform point cloud feature representation, which can combine with multiple rigid/non-rigid registration algorithms based on point feature correspondences. Only four algorithms, CPD, CorrNet3D, RMANet, and NGRLK, compute the EMD distance in the training dataset Surreal, which are shown in Figure 8. The experimental visual effects of the non-rigid registration on the test dataset SHREC and real sampled data are shown in Figure 9 and Figure 10, respectively.
Figure 6 shows the Corr (%) result of F3D, CorrNet3D, Lepard, and NGRLK algorithms, which indicates that the performance of CorrNet3D, Lepard, and NGRLK is close to each other and significantly better than that of F3D up to 5% error tolerance. When the error tolerance rate is increased to 15%, the CorrNet3D and the NGRLK still show similar performance; the Lepard gradually highlights the advantage of algorithm performance with a 5% higher Corr (%) value. As the error tolerance is increased to 20%, compared with the CorrNet3D, the Lepard outperforms by about 8%, and NGRLK also shows better performance with a 3% higher Corr (%). Throughout the process, the CorrNet3D consistently outperforms the F3D by 5–10% Corr (%). The correspondence percentage can be effectively improved by enhancing the point feature expression capability or adopting more advanced feature-matching algorithms. As the red line indicates the correspondence between the origin and the target point, Figure 7 presents visually that Lepard and NGRLK produce more accurate corresponding points at the same error tolerance. The F3D represents points by a hierarchical feature extraction method (similar to PointNet++ [47]), and both CorrNet3D and NGRLK characterize points by a DGCNN-like approach. The DGCNN is more capable of representing point features and proves that PointNet++ is a subclass feature extraction process, which corresponds to the F3D having the lowest Corr (%) value. Lepard represents point features by synthesizing feature space, position space, and 3D relative distance, which has a strong feature expression capability. More experiment details for the effect of different feature extraction methods on the Corr (%) value are recorded in the subsequent ablation experiments.
The EMD results of the CPD, CorrNet3D, RMANet, and NGRLK non-rigid registration algorithms for the same object are recorded in Figure 8. The CPD has the largest EMD of 0.07937, which is about 6.5 times higher than the other three algorithms. Compared with the advanced non-rigid alignment algorithms based on unsupervised learning in recent years, the EMD 0.01219 of NRGLK is lower than the EMD 0.01728 of CorrNet3D and EMD 0.01462 of RMANet. The NGRLK proposed in this paper has the smallest registration error. Figure 9 shows the visual registration results of the CPD, CorrNet3D, RMANet, and NGRLK, where the blue represents the target object, the green is the original object and the red shows the registrated object with different methods. From the visual perception, the NGRLK, CorrNet3D, and RMANet can perform non-rigid registration. When dealing with a larger degree of non-rigid deformation, the CPD registration performance decreases, and the NGRLK has the smallest alignment error.
Figure 10 exhibits the alignment results of NGRLK on real data to validate the availability of the algorithm, where Figure 10a denotes the source-target point cloud of self-collected half-body point cloud data, and Figure 10b denotes the non-rigid alignment results. The first set of self-sampled half-body point cloud data has noise and a small range of deformation, and the second set of self-sampled full-body point cloud data has noise, rotational bias, and large deformation. The registration results validate that the NGRLK can perform non-rigid registration on real sampled non-rigid data with good generalization ability and usability.

5.4. Rigid Experiment

For the rigid registration experiment analysis, the RPMNet, CorrNet3D, and RMANet algorithms were selected for comparison. The RPMNet, CorrNet3D, and NGRLK (RMANet does not reflect the correspondence) performed the Corr (%) computation, which is recorded in Figure 11. The registration performance of the RPMNet, CorrNet3D, RMANet, and NGRLK is compared on the ModelNet40 dataset, and the RMSE value is recorded in Figure 11. Moreover, the alignment results of the four registration models on ModelNet40 and real sampled data are shown in Figure 12 and Figure 13, respectively.
Figure 11 records the rigid registration Corr (%) of the RPMNet, CorrNet3D, and NGRLK and indicates good performance. When the error tolerance exceeds 7.5%, the Corr (%) of both CorrNet3D and NGRLK is above 80%, and the Corr (%) of RPMNet is only 40%. While the error tolerance is increased to 15%, the Corr (%) of all three algorithms exceeds 70%, and that of both CorrNet3D and NGRLK in particular exceeds 90%. The NGRLK has the highest Corr (%) at any error tolerance and presents the best performance.
Figure 12 records the numerical analysis of the registration errors for RPMNet, CorrNet3D, RMANet, and NGRLK. As the representative of the rigid registration algorithm, RPMNet has the smallest RMSE of 0.0098. Among CorrNet3D, RMANet, and NGRLK, NGRLK has the smallest RMSE of 0.0153, which is about 56% higher than that of RPMNet. CorrNet3D and RMANet have larger RMSEs of 0.0216 and 0.0164, which are about 2.2 and 1.67 times higher than that of RPMNet, respectively.
Figure 13 demonstrates the rigid registration results of RPMNet, CorrNet3D, RMANet, and NGRLK on the ModelNet40 noisy dataset, which is generated in the same way as our previous research DIRGMR [41]. The intuitive visual comparison shows that all four algorithms can complete the rigid registration on noisy data well. To further validate the usability of the registration model proposed in this paper, Figure 14 summarizes the rigid registration results of the NGRLK on platform-sampled data. The real sampled data are noisy and occluded, and the visual registration verifies that the NGRLK algorithm can complete rigid alignment on the noisy data well.

5.5. Ablation Experiment

In this subsection, we perform sufficient ablation tests on the loss function of L s u r f , L s p a c e , and L C D in NGRLK with different feature extraction modules. By default, the pluggable feature extraction module in NGRLK adopts DGCNN (2019) [40], which is compared to the feature description method adapted with PointNet (2017) [27] and the feature extraction module proposed with non-regular shape factors in DIRGMR (2024) [41], respectively. With other experiment conditions unchanged, Figure 15 records the registration Corr (%) corresponding to the different feature extraction methods. The analysis points out that when the error tolerance is less than 2.5%, the NGRLK is almost unaffected by the expression ability of the point features, and the registration Corr (%) is relatively low (less than 10%). While the error tolerance is increased to 5%, the registration Corr (%) is affected by the feature extraction method. In particular, when the error tolerance rises to 10%, the intuitive visualization shows that the default configuration of the registration Corr (%) improves by about 5% over the NRGLK main framework adapted to the PointNet feature extraction approach. As the error tolerance continues increasing to 20%, the Corr (%) of the NGRLK fused with DIRGMR is consistently higher than that of the default configured NGRLK and the NGRLK main framework adapted to PointNet, with a maximum difference of about 12%. The feature representation fused with the irregular shape factor can effectively improve the accuracy of the registration Corr (%). At the same error tolerance, the registration Corr (%) of the NGRLK with default configuration is also consistently higher than that fused with PointNet by between 3 and 5%, which is presumably because PointNet is not as expressive as DGCNN for point feature extraction.
Then, tests are mainly conducted to analyze the effect of different loss functions on EMD. Specifically, the effects of L s u r f and L s p a c e on the accuracy of the registration results are compared at different loop stages. When the number of cycles is set to 5, 8, and 11, Table 1 records the registration EMD corresponding to L s u r f   +   L s u r f , missing both L s u r f   +   L s u r f , and missing either L s u r f or L s p a c e . Figure 16 compares the effect of loss functions on the accuracy of the registration results through histograms.
Combining Table 1 and Figure 16, the analysis reveals that the EMD decreases from 15.79 to 14.37 when the number of cycles increases from 5 to 8, without L s u r f and L s p a c e . When the number of cycles continuously increases to 11, the EMD only decreases by 0.21. When different loss functions are used, the analysis points out that the EMD values change with a similar pattern. For the records of using L s u r f or L s p a c e alone, the analysis validates that the EMD values change with a similar pattern, which is the small decrease in EMD when the number of loops increases from 5 to 8, and the lower decrease in EMD when the number of loops increases to 11. The above tests illustrate the limited improvement in model training accuracy by the number of loops without adapting an effective loss function. However, when the number of cycles is fixed, both L s u r f and L s p a c e can improve the registration accuracy substantially (EMD values decrease by a factor of 3–30), where L s u r f having a greater impact than L s p a c e for L s u r f imposes constraints on deformations directly, whereas L s p a c e imposes constraints on the relative position of the points in local space that remain unchanged after a single rigid change. In addition, the superposition of L s u r f and L s p a c e can reduce the EMD by nearly 0.7 times, which also indicates that the designed loss function can effectively constrain the learning direction of the NGRLK parameters.

6. Conclusions

Here, we propose an unsupervised non-rigid registration model of dual recurrent computation. Through elastic fine-tuning incremental transformations, a more detailed rigid transformation can be obtained. Combining the attention mechanism and recurrent neural network, the weight of each rigid transformation is updated, and the influence of invalid transformations can be effectively eliminated. In the unsupervised learning framework, a two-dimensional manifold representation is utilized for both the dimensional reduction and reconstruction of point cloud data, complemented by a specifically designed loss function that enforces deformation similarity constraints to guide model training. Moreover, the spatial distribution loss function is designed to make the rigid transformation as rigid as possible. Compared with representative advanced registration models in recent years, NGRLK achieves state-of-the-art performance with 0.01219 earth mover’s distance (EMD) for non-rigid registration and 0.0153 root mean square error (RMSE) for rigid alignment, outperforming approaches it has been compared to by averages of 15% and 50%, respectively, on standard benchmarks. Currently, the proposed NGRLK conducts the experimental scene mainly for indoor object-level and asynchronous interactive scenarios. In the future, research works should explore outdoor and real-time registration performance to expand the application.

Author Contributions

Conceptualization and methodology, M.Y.; investigation, H.T. and X.L.; writing—original draft preparation, M.Y.; writing—review and editing, X.L.; visualization, H.T.; supervision and project administration, X.L. and H.T.; resources, M.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Open Research Project of Zhongke Technology Achievement Transfer and Transformation Center of Henan Province under grant No. 2024142.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, D.; Chen, Y.; Li, J. Remote sensing image registration based on dual-channel neural network and robust point set registration algorithm. In Proceedings of the 2020 5th International Conference on Intelligent Informatics and Biomedical Sciences (ICIIBMS), Okinawa, Japan, 18–20 November 2020; pp. 208–215. [Google Scholar]
  2. Yoo, H.; Choi, A.; Mun, J.H. Acquisition of point cloud in CT image space to improve accuracy of surface registration: Application to neurosurgical navigation system. J. Mech. Sci. Technol. 2020, 34, 2667–2677. [Google Scholar] [CrossRef]
  3. Lu, W.; Zhou, Y.; Wan, G.; Hou, S.; Song, S. L3-net: Towards learning based lidar localization for autonomous driving. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 6382–6391. [Google Scholar]
  4. Wang, M.; Yue, G.; Xiong, J.; Tian, S. Intelligent Point Cloud Processing, Sensing, and Understanding. Sensors 2024, 24, 283. [Google Scholar] [CrossRef]
  5. Li, Y.; Harada, T. Lepard: Learning partial point cloud matching in rigid and deformable scenes. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5544–5554. [Google Scholar]
  6. Feng, W.; Zhang, J.; Cai, H.; Xu, H.; Hou, J.; Bao, H. Recurrent Multi-view Alignment Network for Unsupervised Surface Registration. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 10292–10302. [Google Scholar]
  7. Zeng, Y.; Qian, Y.; Zhu, Z.; Hou, J.; Yuan, H.; He, Y. CorrNet3D: Unsupervised End-to-end Learning of Dense Correspondence for 3D Point Clouds. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6048–6057. [Google Scholar]
  8. Wang, M.; Tian, S. A Brief Introduction to Intelligent Point Cloud Processing, Sensing, and Understanding: Part II. Sensors 2025, 25, 1310. [Google Scholar] [CrossRef]
  9. Newcombe, R.A.; Fox, D.; Seitz, S.M. DynamicFusion: Reconstruction and tracking of non-rigid scenes in real-time. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 343–352. [Google Scholar]
  10. Xu, L.; Su, Z.; Han, L.; Yu, T.; Liu, Y.; Fang, L. UnstructuredFusion: Realtime 4D Geometry and Texture Reconstruction Using Commercial RGBD Cameras. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2508–2522. [Google Scholar] [CrossRef] [PubMed]
  11. Yao, Y.; Deng, B.; Xu, W.; Zhang, J. Quasi-Newton Solver for Robust Non-Rigid Registration. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 7597–7606. [Google Scholar]
  12. Myronenko, A.; Song, X.; Carreira-Perpinan, M. Non-rigid point set registration: Coherent Point Drift. In Advances in Neural Information Processing Systems 19; MIT Press: Cambridge, MA, USA, 2007; pp. 1009–1016. [Google Scholar]
  13. Anguelov, D.; Srinivasan, P.; Pang, H.C.; Koller, D.; Thrun, S.; Davis, J. The correlated correspondence algorithm for unsupervised registration of nonrigid surfaces. In Advances in Neural Information Processing Systems; MIT Press: Vancouver, BC, Canada, 2004; pp. 33–40. [Google Scholar]
  14. Sahillioğlu, Y. A Genetic Isometric Shape Correspondence Algorithm with Adaptive Sampling. ACM Trans. Graph. (TOG) 2018, 37, 1–14. [Google Scholar] [CrossRef]
  15. Ovsjanikov, M.; Merigot, Q.; Memoli, F.; Guibas, L. One Point Isometric Matching with the Heat Kernel. Comput. Graph. Forum 2010, 29, 1555–1564. [Google Scholar] [CrossRef]
  16. Huang, Q.-X.; Adams, B.; Wicke, M.; Guibas, L.J. Non-rigid registration under isometric deformations. Comput. Graph. Forum 2008, 27, 1449–1457. [Google Scholar] [CrossRef]
  17. Rodola, E.; Bronstein, A.M.; Albarelli, A.; Bergamasco, F.; Torsello, A. A game-theoretic approach to deformable shape matching. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 182–189. [Google Scholar]
  18. Rampini, A.; Tallini, I.; Ovsjanikov, M.; Bronstein, A.M.; Rodolà, E. Correspondence-Free Region Localization for Partial Shape Similarity via Hamiltonian Spectrum Alignment. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Quebec City, QC, Canada, 16–19 September 2019; pp. 37–46. [Google Scholar]
  19. Attaiki, S.; Pai, G.; Ovsjanikov, M. DPFM: Deep Partial Functional Maps. In Proceedings of the 2021 International Conference on 3D Vision (3DV), London, UK, 1–3 December 2021; pp. 175–185. [Google Scholar]
  20. Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; Black, M.J. SMPL: A Skinned Multi-Person Linear Model. Acm Trans. Graph. 2015, 34, 248:1–248:16. [Google Scholar] [CrossRef]
  21. Groueix, T.; Fisher, M.; Kim, V.G.; Russell, B.C.; Aubry, M. 3D-CODED: 3D Correspondences by Deep Deformation. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Volume 11206, pp. 235–251. [Google Scholar]
  22. Wang, S.; Geiger, A.; Tang, S. Locally Aware Piecewise Transformation Fields for 3D Human Mesh Registration. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 7635–7644. [Google Scholar]
  23. Wang, L.; Chen, J.; Li, X.; Fang, Y. Non-Rigid Point Set Registration Networks. arXiv 2019, arXiv:1904.01428. [Google Scholar]
  24. Wang, L.; Li, X.; Chen, J.; Fang, Y. Coherent Point Drift Networks: Unsupervised Learning of Non-Rigid Point Set Registration. arXiv 2019, arXiv:1906.03039. [Google Scholar]
  25. Halimi, O.; Litany, O.; Rodola, E.; Bronstein, A.M.; Kimmel, R. Unsupervised Learning of Dense Shape Correspondence. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4365–4374. [Google Scholar]
  26. Bozic, A.; Zollhofer, M.; Theobalt, C.; Nießner, M. DeepDeform: Learning Non-Rigid RGB-D Reconstruction With Semi-Supervised Data. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 7000–7010. [Google Scholar]
  27. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  28. Zhou, W.; Lu, J.; Yue, W. A New Semantic Segmentation Method of Point Cloud Based on PointNet and VoxelNet. In Proceedings of the 2019 Chinese Control and Decision Conference (CCDC), Nanchang, China, 3–5 June 2019; pp. 803–808. [Google Scholar]
  29. Zhao, K.; Lu, H.; Li, Y. PointNetX: Part Segmentation Based on PointNet Promotion. Commun. Comput. Inf. Sci. 2023, 1732, 65–76. [Google Scholar]
  30. Deng, H.; Birdal, T.; Ilic, S. PPFNet: Global Context Aware Local Features for Robust 3D Point Matching. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 195–205. [Google Scholar]
  31. Deng, H.; Birdal, T.; Ilic, S. PPF-FoldNet: Unsupervised Learning of Rotation Invariant 3D Local Descriptors. In European onference on Computer Vision; Springer: Munich, Germany, 2018; Volume 1808.10322, pp. 1–19. [Google Scholar]
  32. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. PointNetLK: Robust & Efficient Point Cloud Registration Using PointNet. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7156–7165. [Google Scholar]
  33. Sarode, V.; Li, X.; Goforth, H.; Aoki, Y.; Srivatsan, R.A.; Lucey, S.; Choset, H. Pcrnet: Point cloud registration network using pointnet encoding. arXiv 2019, arXiv:1908.07906. [Google Scholar]
  34. Sharp, N.; Attaiki, S.; Crane, K.; Ovsjanikov, M. DiffusionNet: Discretization Agnostic Learning on Surfaces. ACM Trans. Graph. (TOG) 2020, 41, 1–16. [Google Scholar] [CrossRef]
  35. Shimada, S.; Golyanik, V.; Tretschk, E.; Stricker, D.; Theobalt, C. DispVoxNets: Non-Rigid Point Set Alignment with Supervised Learning Proxies. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Québec City, QC, Canada, 16–19 September 2019; pp. 27–36. [Google Scholar]
  36. Trappolini, G.; Cosmo, L.; Moschella, L.; Marin, R.; Melzi, S.; Rodolà, E. Shape registration in the time of transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 5731–5744. [Google Scholar]
  37. Boscaini, D.; Masci, J.; Rodolà, E.; Bronstein, M. Learning shape correspondence with anisotropic convolutional neural networks. In 30th Conference on Neural Information Processing Systems (NIPS 2016); MIT Press: Barcelona, Spain, 2016; Volume 29, pp. 1–9. [Google Scholar]
  38. Wang, L.; Zhou, N.; Huang, H.; Wang, J.; Li, X.; Fang, Y. GP-Aligner: Unsupervised Groupwise Nonrigid Point Set Registration Based on Optimizable Group Latent Descriptor. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  39. Baker, S.; Matthews, I. Lucas-Kanade 20 Years On: A Unifying Framework. Int. J. Comput. Vis. 2004, 56, 221–255. [Google Scholar] [CrossRef]
  40. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph CNN for learning on point clouds. ACM Trans. Graph. (TOG) 2019, 38, 1–12. [Google Scholar] [CrossRef]
  41. Yuan, M.; Li, X. Point cloud registration method for indoor depth sensor acquisition system based on dual graph computation with irregular shape factors. IET Image Process. 2024, 18, 2161–2178. [Google Scholar] [CrossRef]
  42. Vakalopoulou, M.; Chassagnon, G.; Bus, N.; Marini, R.; Zacharaki, E.I.; Revel, M.P.; Paragios, N. Atlasnet: Multi-atlas non-linear deep networks for medical image segmentation. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, 16–20 September 2018; Springer: Granada, Spain, 2018; pp. 658–666. [Google Scholar]
  43. Puy, G.; Boulch, A.; Marlet, R. Flot: Scene flow on point clouds guided by optimal transport. In European Conference on Computer Vision; Springer: Glasgow, UK, 2020; pp. 527–544. [Google Scholar]
  44. Yuan, M.; Li, X.; Xu, J.; Jia, C.; Li, X. 3D foot scanning using multiple RealSense cameras. Multimed. Tools Appl. 2021, 80, 22773–22793. [Google Scholar] [CrossRef]
  45. Liu, X.; Qi, C.R.; Guibas, L.J. Flownet3d: Learning scene flow in 3d point clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 529–537. [Google Scholar]
  46. Yew, Z.J.; Lee, G.H. RPM-Net: Robust Point Matching Using Learned Features. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11821–11830. [Google Scholar]
  47. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 14. [Google Scholar]
Figure 1. The main framework of NGRLK.
Figure 1. The main framework of NGRLK.
Sensors 25 03525 g001
Figure 2. The computation and update of each unit.
Figure 2. The computation and update of each unit.
Sensors 25 03525 g002
Figure 3. The deformation and reconstruction process from S to T .
Figure 3. The deformation and reconstruction process from S to T .
Sensors 25 03525 g003
Figure 4. The local distribution of a point in point cloud.
Figure 4. The local distribution of a point in point cloud.
Sensors 25 03525 g004
Figure 5. The self-sampled data system. (a) The real data sampling platform; (b) the design structure sampled camera.
Figure 5. The self-sampled data system. (a) The real data sampling platform; (b) the design structure sampled camera.
Sensors 25 03525 g005
Figure 6. The Corr% of F3D (2019) [45], CorrNet3D (2021) [7], Lepard (2022) [5], and NGRLK.
Figure 6. The Corr% of F3D (2019) [45], CorrNet3D (2021) [7], Lepard (2022) [5], and NGRLK.
Sensors 25 03525 g006
Figure 7. The point correspondences estimation of F3D (2019) [45], CorrNet3D (2021) [7], Lepard (2022) [5], and NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the correspondence estimation of F3D; (c) the correspondence estimation of CorrNet3D; (d) the correspondence estimation of Lepard; (e) the correspondence estimation of NGRLK.
Figure 7. The point correspondences estimation of F3D (2019) [45], CorrNet3D (2021) [7], Lepard (2022) [5], and NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the correspondence estimation of F3D; (c) the correspondence estimation of CorrNet3D; (d) the correspondence estimation of Lepard; (e) the correspondence estimation of NGRLK.
Sensors 25 03525 g007
Figure 8. The EMD of CPD (2007) [12], CorrNet3D (2021) [7], RPMNet (2020) [46], and NGRLK.
Figure 8. The EMD of CPD (2007) [12], CorrNet3D (2021) [7], RPMNet (2020) [46], and NGRLK.
Sensors 25 03525 g008
Figure 9. The registration results of CPD (2007) [12], CorrNet3D (2021) [7], RPMNet (2020) [46], and NGRLK on the training dataset. (a) The source point cloud and target point cloud (ground truth); (b) the registration results of CPD; (c) the registration results of CorrNet3D; (d) the registration results of RMANet; (e) the registration results of NGRLK.
Figure 9. The registration results of CPD (2007) [12], CorrNet3D (2021) [7], RPMNet (2020) [46], and NGRLK on the training dataset. (a) The source point cloud and target point cloud (ground truth); (b) the registration results of CPD; (c) the registration results of CorrNet3D; (d) the registration results of RMANet; (e) the registration results of NGRLK.
Sensors 25 03525 g009
Figure 10. The non-rigid registration results of NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the non-rigid registration results of NGRLK on training data.
Figure 10. The non-rigid registration results of NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the non-rigid registration results of NGRLK on training data.
Sensors 25 03525 g010
Figure 11. The Corr% of RPMNet (2020) [46], CorrNet3D (2021) [7], and NGRLK.
Figure 11. The Corr% of RPMNet (2020) [46], CorrNet3D (2021) [7], and NGRLK.
Sensors 25 03525 g011
Figure 12. The RMSE of RPMNet (2020) [46], CorrNet3D (2021) [7], RMANet (2021) [6], and NGRLK.
Figure 12. The RMSE of RPMNet (2020) [46], CorrNet3D (2021) [7], RMANet (2021) [6], and NGRLK.
Sensors 25 03525 g012
Figure 13. The registration results of RPMNet (2020) [46], CorrNet3D (2021) [7], RMANet (2021) [6], and NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the registration results of RPMNet; (c) the registration results of CorrNet3D; (d) the registration results of RMANet; (e) the registration results of NGRLK.
Figure 13. The registration results of RPMNet (2020) [46], CorrNet3D (2021) [7], RMANet (2021) [6], and NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the registration results of RPMNet; (c) the registration results of CorrNet3D; (d) the registration results of RMANet; (e) the registration results of NGRLK.
Sensors 25 03525 g013
Figure 14. The registration results of NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the registration results of NGRLK on real sampled data.
Figure 14. The registration results of NGRLK. (a) The source point cloud and target point cloud (ground truth); (b) the registration results of NGRLK on real sampled data.
Sensors 25 03525 g014
Figure 15. The impact of DGCNN (2019) [40], PointNet (2017) [27], and DIRGMR (2024) [41] feature extraction on Corr (%).
Figure 15. The impact of DGCNN (2019) [40], PointNet (2017) [27], and DIRGMR (2024) [41] feature extraction on Corr (%).
Sensors 25 03525 g015
Figure 16. The loss function effect on registration accuracy.
Figure 16. The loss function effect on registration accuracy.
Sensors 25 03525 g016
Table 1. The registration accuracy comparison results of L s u r f and L s p a c e .
Table 1. The registration accuracy comparison results of L s u r f and L s p a c e .
LossRecurrent NumberEMD
-515.79
-814.37
-1114.16
L s u r f 51.26
L s p a c e 54.37
L s u r f + L s p a c e 50.924
L s u r f 80.943
L s p a c e 84.21
L s u r f + L s p a c e 80.505
L s u r f 110.732
L s p a c e 113.879
L s u r f + L s p a c e 110.503
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, M.; Li, X.; Tan, H. An Elastic Fine-Tuning Dual Recurrent Framework for Non-Rigid Point Cloud Registration. Sensors 2025, 25, 3525. https://doi.org/10.3390/s25113525

AMA Style

Yuan M, Li X, Tan H. An Elastic Fine-Tuning Dual Recurrent Framework for Non-Rigid Point Cloud Registration. Sensors. 2025; 25(11):3525. https://doi.org/10.3390/s25113525

Chicago/Turabian Style

Yuan, Munan, Xiru Li, and Haibao Tan. 2025. "An Elastic Fine-Tuning Dual Recurrent Framework for Non-Rigid Point Cloud Registration" Sensors 25, no. 11: 3525. https://doi.org/10.3390/s25113525

APA Style

Yuan, M., Li, X., & Tan, H. (2025). An Elastic Fine-Tuning Dual Recurrent Framework for Non-Rigid Point Cloud Registration. Sensors, 25(11), 3525. https://doi.org/10.3390/s25113525

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop