Next Article in Journal
A Systematic Literature Review of Handheld Augmented Reality Solutions for People with Disabilities
Previous Article in Journal
Assessment of the Performance of a Portable, Low-Cost and Open-Source Device for Luminance Mapping through a DIY Approach for Massive Application from a Human-Centred Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Feature Matching for 3D Point Clouds with Progressive Consistency Voting

School of Electronics and Control Engineering, Chang’an University, Xi’an 710064, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(20), 7718; https://doi.org/10.3390/s22207718
Submission received: 4 September 2022 / Revised: 30 September 2022 / Accepted: 4 October 2022 / Published: 11 October 2022
(This article belongs to the Section Optical Sensors)

Abstract

:
Feature matching for 3D point clouds is a fundamental yet challenging problem in remote sensing and 3D computer vision. However, due to a number of nuisances, the initial feature correspondences generated by matching local keypoint descriptors may contain many outliers (incorrect correspondences). To remove outliers, this paper presents a robust method called progressive consistency voting (PCV). PCV aims at assigning a reliable confidence score to each correspondence such that reasonable correspondences can be achieved by simply finding top-scored ones. To compute the confidence score, we suggest fully utilizing the geometric consistency cue between correspondences and propose a voting-based scheme. In addition, we progressively mine convincing voters from the initial correspondence set and optimize the scoring result by considering top-scored correspondences at the last iteration. Experiments on several standard datasets verify that PCV outperforms five state-of-the-art methods under almost all tested conditions and is robust to noise, data decimation, clutter, occlusion, and data modality change. We also apply PCV to point cloud registration and show that it can significantly improve the registration performance.

1. Introduction

Feature matching for 3D point clouds is a fundamental problem in remote sensing and 3D computer vision with a number of applications such as point cloud registration [1,2,3,4,5], 3D object recognition [6], and localization [7]. Its objective is to establish reasonable point-to-point correspondences between two given point clouds (this paper focuses on 3D rigid point clouds). Usually, such correspondences are a subset of initial feature correspondences with outliers (incorrect correspondences) generated by matching local keypoint descriptors [8,9,10]. Therefore, the key issue to robust feature matching for point clouds is how to distinguish inliers from outliers. This task is particularly challenging when the initial correspondence set is contaminated by severe outliers [11,12], which is generally caused by the following factors. (1) Keypoint detection errors. The repeatability of existing 3D keypoint detectors [13] is limited to real-world data, especially for data captured by low-cost sensors. In this case, some truly corresponding points cannot be discovered. (2) Limited descriptiveness of 3D local descriptors. The premise for correct correspondences is that the similarity of two local surface patches can be reasonably estimated, which is accomplished by 3D local descriptors. However, most of the existing 3D local descriptors show limited descriptiveness in the presence of partial overlap, clutter, and occlusion [14]. (3) Data nuisances. Due to the limitations of current 3D acquisition systems, the raw captured point clouds are usually quite noisy. Variation of the distance from the sensor to the object/scene will further result in varying data resolutions. In addition, complex scenes with self-occlusion, occlusion, and clutter may lead to point clouds with incomplete or redundant data. These data nuisances further possess great challenges to robust point cloud correspondences.
To overcome the above difficulties, many attempts have been done by researchers in the past decade. Some methods select candidates from the initial correspondence set solely using the feature similarity cue, e.g., correspondence selection based on the nearest neighbor (NN) [15,16] or the nearest neighbor similarity ratio (NNSR) [17,18]. Although simple and fast, methods relying on feature similarity are very sensitive to outliers because the distinctiveness of local descriptors cannot be guaranteed on challenging data. The popular estimator random sample consensus (RANSAC) [19] can also be employed for outlier rejection [2,20] in the context of point cloud feature correspondences. Nonetheless, it is not appropriate to directly apply RANSAC or its variants to find consistent correspondences from the initial set because these estimators usually require a huge amount of repetitions and may still fail to find the optimal solution. A more reasonable way, as suggested by many studies [21,22], is to first score feature correspondences and then rank them to screen out top-scored correspondences as the final feature matching result. Some advances have been made along this research line, e.g., the spectral technique (ST) [23] and search of inliers (SI) [21]. Note that NN and NNSR can be also categorized as correspondence scoring techniques since the similarity of two keypoints can be served as the confidence score as well. Unfortunately, as revealed by a recent evaluation study [11], the performance of existing scoring methods is still limited when faced with severe outliers.
Under these considerations, we propose a novel scoring technique named progressive consistency voting (PCV) to achieve robust feature matching for 3D point clouds. Given a set of initial correspondences generated by matching local keypoint descriptors, PCV assigns a confidence score to each correspondence. Accordingly, a subset of high-quality correspondences, i.e., the expected output of feature matching, can be screened out by simply considering top-scored candidates. The basic idea of PCV is finding a reliable voting set from the initial set and aggregating the geometric consistencies between the correspondence to be scored and these candidates in the voting set under a voting scheme. Although some voting-based scoring methods already exist [21,22], they underestimate the significance of the voting set, which is critical to achieve robust feature matching, as will be verified in Section 4.4. By contrast, the proposed PCV, to the best of our knowledge, is the first method that highlights the issue of mining reliable voters and optimizes the voting set in a progressive manner. Moreover, unlike previous voting-based methods needing special care for the design of geometric constraints, PCV manages to achieve decent performance using very simple geometric constraints. Experiments on several public datasets verify that PCV is robust to a number of nuisances. A comparison with six state-of-the-art methods further confirms the overall superiority of the proposed method. In a nutshell, this paper presents three main contributions.
  • We propose a feature matching method PCV for 3D point clouds, which progressively optimizes a voting set and the confidence scores of correspondences based on the geometric consistencies between correspondences. Experiments and comparison with state-of-the-art methods demonstrate that PCV is robust to noise, data decimation, clutter, occlusion, and data modality change. Moreover, PCV can achieve outstanding performance with very simple geometric constraints.
  • PCV highlights and demonstrates the significance of the voting set to robust 3D feature matching for voting-based methods, which may inspire other related methods.
  • We show that PCV can be successfully applied to point cloud registration and improve the registration performance under existing pipelines.
The remainder of this paper is organized as follows. Section 2 presents a brief overview of existing 3D feature matching materials and methods. Section 3 elaborates the proposed PCV method. Section 4 presents the experimental results on several standard datasets and the comparative results with some state-of-the-art methods. Finally, Section 5 draws the conclusions.

2. Related Works

Existing feature matching methods for point clouds tend to assign confidence scores to feature correspondences. There also exist some methods that assign binary labels (i.e., true or false) to correspondences.

2.1. Methods Computing Binary Labels

RANSAC [19] distinguishes inliers from outliers based on a “hypothesis generation and verification” mechanism; three or more correspondences are first randomly sampled from the initial correspondence set; a hypothesis is then generated with these samples and candidates agreeing the hypothesis are judged as inliers at this iteration; the above processes are repeated, and the inlier set with the maximum cardinality is served as the final result. Some variants of RANSAC have also been proposed in 3D domain, e.g., the optimal sample consensus (OSAC) [15], 1-point RANSAC (1P-RANSAC) [6], RANSAC with different inlier metrics [24]. A recent comprehensive evaluation on RANSAC methods for 3D correspondence grouping and registration can be referred to [25]. However, the RANSAC family generally holds limited scalability and can hardly find the optimal solution when confronted with massive outliers [11]. Chen and Bhanu [26] grouped correspondences with a geometric consistency (GC) framework; they measured the consistency score between any two correspondences and formed a cluster for each correspondence with its compatible ones; the maximum cluster is treated as the inlier cluster. Tombari and Stefano [27] presented a 3D Hough voting (3DHV) approach that projects each correspondence to a 3D Hough space using the local reference frame (LRF) of each keypoint; projected points forming a cluster in the Hough space are identified as inliers. Ma et al. proposed robust hypothesize-and-verify methods that rely on vector field consensus (VFC) [28] and robust point matching via L 2 E [29], respectively. Lu et al. [30] employed redundant geometric constraints to determine the best transformation and correspondences agreeing with the best transformation are grouped as inliers. More recently, Ma et al. [31] and Zhao et al. [32] cast the feature matching problem into a two-class classification problem and learned a general classifier to determine the correctness of a correspondence. In order to maintain the local neighborhood structures of potential inliers, Ma et al. [33] proposed a locality preserving matching (LPM) model to identify correct correspondences.
The above methods are either parametric (e.g., RANSAC) or non-parametric (e.g., GC and 3DHV). Parametric methods cannot scale well to large-scale feature matching problems and are sensitive to outliers; the precision performance of many non-parametric methods remains limited [11].

2.2. Methods Computing Confidence Scores

A straightforward correspondence scoring approach is serving the feature similarity between the two associated keypoints of a correspondence as the confidence score [15,34]. Typically, L 2 distance is employed as the distance metric. Rusu et al. [35] and Yang et al. [16] investigated the effect when using other different distance metrics on point cloud feature matching. Because local descriptors are sensitive to repetitive patterns, Lowe [17] presented a nearest neighbor similarity ratio (NNSR) method to prefer distinctive correspondences. Unfortunately, purely leveraging the feature similarity cue inherits the sensitivity of the employed local descriptor to some certain nuisances such as clutter and occlusion.
Compared with feature similarity, geometric constraints are more robust cues. Leordeanu and Hebert [23] introduced a spectral technique (ST) to recover inliers from initial correspondences through analyzing the adjacency matrix of a graph with each node being a correspondence and each edge being the pairwise consistency of two nodes. Rodolà et al. [36] selected correspondences that satisfy global geometric consistency constraints with a game theory matching (GTM) model. Cirujeda et al. [37] presented a variant of GTM by adding feature constraints to the pay-off function. A limitation of GTM-based methods is that they are highly selective, thus failing to strike a good balance between precision and recall. Buch et al. [21] proposed finding inliers by local and global voting; the local voting stage checks the geometric consistency of a correspondence with its neighbors via a distance constraint; the global stage judges the global geometric consistency using an LRF constraint. Following the voting scheme, Yang et al. [22] proposed a consistency voting (CV) approach that treats distinctive correspondences as the voting set and assign voting scores to correspondences based on the rigidity and LRF constraints. Note that although the proposed PCV method also follows a voting fashion like SI and CV, it holds some unique characteristics. (1) SI and CV ignore the significance of the definition of voting set, and they regard either spatially close or distinctive correspondences as voters. This cannot guarantee the quality of the voting set and fails to assign reasonable voting scores to correspondence as will be verified in Section 4.4. By contrast, our PCV highlights the significance of the voting set and progressively optimizes the voting set. (2) Unlike SI and CV needing carefully designed geometric constraints, the proposed PCV method is able to achieve the state-of-the-art performance with a simple distance constraint.

3. Methods

This section elaborates the technique details of the proposed PCV method (Figure 1). PCV iteratively refines the confidence score assigned to each correspondence. Each iteration contains three main parts, i.e., definition/update of the voting set, pairwise compatibility assessment, and voting score calculation.

3.1. Notations and Problem Formulation

Let P s , P t , and C i n i t i a l be the source point cloud, target point cloud, and the initial correspondence set between them, respectively. C i n i t i a l is usually generally obtained by matching local keypoint geometric descriptors such as the signature of histograms of orientation (SHOT) [38] and local feature statistics histograms (LFSH) [15]. A candidate in C i n i t i a l can be parametrized as c = ( p s ,   p t ,   s f ( p s ,   p t ) ) , where p s and p t , respectively, are two keypoints in P s and P t , and s f ( p s ,   p t ) is the feature matching score of p s and p t . By default, we use Lowe’s ratio technique [17] to assign s f ( p s ,   p t ) to c.
The objective is to assign a confidence score s ( c ) to each correspondence. Accordingly, an inlier set C i n l i e r can be efficiently calculated by finding top-scored candidates in C i n i t i a l .

3.2. Pairwise Compatibility Assessment

The proposed method iteratively checks the compatibility of each correspondence in C i n i t i a l and all candidates in a predefined voting set C v o t i n g . The motivation behind is that, if a correspondence is correct, it will receive a high voting score if most of the candidates in the voting set are correct as well, because only correct correspondences are geometrically compatible with each other [11]. Therefore, the calculation of the voting score relies on the assessment of the pairwise compatibility of two correspondences. The definition and update of the voting set C v o t i n g will be detailed in Section 3.4, and in the following we assume that C v o t i n g has been given.
Now, the first question is how to measure the compatibility of two correspondences c 1 and c 2 . In the literature, many effective geometric constraints have been proposed for the assessment of correspondence compatibility, where L 2 distance [21,36], normal [39,40], and LRF [21,22] are three typical ones. To better understand and compare the three constraints, we give a schematic illustration for them in Figure 2.
The rigidity constraint (Figure 2a) is defined as:
E L 2 = | p 1 s p 2 s p 1 t p 2 t | .
The normal constraint (Figure 2b) is defined as:
E n o r m a l = acos ( n 1 s · n 2 s ) acos ( n 1 t · n 2 t ) .
where n is the normal vector of p . Let L be the LRF at p . LRF is a local canonical frame established at keypoint p to make the local keypoint descriptor rotation invariant. An LRF can be mathematically represented by a 3 × 3 matrix constituted by three orthogonal unit vectors. The LRF constraint (Figure 2c) is defined as:
E l r f = l d i f f s l d i f f t ,
where
l d i f f = acos trace ( L 1 L 2 1 ) 1 2 180 π .
Most of existing methods [21,22,40] make great efforts to either carefully combine above geometric constraints or devise new ones. By contrast, our method, with progressively optimized voters, can achieve decent performance with the simple L 2 distance constraint, which is quite efficient for calculation (as will be verified in Section 4.4). Note that the L 2 distance constraint only needs keypoint locations, which are already provided at the keypoint detection stage, while other constraints such as normal and LRF need additional computational costs as well as efforts on tuning associated parameters. Therefore, we define the compatibility score F ( c 1 ,   c 2 ) between c 1 and c 2 as:
F ( c 1 ,   c 2 ) = exp E L 2 2 2 τ 2 ,
where τ is a distance parameter. Based on empirical tests, τ is set to 10 pr. The unit pr, here and hereafter, denotes the point cloud resolution, i.e., the average shortest distance between each point and its neighbors in a point cloud, which can be mathematically defined as:
p r = 1 | P | p i P | | p i p n n | | ,
where p i represents a generic point in point cloud P and p n n is the nearest neighbor of p i .

3.3. Voting Score Calculation

At each iteration of PCV, e.g., the ith iteration, a voting score s i ( c ) will be assigned to each correspondence c. s i ( c ) is the integration of the compatibility scores of c with each candidate in the voting set C v o t i n g i at the current iteration. More specifically, s i ( c ) is defined as:
s i ( c ) = c j C v o t i n g i F ( c ,   c j ) .
From the above equation, one can find that the voting score is independent from feature similarities that are sensitive to clutter, occlusion, and limited overlap. We repeat the voting score calculation process for N i t e r times and serve the voting scores at the last iteration as the final scoring result.

3.4. Definition and Update of the Voting Set

The definition and update of the voting set is a peculiarity of the proposed PCV method and acts as a critical component. The basic goal is to find a subset from C i n i t i a l with a high inlier ratio as the voting set C v o t i n g , so the voters are more convincing and the voting scores are more distinguishable. However, the label of each correspondence is unknown and in fact it is the objective of feature matching. To solve this problem, we propose the following method to progressively optimize the voting set.
At the first iteration of PCV, we define the voting set as the collection of correspondences with top s f ( p s ,   p t ) values, denoted by C v o t i n g 0 . Remarkably, the definition C v o t i n g 0 has a relatively slight impact on the performance of PCV. PCV is even able to achieve good performance with C v o t i n g 0 composed of randomly selected correspondences (as will be verified in Section 4.4).
Because s f ( p s ,   p t ) is sensitive to clutter, occlusion, and limited overlap, C v o t i n g 0 may contain many outliers. To optimize C v o t i n g 0 , we suggest updating C v o t i n g based on the scoring result of the previous iteration. This is reasonable because [22] has verified that the confidence scores of correspondences can be optimized within the consistency voting framework. Moreover, we experimentally observed that the quality of the correspondence scoring result generally improves during iterations (as will be shown in Section 4.4). As such, C v o t i n g i at the ith iteration of PCV can be updated as follows:
C v o t i n g i = c j C i n i t i a l i 1 j = 1 κ ,
where C i n i t i a l i 1 is the sorted C i n i t i a l according to the scoring result (Equation (7)) in a descending order at the previous iteration and κ is the cardinality of C v o t i n g i .
The parameter κ requires special attention, because the objective is to bring as many inliers as possible in C v o t i n g i while rejecting outliers. If we fix κ , small values of κ will fail to include all potential inliers and large values of κ will drag in outliers if the number of inliers in C i n i t i a l is smaller than κ . To overcome this problem, we propose a method to make κ adaptive. Specifically, we apply Otsu’s thresholding method [41] on C i n i t i a l i 1 to obtain an adaptive threshold τ o t s u . Then, we define κ as the number of correspondences whose voting scores are greater than τ o t s u . The reason behind is that the voting scores calculated using geometric constraints hold certain discriminatory power, which is also demonstrated in Figure 3.
Therefore, the voting set in PCV on one hand is progressively optimized using the scoring result at the previous iteration and on the other has an adaptive cardinality to cope with initial correspondence sets with different scales and inlier ratios.

3.5. Method Summarization

Algorithm 1 summarizes the key steps of the proposed PCV method. Remarkably, once voting scores have been assigned to correspondences by PCV, we can define inliers in two ways. One is setting a threshold based on Otsu’s thresholding method [41] to split the correspondence set by comparing voting scores with the threshold. The other is simply serving K of top-scored correspondences as inliers, where K is an empirical threshold depending on application scenarios.
Algorithm 1 Feature matching for point clouds based on PCV.
Require: The source point cloud P s and the target point cloud P t .
Ensure: Consistent feature correspondences C i n l i e r between P s and P t .
 1: Generate C i n i t i a l by matching local keypoint descriptors;
 2: Initialize the iteration index i = 0 and C i n l i e r = Ø ;
 3: Define the initial voting set C v o t i n g 0 based on s f ( p s ,   p t ) ;
 4: while   i < N i t e r   do
 5:    Calculate the compatibility score between c C i n i t i a l and c j C v o t i n g i 1 using Equation (5);
 6:    Calculate the voting score for c C i n i t i a l using Equation (7);
 7:    Update C v o t i n g i using Equation (8);
 8:     i = i + 1 ;
 9: end while
 10: Sort C i n i t i a l based on the scoring result at the last iteration;
 11: Push top-scored correspondences into C i n l i e r ;
 12: Return C i n l i e r .

4. Experiments

4.1. Datasets

The experimental datasets include the Bologna Dataset1 (BoD1) dataset [42], the UWA 3D object recognition (U3OR) dataset [43], and the Bologna Dataset5 (BoD5) dataset [42]. BoD1 is a synthetic recognition dataset consisting of six models and 45 scenes, where the scenes were generated by randomly rotating three to five models to create clutter and pose variations. U3OR is a renowned benchmark in 3D computer vision [14]. It has five models and 50 real scenes. This dataset also provides the quantized information about clutter and occlusion. Both BoD1 and U3OR are composed of LiDAR-scanned point clouds, while BoD5 is obtained via a Kinect sensor. BoD5 contains 43 matching pairs with severe real noise. Some sample views of the point clouds in these datasets are presented in Figure 4.

4.2. Criterion

Following [22,44], we employ the recall of inlier (ROI) metric to evaluate the quality of a point cloud feature matching method. The definition is given as follows.
Let C i n l i e r K be the set of K top-scored correspondences computed by a correspondence scoring method, the recall respecting a given K is defined as:
r e c a l l K = # inliers in C i n l i e r K # inliers in C i n i t i a l .
By varying K, a curve can be generated. To judge if a correspondence c = ( p s ,   p t ) is an inlier, we use the ground truth rotation matrix R g t and translation vector t g t and check if R g t p s + t g t p t is smaller than a distance threshold τ i n l i e r . We set τ i n l i e r to 5 pr in the experiments.

4.3. Compared Methods and Implementation Details

Five methods are considered for comparative evaluation, including the nearest neighbor (NN) [2,15], nearest neighbor similarity ratio (NNSR) [17], spectral technique (ST) [23], search of inliers (SI), and consistency voting (CV) [22]. NN and NNSR are served as two baselines purely rely on the feature similarity cue; ST is a well-known graph-based method for feature matching; SI and CV are two recent methods based on the voting scheme as well. The parameters of all compared methods are kept identical to the settings in the original papers.
To generate the initial correspondence set C i n i t i a l , we detect keypoints on point clouds via the Harris 3D [45] detector and use the SHOT [38] descriptor to perform local feature description. Then, we perform brute force feature matching based on the L 2 distance metric to generate C i n i t i a l . Note that the we are not aiming at finding the best combination of detector and descriptor to generate high quality C i n i t i a l . Instead, we simply choose a combination of popular detectors and descriptors to obtain initial correspondences with various qualities when confronted with different challenges. This is necessary for a comprehensive evaluation of feature matching methods.

4.4. Method Analysis

Before evaluating PCV’s overall performance, we are amenable to validate the effectiveness of some key components of it, to make PCV explainable. Some key parameters also need to be experimentally analyzed. We conduct these experiments on a “tuning” BoD1 dataset where the scenes are first down-sampled to 1 4 of the original resolution and then injected with 0.3 pr Gaussian noise.
  • The significance of high-quality voting set. To demonstrate the significance of the voting set to the final feature matching performance, we generate voting sets with different inlier ratios, i.e., 0.1 to 1.0 with a gap of 0.1, and perform consistency voting to compute voting scores. Figure 5a presents the ROI performance ( K = 100 ) when using voting sets with various inlier ratios. Clearly, one can find that more inliers in the voting set result in better feature matching performance. This validates the rationality of our motivation to pursue high-quality voters.
  • How many iterations are required? PCV progressively optimizes the voting set and voting scores. Since more iterations would result in more time consumption, we should determine a proper N i t e r to achieve a trade-off between time efficiency and ROI performance. Figure 5b reports the efficiency and ROI results ( K = 100 ) when varying N i t e r from 1 to 8. It suggests that the ROI performance generally improves as N i t e r increases. This also indicates the effectiveness of the proposed progressive voting framework because more iterations achieve better performance. When N i t e r is greater than 2, the performance remains almost stable. Thus, we set N i t e r to 3 in this paper.
  • The definition of C v o t i n g 0 . By default, we use Lowe’s ratio rule [17] (denoted by NNSR) to determine the initial voting set C v o t i n g 0 . In this experiment, initial voting sets determined by selecting correspondences with top-ranked feature similarity values (denoted by NN) and random selection in C i n i t i a l are also taken into consideration. Table 1 presents the results. It suggests that NNSR-based C v o t i n g 0 achieves the best performance, possibly owing to the fact that more inliers are included in C v o t i n g 0 when sorting C i n i t i a l based on NNSR. However, using randomly selected correspondences as C v o t i n g 0 also achieves promising result. We can deduce that PCV is able to effectively refine C v o t i n g in the subsequent iterations even with poor initializations.
  • The cardinality of C v o t i n g 0 . Although we have proposed an approach to make the cardinality C v o t i n g adaptive during the iterations, the cardinality of the initial voting set C v o t i n g 0 should be pre-defined. To examine its influence, we vary C v o t i n g 0 from 20 to 1000 and compute the ROI results. As shown in Figure 5c, we can see that the ROI performance ( K = 100 ) with respect to different values of C v o t i n g 0 is generally stable. The only salient improvement can be found when C v o t i n g 0 increases from 20 to 100. It is possibly due to the fact that few inliers are included with small | C v o t i n g 0 | . Because PCV updates | C v o t i n g i | adaptively at the subsequent iterations (i.e., i > 0 ), PCV is not very sensitive to the parameter | C v o t i n g 0 | as long as it is not extremely small.
  • The effectiveness of making κ adaptive. During the iterations of PCV, we make the cardinality of the voting set κ (Equation (8)) adaptive, because fixing it can hardly guarantee the inlier ratio of the voting set. For instance, if there are 100 inliers in the initial correspondence set and we fix κ to 200, the highest inlier ratio of the voting set is 0.5. However, we have already shown that voting sets with higher inlier ratios can achieve better performance in Figure 5a. Therefore, we compare the performance of the standard PCV and PCV without adaptive voting sets (denoted by “PCV w/o AV”). The result is shown in Figure 5d. Obviously, better performance is achieved with adaptive voting sets.
  • The selection of geometric constraints. Geometric constraints are employed to measure the compatibility score between correspondences. We examine the performance of PCV when using different geometric constraints (the combinations of constraints in Equations (1)–(3)). Specifically, the L 2 distance, L 2 distance+normal, and L 2 distance+LRF constraints have been employed in existing studies [21,22,36,40]. We also consider the L 2 distance+normal+LRF constraint in this experiment, generating a total of four different geometric constraints. The results are reported in Table 2.
It shows that PCV is not sensitive to the selection of geometric constraints, since almost comparable performance is achieved when using tested constraints. Moreover, it achieves outstanding performance with the simple L 2 distance constraint. Note that other constraints such as normal and LRF need additional time consumption as well as efforts for parameter tuning. By contrast, PCV allows the use of a simple distance constraint that only requires the coordinate information of keypoints. It is potentially because PCV can mine high-quality voters and thus does not have strict demands on the selection of geometric constraints. This makes PCV quite flexible. In particular, normals and LRFs are usually computed by feature descriptors, and the time costs of employing the above different constraints are identical.

4.5. Feature Matching Performance

The feature matching performance of PCV and the compared methods is evaluated under various challenging conditions, including Gaussian noise, data decimation, clutter, occlusion, and data modality change.
  • Robustness w.r.t. Gaussian noise. We add Gaussian noise with 0.1 pr, 0.3 pr, and 0.5 pr standard deviations to the scenes of the BoD1 dataset, respectively. The results are shown in Figure 6a–c.
With 0.1 pr Gaussian noise, all tested methods generally behave comparably and our PCV is slightly better than others. As the standard deviation of Gaussian noise increases, PCV achieves the best performance and the gap between it and other compared methods is more clear. Besides, NN and NNSR are inferior to others, verifying our statement that methods simply relying on the feature similarity are not robust to some common nuisances. CV, ST, and SI deliver similar performance in this case.
  • Robustness with reference to data decimation. To assess a method’s robustness to data decimation, we down-sample the scenes of the BoD1 dataset to 1 2 , 1 4 , and 1 8 of their original resolutions. The results are presented in Figure 6d–f.
Compared with Gaussian noise, data decimation has a more severe impact on feature matching. Even though, PCV achieves the best performance under all levels of data decimation. The CV method that ranks the second with 1 2 data decimation meets a clear performance deterioration as the point clouds are further down-sampled. This is due to two factors. First, CV relies on the LRF constraint, and the calculation of LRF is sensitive to data decimation [38], thus affecting the calculation of compatibility scores. Second, CV simply defines the voting set based on feature similarities, as the matching case becomes challenging, the voting set contains less inliers and results in unconvincing voting scores. Overall, ST appears to be the second best method when faced with data decimation.
  • Robustness with reference to clutter and occlusion. The U3OR dataset provides the quantitative information of clutter and occlusion for each matching pair. We first split the matching pairs into three groups with [ 65 % ,   75 % ) , [ 75 % ,   85 % ) , and [ 85 % ,   95 % ) clutter, respectively. Analogously, we then generate another three groups with [ 60 % ,   70 % ) , [ 70 % ,   80 % ) , and [ 80 % ,   90 % ) occlusion, respectively. The results are shown in Figure 6g–l.
From these figures, one can see that PCV outperforms all compared methods under all levels of clutter and occlusion, followed by CV. Although CV and SI are also voting-based methods, they require more complex geometric constraints and are still inferior to PCV. It also reflects the effectiveness of the proposed progressive scheme for voting set optimization. Another remarkable phenomenon is the performance of ST. When faced with clutter and occlusion, ST is even inferior to NN and is the bottom-placed method. This is because ST has an assumption that inliers should appear in a cluster form, which is of rare occurrence when the point clouds are contaminated by severe clutter and occlusion. Generally, three voting-based methods, i.e., PCV, CV, and SI, are more robust to clutter and occlusion than other tested methods. Because voting-based methods search inliers individually, some isolated inliers may be identified.
  • Robustness with reference to data modality change. Both BoD1 and U3OR are LiDAR datasets. To test a method’s robustness to data modality change, we conduct experiments on the Kinect-captured BoD5 dataset. The results are shown in Figure 7.
As witnessed by Figure 7a, PCV consistently surpasses other compared methods on this dataset, showing strong resilience to data modality variation. We also present some visual results of PCV in Figure 7b. One can see that the confidence scores assigned to correspondences are iteratively optimized. More specifically, voting scores become more distinguishable as rendered by pseudo-color lines. Therefore, consistent correspondences can be found by ranking correspondences based on PCV’s confidence scores.
  • Correspondence grouping performance. Because one can also assign binary labels to correspondences based on PCV, we test the correspondence grouping (mismatch removal) performance of feature matchers on the whole U3OR dataset. The objective of correspondence grouping is to split the initial correspondence set into two subsets composed by inliers and outliers, respectively [11]. This experiment additionally takes the recent LPM [33] matcher into consideration. Except for LPM, other tested methods (confidence score-based ones) group correspondences based on Otsu’s thresholding strategy [41] as in [21]. We evaluate the correspondence grouping performance via precision, recall, and F-score metrics [11]. The results are shown in Table 3.
In terms of the F-score performance, the proposed PCV is the best competitor, followed by CV and LPM. In addition, PCV delivers the best precision performance and the second best recall performance. The comparison between PCV and CV again verifies the rationality of mining good voters. PCV also achieves better performance than LPM. A potential reason is that PCV defines the correctness of correspondences based on global geometric consistencies while LPM prefers correspondences preserving local structures. In particular, regarding the recall performance, PCV is slightly worse than LPM. This is because PCV is an iterative method, which imposes tighter constraints to correspondences than the non-iterative LPM method. For point clouds undergoing a rigid transformation model, global consistencies are supposed to be more reliable.
In general, it is very difficult to achieve good performance under a variety of nuisances and we can find the performance of most compared methods fluctuates significantly under different nuisances. However, PCV consistently achieves the best performance under all tested conditions. It can be explained from at least two aspects. (1) PCV is aware of the significance of finding a high-quality voting set, and proposes a progressive voter mining solution. As validated in Figure 5, PCV achieves outstanding performance due to the mined convincing voters. (2) Because of the high-quality voting set, PCV adapts well with the simple L 2 distance constraint. Although this constraint is ambiguous sometimes [22], it holds strong robustness with respect to many common nuisances. This trait makes the compatibility check process robust as well under different challenging conditions.

4.6. Time Efficiency

To compare the efficiency performance of tested methods, we record the time costs of these methods when processing initial correspondence sets with different cardinalities. This process is repeated for ten times, and the mean values are retained. The results are reported in Table 4.
Three main observations can be made from the figure. First, NN and NNSR can deal with thousands of correspondences in real-time. Nonetheless, both methods are sensitive to a number of nuisances. Second, ST is the most time-consuming method. Its time cost is significantly greater than others’ with dense initial correspondences. It is because the eigenvalue decomposition operation of ST is very time-consuming for high-order matrices. Third, CV is the most efficient one among three tested voting-based methods, followed by our PCV. Note that PCV obtains better feature matching performance than CV with respect to a number of nuisances. Overall, PCV is still a good option for applications with a strict demand on time consumption, because it can handle ∼1000 correspondences within 0.2 s.

4.7. Application to 3D Scene Registration

Feature matching usually plays a critical role in point cloud registration applications. Thus, we test PCV’s performance when applied to 3D scene registration. We conduct this experiment on the Aug_ICL-NUIM benchmark [46]. This benchmark assesses the registration performance via three metrics, i.e., precision, recall, and F-score (please refer to [46] for more details). Specifically, we perform 3D registration with a standard local feature matching-based pipeline proposed in [2]. This pipeline has been implemented in the point cloud library [47], and we dub the method as PCL. We propose a variant of PCL, namely PCL+PCV, which first selects the top-100 scored correspondences based on PCV and then feds them to the RANSAC estimator. The results are shown in Table 5.
It is clear that using PCV-selected correspondences can effectively boost the registration performance, as a gain of 11.4 percentages is achieved in terms of the F-score performance when performing PCV within the PCL framework. PCL+PCV also achieves the best F-score performance, the best recall performance, and the second best precision performance among all compared registration methods. This is reasonable because the correspondences selected by PCV are very consistent and RANSAC can estimate accurate transformations for correspondences sets with high inlier ratios. Note that PCL+PCV is slightly inferior to FGR in terms of precision. This is because FGR employs a more strict criterion for judging if a registration is correct. It results in better precision performance yet limited recall performance. Figure 8 presents some visual registration results obtained via PCL+PCV, indicating that consistent correspondences are generated by PCV and scene fragments are precisely aligned.

5. Conclusions

This paper has presented a robust feature matching method for 3D point clouds named PCV. The basic idea is to find a reliable voting set and assess the geometric consistency of each correspondence with the components in the voting set. Compared with existing methods, PCV holds two unique properties:
  • PCV is the first method that highlights the significance of voting set definition for voting-based methods. It also affords a progressive and adaptive approach to mine convincing voters.
  • Unlike existing methods that make great efforts to carefully design or combine geometric constraints, PCV is able to deliver outstanding performance with very simple constraints.
Feature matching experiments on three standard datasets and comparison with several state-of-the-art methods demonstrate that PCV is robust to noise, data decimation, clutter, occlusion, and data modality change. 3D registration experiments further show that PCV can improve the registration performance with existing pipelines.
Our PCV could be applied to the following application scenarios. First, 3D reconstruction from point clouds, which need to establish consistent point-to-point correspondences between point cloud views. Second, model-based 3D object recognition. This requires correspondences between the 3D model and the point cloud scene, which can be provided by PCV.
In the future, we plan to further improve PCV from the following two aspects. First, a better initialization could be achieved for PCV to determine the voting set, such as using existing feature matchers (e.g., GTM [36]). Second, an automatic stopping criterion for PCV iterations is desired, because correspondence sets with different inlier ratios may need different iterations.

Author Contributions

Supervision, S.Q.; Validation, K.Y. (Kaixiao Ye); Visualization, K.N.; Writing—original draft, K.Y. (Kunpeng Yin); Writing—review & editing, S.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China (62006025), Chinese Post-Doctoral Foundation (2020M673319), Natural Science Basic Research Plan in Shaanxi Province of China (Grant No. 2021JQ-290), and Shaanxi Provincial Key Research and Development Program (No. 2020GY-060).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, J.; Xiao, Y.; Cao, Z. Toward the Repeatability and Robustness of the Local Reference Frame for 3D Shape Matching: An Evaluation. IEEE Trans. Image Process. 2018, 27, 3766–3781. [Google Scholar] [CrossRef]
  2. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  3. Wang, B.; Zhou, J.; Huang, Y.; Wang, Y.; Huang, B. Improved Registration Algorithm Based on Double Threshold Feature Extraction and Distance Disparity Matrix. Sensors 2022, 22, 6525. [Google Scholar] [CrossRef]
  4. Hu, X.; Zhang, D.; Chen, J.; Wu, Y.; Chen, Y. NrtNet: An Unsupervised Method for 3D Non-Rigid Point Cloud Registration Based on Transformer. Sensors 2022, 22, 5128. [Google Scholar] [CrossRef]
  5. Souza Neto, P.; Marques Soares, J.; Pereira Thé, G.A. Uniaxial Partitioning Strategy for Efficient Point Cloud Registration. Sensors 2022, 22, 2887. [Google Scholar] [CrossRef]
  6. Guo, Y.; Sohel, F.; Bennamoun, M.; Lu, M.; Wan, J. Rotational projection statistics for 3D local surface description and object recognition. Int. J. Comput. Vis. 2013, 105, 63–86. [Google Scholar] [CrossRef] [Green Version]
  7. Tateno, K.; Tombari, F.; Navab, N. When 2.5 D is not enough: Simultaneous reconstruction, segmentation and recognition on dense SLAM. In Proceedings of the IEEE International Conference on Robotics and Automation, Stockholm, Sweden, 16–21 May 2016; pp. 2295–2302. [Google Scholar]
  8. Mian, A.S.; Bennamoun, M.; Owens, R.A. Automatic correspondence for 3D modeling: An extensive review. Int. J. Shape Model. 2005, 11, 253–291. [Google Scholar] [CrossRef]
  9. Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J. 3D object recognition in cluttered scenes with local surface features: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2270–2287. [Google Scholar] [CrossRef]
  10. Ma, J.; Zhou, H.; Zhao, J.; Gao, Y.; Jiang, J.; Tian, J. Robust feature matching for remote sensing image registration via locally linear transforming. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6469–6481. [Google Scholar] [CrossRef]
  11. Yang, J.; Xian, K.; Xiao, Y.; Cao, Z. Performance Evaluation of 3D Correspondence Grouping Algorithms. In Proceedings of the International Conference on 3D Vision, Qingdao, China, 10–12 October 2017; pp. 467–476. [Google Scholar]
  12. Yang, J.; Xian, K.; Wang, P.; Zhang, Y. A Performance Evaluation of Correspondence Grouping Methods for 3D Rigid Data Matching. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1859–1874. [Google Scholar] [CrossRef]
  13. Tombari, F.; Salti, S.; Di Stefano, L. Performance evaluation of 3D keypoint detectors. Int. J. Comput. Vis. 2013, 102, 198–220. [Google Scholar] [CrossRef]
  14. Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J.; Kwok, N.M. A comprehensive performance evaluation of 3D local feature descriptors. Int. J. Comput. Vis. 2016, 116, 66–89. [Google Scholar] [CrossRef]
  15. Yang, J.; Cao, Z.; Zhang, Q. A fast and robust local descriptor for 3D point cloud registration. Inf. Sci. 2016, 346, 163–179. [Google Scholar] [CrossRef]
  16. Yang, J.; Zhang, Q.; Xian, K.; Xiao, Y.; Cao, Z. Rotational contour signatures for both real-valued and binary feature representations of 3D local shape. Comput. Vis. Image Underst. 2017, 160, 133–147. [Google Scholar] [CrossRef]
  17. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  18. Guo, Y.; Sohel, F.; Bennamoun, M.; Wan, J.; Lu, M. A novel local surface feature for 3D object recognition under clutter and occlusion. Inf. Sci. 2015, 293, 196–213. [Google Scholar] [CrossRef]
  19. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  20. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
  21. Buch, A.G.; Yang, Y.; Krüger, N.; Petersen, H.G. In search of inliers: 3d correspondence by local and global voting. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2075–2082. [Google Scholar]
  22. Yang, J.; Xiao, Y.; Cao, Z.; Yang, W. Ranking 3D feature correspondences via consistency voting. Pattern Recognit. Lett. 2019, 117, 1–8. [Google Scholar] [CrossRef]
  23. Leordeanu, M.; Hebert, M. A spectral technique for correspondence problems using pairwise constraints. Int. Conf. Comput. Vis. 2005, 2, 1482–1489. [Google Scholar]
  24. Yang, J.; Huang, Z.; Quan, S.; Zhang, Q.; Zhang, Y.; Cao, Z. Toward Efficient and Robust Metrics for RANSAC Hypotheses and 3D Rigid Registration. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 893–906. [Google Scholar] [CrossRef]
  25. Yang, J.; Huang, Z.; Quan, S.; Cao, Z.; Zhang, Y. RANSACs for 3D Rigid Registration: A Comparative Evaluation. IEEE/CAA J. Autom. Sin. 2022, 9, 1861–1878. [Google Scholar] [CrossRef]
  26. Chen, H.; Bhanu, B. 3D free-form object recognition in range images using local surface patches. Pattern Recognit. Lett. 2007, 28, 1252–1262. [Google Scholar] [CrossRef]
  27. Tombari, F.; Di Stefano, L. Object recognition in 3D scenes with occlusions and clutter by Hough voting. In Proceedings of the Pacific-Rim Symposium on Image and Video Technology, Singapore, 14–17 November 2010; pp. 349–355. [Google Scholar]
  28. Ma, J.; Zhao, J.; Tian, J.; Yuille, A.L.; Tu, Z. Robust point matching via vector field consensus. IEEE Trans. Image Process. 2014, 23, 1706–1721. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Ma, J.; Qiu, W.; Zhao, J.; Ma, Y.; Yuille, A.L.; Tu, Z. Robust L2E estimation of transformation for non-rigid registration. IEEE Trans. Signal Process. 2015, 63, 1115–1129. [Google Scholar] [CrossRef]
  30. Lu, R.; Zhu, F.; Wu, Q.; Fu, X. Search inliers based on redundant geometric constraints. Vis. Comput. 2018, 36, 253–266. [Google Scholar] [CrossRef]
  31. Ma, J.; Jiang, X.; Jiang, J.; Zhao, J.; Guo, X. Lmr: Learning a two-class classifier for mismatch removal. IEEE Trans. Image Process. 2019, 28, 4045–4059. [Google Scholar] [CrossRef]
  32. Zhao, C.; Cao, Z.; Li, C.; Li, X.; Yang, J. NM-Net: Mining Reliable Neighbors for Robust Feature Correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 215–224. [Google Scholar]
  33. Ma, J.; Zhao, J.; Jiang, J.; Zhou, H.; Guo, X. Locality preserving matching. Int. J. Comput. Vis. 2019, 127, 512–531. [Google Scholar] [CrossRef]
  34. Mian, A.S.; Bennamoun, M.; Owens, R.A. A novel representation and feature matching algorithm for automatic pairwise registration of range images. Int. J. Comput. Vis. 2006, 66, 19–40. [Google Scholar] [CrossRef] [Green Version]
  35. Rusu, R.B.; Marton, Z.C.; Blodow, N.; Beetz, M. Persistent point feature histograms for 3D point clouds. In Proceedings of the International Conference on Intelligent Autonomous Systems, Baden Baden, Germany; 2008; pp. 119–128. [Google Scholar]
  36. Rodolà, E.; Albarelli, A.; Bergamasco, F.; Torsello, A. A scale independent selection process for 3d object recognition in cluttered scenes. Int. J. Comput. Vis. 2013, 102, 129–145. [Google Scholar] [CrossRef] [Green Version]
  37. Cirujeda, P.; Cid, Y.D.; Mateo, X.; Binefa, X. A 3D scene registration method via covariance descriptors and an evolutionary stable strategy game theory solver. Int. J. Comput. Vis. 2015, 115, 306–329. [Google Scholar] [CrossRef]
  38. Tombari, F.; Salti, S.; Di Stefano, L. Unique signatures of histograms for local surface description. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
  39. Drost, B.; Ulrich, M.; Navab, N.; Ilic, S. Model globally, match locally: Efficient and robust 3D object recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 998–1005. [Google Scholar]
  40. Yang, J.; Zhang, Q.; Cao, Z. Multi-attribute statistics histograms for accurate and robust pairwise registration of range images. Neurocomputing 2017, 251, 54–67. [Google Scholar] [CrossRef]
  41. Otsu, N. A threshold selection method from gray-level histograms. Automatica 1975, 11, 23–27. [Google Scholar] [CrossRef]
  42. Salti, S.; Tombari, F.; Di Stefano, L. SHOT: Unique signatures of histograms for surface and texture description. Comput. Vis. Image Underst. 2014, 125, 251–264. [Google Scholar] [CrossRef]
  43. Mian, A.S.; Bennamoun, M.; Owens, R. Three-dimensional model-based object recognition and segmentation in cluttered scenes. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1584–1601. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Cho, M.; Sun, J.; Duchenne, O.; Ponce, J. Finding matches in a haystack: A max-pooling strategy for graph matching in the presence of outliers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2083–2090. [Google Scholar]
  45. Sipiran, I.; Bustos, B. Harris 3D: A robust extension of the Harris operator for interest point detection on 3D meshes. Vis. Comput. 2011, 27, 963–976. [Google Scholar] [CrossRef]
  46. Choi, S.; Zhou, Q.Y.; Koltun, V. Robust reconstruction of indoor scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5556–5565. [Google Scholar]
  47. Rusu, R.B.; Cousins, S. 3d is here: Point cloud library (pcl). In Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 1–4. [Google Scholar]
  48. Mellado, N.; Aiger, D.; Mitra, N.J. Super 4pcs fast global pointcloud registration via smart indexing. Comput. Graph. Forum. 2014, 33, 205–215. [Google Scholar] [CrossRef]
  49. Zhou, Q.Y.; Park, J.; Koltun, V. Fast global registration. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 766–782. [Google Scholar]
Figure 1. Pipeline of the proposed PCV method. Given initial correspondences (cyan lines), PCV iteratively updates the voting set, checks the geometric conistency of each correspondence with the voters in the voting set, and assigns a voting/confidence score to each correspondence (rendered in colored lines). Consistent correspondences (green lines) can be found by simply ranking the correspondences based on voting scores.
Figure 1. Pipeline of the proposed PCV method. Given initial correspondences (cyan lines), PCV iteratively updates the voting set, checks the geometric conistency of each correspondence with the voters in the voting set, and assigns a voting/confidence score to each correspondence (rendered in colored lines). Consistent correspondences (green lines) can be found by simply ranking the correspondences based on voting scores.
Sensors 22 07718 g001
Figure 2. Three popular geometric constraints for the assessment of correspondence compatibility (illustrated in 2D).
Figure 2. Three popular geometric constraints for the assessment of correspondence compatibility (illustrated in 2D).
Sensors 22 07718 g002
Figure 3. An adaptive strategy to determine the cardinality of the voting set. C i n i t i a l is first sorted according to the voting scores at the i 1 th iteration; an adaptive threshold τ o t s u is then computed based on Otsu’s thresholding method [41]; the number of correspondences whose voting scores are greater than τ o t s u is defined as κ . The point cloud pair used to generate this figure is taken from the Bologna Dataset1 (BoD1) [42] dataset.
Figure 3. An adaptive strategy to determine the cardinality of the voting set. C i n i t i a l is first sorted according to the voting scores at the i 1 th iteration; an adaptive threshold τ o t s u is then computed based on Otsu’s thresholding method [41]; the number of correspondences whose voting scores are greater than τ o t s u is defined as κ . The point cloud pair used to generate this figure is taken from the Bologna Dataset1 (BoD1) [42] dataset.
Sensors 22 07718 g003
Figure 4. Visualization of three sample point cloud pairs from the experimental datasets.
Figure 4. Visualization of three sample point cloud pairs from the experimental datasets.
Sensors 22 07718 g004
Figure 5. Experimental analysis for PCV. (a) The effect when using voting sets with different inlier ratios on the feature matching performance. (b) Parameter analysis for N i t e r . (c) Parameter analysis for | C v o t i n g 0 | . (d) The effectiveness of making the cardinality of the voting set κ (Equation (8)) adaptive.
Figure 5. Experimental analysis for PCV. (a) The effect when using voting sets with different inlier ratios on the feature matching performance. (b) Parameter analysis for N i t e r . (c) Parameter analysis for | C v o t i n g 0 | . (d) The effectiveness of making the cardinality of the voting set κ (Equation (8)) adaptive.
Sensors 22 07718 g005
Figure 6. ROI performance of PCV and compared methods under different levels of (ac) Gaussian noise, (df) data decimation, (gi) clutter, and (jl) occlusion.
Figure 6. ROI performance of PCV and compared methods under different levels of (ac) Gaussian noise, (df) data decimation, (gi) clutter, and (jl) occlusion.
Sensors 22 07718 g006
Figure 7. Feature matching results on the BoD5 dataset. (a) The ROI performance of six tested methods. (b) Visualization of the scoring result at each iteration of PCV (voting scores are encoded with pseudo-color).
Figure 7. Feature matching results on the BoD5 dataset. (a) The ROI performance of six tested methods. (b) Visualization of the scoring result at each iteration of PCV (voting scores are encoded with pseudo-color).
Sensors 22 07718 g007
Figure 8. Sample registration results from the Aug_ICL-NUIM benchmark based on PCL+PCV. From left to right: initial point clouds, scoring results by PCV, top-scored correspondences selected by PCV, point clouds after registration.
Figure 8. Sample registration results from the Aug_ICL-NUIM benchmark based on PCL+PCV. From left to right: initial point clouds, scoring results by PCV, top-scored correspondences selected by PCV, point clouds after registration.
Sensors 22 07718 g008
Table 1. The ROI performance (%) of PCV when using different initial voting sets.
Table 1. The ROI performance (%) of PCV when using different initial voting sets.
C voting 0 K = 20 K = 50 K = 100 K = 500
NNSR11.7227.7449.1296.50
NN11.2527.7447.9694.20
Random9.9122.5345.2691.08
Table 2. The ROI performance (%) of PCV when using different geometric constraints.
Table 2. The ROI performance (%) of PCV when using different geometric constraints.
Constraints K = 20 K = 50 K = 100 K = 500
L 2 11.7227.7449.1296.50
L 2 + Normal12.0527.7450.0296.61
L 2 + LRF11.2726.4048.4995.43
L 2 + Normal + LRF11.9027.5849.4096.50
Table 3. Correspondence grouping performance of tested methods on the U3OR dataset.
Table 3. Correspondence grouping performance of tested methods on the U3OR dataset.
Precision (%)Recall (%)F-Score (%)
NN8.250.314.1
NNSR20.847.314.5
ST [23]14.530.119.6
SI [21]22.050.830.7
CV [22]30.268.441.9
LPM [33]21.473.233.1
PCV31.872.944.3
Table 4. Time costs (measured by second) of six feature matching methods respecting different numbers of initial correspondences. The symbol “−” denotes the time cost smaller than 1.0 × 10 4 s.
Table 4. Time costs (measured by second) of six feature matching methods respecting different numbers of initial correspondences. The symbol “−” denotes the time cost smaller than 1.0 × 10 4 s.
# CorrespondencesNNNNSRSTSICVPCV
500.0430.0550.0010.002
1000.4440.2590.0040.010
2001.6480.6120.0150.038
5000.000100.000109.9340.8560.0370.081
10000.000130.0001339.9551.1590.0740.182
20000.000180.00018157.3581.8140.1460.388
Table 5. 3D scene registration performance on the Aug_ICL-NUIM benchmark [46].
Table 5. 3D scene registration performance on the Aug_ICL-NUIM benchmark [46].
Precision (%)Recall (%)F-Score (%)
OpenCV [39]1.65.32.5
Super 4PCS [48]10.417.813.1
PCL [2]14.044.921.3
FGR [49]23.251.131.9
CZK [46]19.659.229.4
Ours (PCL + PCV)21.865.232.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Quan, S.; Yin, K.; Ye, K.; Nan, K. Robust Feature Matching for 3D Point Clouds with Progressive Consistency Voting. Sensors 2022, 22, 7718. https://doi.org/10.3390/s22207718

AMA Style

Quan S, Yin K, Ye K, Nan K. Robust Feature Matching for 3D Point Clouds with Progressive Consistency Voting. Sensors. 2022; 22(20):7718. https://doi.org/10.3390/s22207718

Chicago/Turabian Style

Quan, Siwen, Kunpeng Yin, Kaixiao Ye, and Kechen Nan. 2022. "Robust Feature Matching for 3D Point Clouds with Progressive Consistency Voting" Sensors 22, no. 20: 7718. https://doi.org/10.3390/s22207718

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop