A Comparative Study of Weighting Methods for Local Reference Frame

: In the ﬁeld of photogrammetric engineering, computer vision, and graphics, local shape description is an active research area. A wide variety of local shape descriptors (LSDs) have been designed for di ﬀ erent applications, such as shape retrieval, object recognition, and 3D registration. The local reference frame (LRF) is an important component of the LSD. Its repeatability and robustness directly inﬂuence the descriptiveness and robustness of the LSD. Several weighting methods have been proposed to improve the repeatability and robustness of the LRF. However, no comprehensive comparison has been implemented to evaluate their performance under di ﬀ erent data modalities and nuisances. In this paper, we focus on the comparison of weighting methods by using six datasets with di ﬀ erent data modalities and application contexts. We evaluate the repeatability of the LRF under di ﬀ erent nuisances, including occlusion, clutter, partial overlap, varying support radii, Gaussian noise, shot noise, point density variation, and keypoint localization error. Through the experiments, the traits, advantages, and disadvantages of weighting methods are summarized. point density variation, and keypoint localization error.


Introduction
Local shape description has been proved to be very successful in the field of photogrammetric engineering, computer vision and graphics. Numerous applications are performed by the local shape description, such as 3D object recognition [1][2][3], simultaneous localization and mapping [4], 3D modeling and scene reconstruction [5,6], shape retrieval [7], and 3D object categorization [8].
In comparison with the global shape descriptor, the local shape descriptor (LSD), which is an important component in the local shape description, holds a number of advantages, including being robust against clutter, occlusion, and missing regions [9]. This leads to that many scholars start to do research on the LSD.
A large number of LSDs have been proposed, which encode the geometric information of the local neighborhood around a specific keypoint by calculating geometric features such as distance [10], local depth [11], and normal deviation [12]. Essentially, the LSD should be invariant to rigid transformation, because point-to-point correspondences are established by matching the LSDs. However, in real applications, the LSD is easily influenced by noise, point density variation, clutter, occlusion, and missing regions. Therefore, an ideal LSD should have the ability to resist these nuisances. Many efforts have been made to deal with these difficulties, such as signature of Although the comparative studies of both the LSDs and the LRF methods have been presented in the literature, the comparative study of weighting methods for the LRF has not been conducted. The performance of the LRF is affected by multiple factors, so comparing the LRF methods directly is very hard to determine which factors make an improvement to the LRF. A weighting method plays an important role in an LRF method. Assigning proper weight values to the points in the neighborhood can improve the repeatability and robustness of the LRF. Therefore, in this paper, we focus on the comparison of weighting methods. Majority of these methods calculate the weight values for the points in the local neighborhood according to the distances from the keypoint to these points, but which methods are better is unknown. In this paper, five weighting methods are compared, and a quantitative assessment of them is presented. Six datasets with different data modalities and application contexts are selected to perform the comparative study. The repeatability of the LRF against various nuisances is tested. In addition, the Gaussian function is introduced to build the LRF. Unlike other weighting methods, as long as the weighting function is determined, the weight of each point is fixed. In the Gaussian function, the weight of each point can be changed to adjust its contribution to the LRF. Through the comparative study, the main conclusions are summarized as follows: (1) When the wights of the distant points in the local neighborhood are too large, the LRF is sensitive to occlusion, clutter, partial overlap, noise, outliers, and keypoint localization error. By contrast, when the wights of the distant points are too small, the LRF is susceptible to noise, varying point density, and keypoint localization error. (2) The weighting method should be mainly designed to achieve the robustness to noise and keypoint localization error. Then, it should be properly adjusted to get a balanced robustness to noise, keypoint localization error, and shape incompleteness. The robustness to point density variation and outliers should be obtained by extra methods. (3) No method is generalized, but the GF can always get good performance on different data modalities by changing the value of the Gaussian parameter. Therefore, GF can be regarded as a generalized method.
The remainder of this paper is organized as follows. In Section 2, five weighting methods are reviewed. The evaluation methodology is described with details in Section 3. The experimental results and analysis are presented in Section 4. In Section 5, the conclusion is drawn.

Overview of Five Weighting Methods
The weighting methods are mainly related to the CA-based methods, so we employ the CA-based methods to perform the comparative study. Furthermore, the CA-based methods generally have higher computational efficiency in comparison with the GA-based methods and are easy to implement. The sign disambiguation technique [13] is applied to eliminate the sign ambiguity of the LRF for all the considered methods because our aim is to compare the weighting methods. In addition, we also use the keypoint to replace the barycenter when computing the covariance matrix, as proposed in reference [13]. The process diagram of building an LRF is shown in Figure 1.
The covariance matrix is computed as where p is the keypoint, N(p) is the radius neighborhood of p, q i is a neighboring point in N(p), w q i is the weight of q i . After performing the eigenvalue decomposition on C(p), we can obtain three eigenvalues (λ 1 , λ 2 , λ 3 ) that satisfy λ 1 ≥ λ 2 ≥ λ 3 and three corresponding eigenvectors (v 1 , v 2 , v 3 ). The three eigenvectors provide a basis for LRF definition. Specifically, the sign disambiguation technique is applied to redetermine the signs of v 1 and v 3 , producing two unambiguous vectors v 1 and v 3 . v 2 is defined as v 1 × v 3 , and v 1 , v 2 and v 3 define the directions of the x, y, and z-axes, respectively.  Figure 1. The process diagram of building a local reference frame.
EM [26]: This method does not take any nuisances (e.g., occlusion and clutter) into consideration. The weights for all points are just set as 1. That is ( It means that all points in the local neighborhood make the same contribution to the covariance matrix. In spite of this, the method may obtain more repeatable and robust LRF for some data modalities, as will be shown in Section 4.1.
SHOT [13]: This method treats the local points unequally to enhance the robustness under 3D object recognition scenarios. The weights are computed as where R is the support radius, p q  i is the 3D Euclidean distance between i q and p . Thus, the distant points (i.e., the neighboring points have larger distances to the keypoint) are assigned smaller weights. In other words, the contributions of the distant points to the covariance matrix are weakened. The purpose is to increase repeatability in the presence of clutter.
BSC [24]: Dong et al. developed a different weighting strategy to determine the weights. It also assigns smaller weights to the distant points. The weights are computed as This weighting strategy is expected to improve the robustness of the LRF against occlusion and clutter because the distant points are easily influenced by occlusion and clutter. Unlike the SHOT method, this method limits the weight values to be between 0 and 1. EM [26]: This method does not take any nuisances (e.g., occlusion and clutter) into consideration. The weights for all points are just set as 1. That is It means that all points in the local neighborhood make the same contribution to the covariance matrix. In spite of this, the method may obtain more repeatable and robust LRF for some data modalities, as will be shown in Section 4.1.
SHOT [13]: This method treats the local points unequally to enhance the robustness under 3D object recognition scenarios. The weights are computed as where R is the support radius, q i − p is the 3D Euclidean distance between q i and p. Thus, the distant points (i.e., the neighboring points have larger distances to the keypoint) are assigned smaller weights. In other words, the contributions of the distant points to the covariance matrix are weakened. The purpose is to increase repeatability in the presence of clutter. BSC [24]: Dong et al. developed a different weighting strategy to determine the weights. It also assigns smaller weights to the distant points. The weights are computed as This weighting strategy is expected to improve the robustness of the LRF against occlusion and clutter because the distant points are easily influenced by occlusion and clutter. Unlike the SHOT method, this method limits the weight values to be between 0 and 1.
TOLDI [18]: Although the method is designed for a GA-based LRF method, it also computes the weight values according to the distances. Therefore, we also include it in the comparison. By using this method, a CA-based LRF method may obtain the more repeatable and robust LRF, and the information Appl. Sci. 2020, 10, 3223 5 of 14 obtained by the comparison also has some reference value for the GA-based methods. The weights are calculated as This weighting strategy is designed to improve the robustness of the LRF against clutter, occlusion, and missing regions because the distant points contribute less to the covariance matrix.
Gaussian function (GF) [27]: Levin used the GF in the moving least-squares approach to approximate the local surface. Here, it is introduced to determine the weights for the LRF, because better performance may be obtained by this method. The weights are calculated by where h = σ · R, and σ is the Gaussian parameter. It can be seen from Equation (6) that the weights decrease as the distance increases, so the distant points are assigned smaller weights, similar to SHOT, BSC, and TOLDI. Therefore, this weighting strategy is also useful to improve the robustness of the LRF against clutter, occlusion, and missing regions. However, the introduction of σ provides flexibility in the weight assignment. The big value of σ slightly weakens the contributions of the distant points, while the small value of σ greatly weakens the contributions of the distant points. That is, a better weight assignment would be achieved if the value σ is properly selected. In Section 4, five different values of σ (i.e., 0.2, 0.4, 0.6, 0.8, and 1) are selected to perform the comparison.

Evaluation Methodology
To assess and compare the weighting methods mentioned in Section 2, we chose 6 benchmark datasets, which are described in detail in Section 3.1, to conduct our experiments. Then, the evaluation criterion is described in Section 3.2, which is used to evaluate the repeatability of the LRF.

Datasets
The selected 6 datasets are Retrieval [28], Laser Scanner [29], Kinect [13], Space Time [28], LiDAR Registration [30], and Kinect Registration [13]. These datasets are acquired by different devices, resulting in different data modalities. The Retrieval dataset is obtained by a Cyberware 3030 MS scanner. It contain 6 noise-free models. The Laser Scanner dataset include 5 models and 50 real scenes, forming 188 model-scene pairs. The real scenes are obtained by a Minolta Vivid 910 scanner. The Kinect dataset is acquired by a Microsoft Kinect sensor. It include 6 models and 16 real scenes, providing 43 model-scene pairs. The Space Time dataset is composed of 8 models and 15 real scenes. This dataset is created by the Space Time Stereo technique. It provide 24 model-scene pairs. The LiDAR Registration dataset is obtained by scanning 4 objects from different views using a Minolta Vivid 910 scanner, deriving 22, 16, 16, and 21 point clouds, respectively. The Kinect Registration dataset consists of 15, 16, 20, 13, 16, and 15 point clouds of 6 objects, which are also acquired by a Microsoft Kinect sensor. The Laser Scanner and LiDAR Registration datasets are high-quality, while the Kinect and Kinect Registration datasets are noisy and sparse. The Space Time dataset is contaminated by noise and outliers. The aim is to make our experiments cover various data modalities (i.e., LiDAR, Kinect and Space Time) and application scenarios (i.e., shape retrieval, 3D object recognition and 3D registration), and different kinds of nuisances, including noise, point density variation, clutter, occlusion, shot noise, and partial overlap (i.e., missing regions) are contained in these datasets. The detailed description of the 6 datasets is listed in Table 1, and the examples of some of them are shown in Figure 2. Note that we call 2 input shapes as model and scene for simplicity, though such appellation was originally used for object recognition. examples of some of them are shown in Figure 2. Note that we call 2 input shapes as model and scene for simplicity, though such appellation was originally used for object recognition.

Evaluation Criterion
In theory, if two LRFs ( and scene are error-free, the two LRFs should conform to the equation GT R is the ground-truth rotation matrix. Therefore, we can use the criteria suggested in reference [1] to measure the error between two LRFs: After computing the errors of all the LRF pairs, we calculate the ratio of the LRF pairs in a dataset whose errors being smaller than 10 degrees by where 10 N denotes the number of the LRF pairs whose errors smaller than 10 degrees, t N denotes the total number of the LRF pairs.

Evaluation Criterion
In theory, if two LRFs (L(p m ), L(p s )) of corresponding keypoints (p m , p s ) in the model and scene are error-free, the two LRFs should conform to the equation L(p s ) = R GT L(p m ), where R GT is the ground-truth rotation matrix. Therefore, we can use the criteria suggested in reference [1] to measure the error between two LRFs: where L(p m ) = R GT L(p m ). For each dataset, the ground-truth transformation is known in advance. The transformation of the Retrieval dataset is artificially set and then used to simulate the point clouds of the corresponding scenes. The transformations of the Laser Scanner, Kinect, and Space Time datasets are provided by the publishers. The transformations of the LiDAR Registration and Kinect Registration datasets are obtained by manual alignment. After computing the errors of all the LRF pairs, we calculate the ratio of the LRF pairs in a dataset whose errors being smaller than 10 degrees by where N 10 denotes the number of the LRF pairs whose errors smaller than 10 degrees, N t denotes the total number of the LRF pairs. Eight kinds of nuisances (i.e., Gaussian noise, shot noise (i.e., outliers), point density variation, keypoint localization error, clutter, occlusion, partial overlap, and varying support radii) are tested in this paper. The detailed definition of each nuisance can be found in reference [20].

Implementation Details
For each dataset, we apply the uniform sampling method [31] to extract 1000 points from the model as the keypoints. Then, the ground-truth transformation is used to transform these keypoints, and these transformed keypoints are used to find the nearest points in the scene. Finally, these nearest points are treated as the corresponding keypoints in the scene. The aim is to remove the effect of keypoint localization error on the LRF because keypoint localization error should be tested independently, as inspired by reference [19]. For the datasets with incomplete shape (i.e., Laser Scanner, Kinect, Space Time, LiDAR Registration, and Kinect Registration), we find out the keypoints in the overlapping region. The LRFs are only calculated on the keypoints in the overlapping region. The support radius R is set as 15pr (pr denotes the point cloud resolution, i.e., the average shortest distance among neighboring points) in all the experiments except the experiments for testing varying support radii.

Test on the Six Datasets
The experimental results of the six datasets are listed in Table 2. For the Retrieval dataset, the 6 noise-free models are transformed to get 6 scenes by artificially setting the ground-truth transformation. Then the 6 scenes are down-sampled (sampling rate of 7/10), and Gaussian noise (standard deviation of 0.3pr) is added to them. The aim is to create a combination of nuisances. The other datasets are real data obtained by laser scanning devices, so no additional operation is implemented on them. In addition, we recommend that readers first see Section 4.4 to know which weighting method makes the distant points contribute more to the covariance matrix and which weighting method makes them contribute less to the covariance matrix. From Table 2, we can see that, on the Retrieval dataset, GF(σ = 0.8) obtains the best performance, followed by GF(σ = 1), GF(σ = 0.6), BSC, and SHOT. TOLDI achieve relatively poor performance. For this dataset, the weighting method is used to reduce the effect of Gaussian noise and point density variation. Because there is no clutter, occlusion, and missing region, the contributions of the distant points do not need to be declined. Therefore, the weights of the distant points should be increased. However, as can be seen from the calculated results, EM and GF(σ = 1) obtain inferior performance. This indicates that too large contributions of the distant points would also degrade the repeatability of the LRF, even without the effect of clutter, occlusion, and missing region.
On the Laser Scanner and LiDAR Registration datasets, GF(σ = 0.4), TOLDI, SHOT, and BSC have relatively good performance, whereas EM obtains poor performance. Both of the two datasets are high-quality, so they are less contaminated by noise. In this case, the contributions of distant points ought to be controlled properly, as will be shown in Section 4.3. Furthermore, their contributions should be weakened to improve the robustness to clutter, occlusion, and missing regions as well. The is the reason why GF(σ = 0.4) and TOLDI achieve good performance, while EM and GF(σ = 1) obtain poor performance on the two datasets.
For the Kinect and Kinect Registration datasets, the ranking of all the methods changes significantly. The three best methods are GF(σ = 0.8), GF(σ = 1), and EM, while TOLDI gets a relatively poor performance. In addition to shape incompleteness (i.e., occlusion, clutter, and missing regions), the two datasets are also rather sparse (i.e., the point density variation is obvious) and noisy. Especially for point density variation, the contributions of the distant points should be increased, as will be shown in Section 4.3. TOLDI is sensitive to point density variation, so it exhibits poor repeatability, even under the case with clutter, occlusion, and missing regions.
For the Space Time dataset, the four methods (GF(σ = 0.6), GF(σ = 0.8), SHOT, and BSC) achieve relatively good performance, among which GF(σ = 0.6) is the best one. In this dataset, besides clutter and occlusion, noise and outliers are the main nuisances. For outliers, the weights of the distant points should be reduced, because outliers are far away from the local surface. Moreover, this is helpful in reducing the effect of clutter and occlusion. However, if the weights of the distant points are too small, the LRF is sensitive to noise. Hence, TOLDI and GF(σ = 0.2) obtain the worst performance.
Based on the above discussion, we can know that no weighting method is always the best for all the datasets, because different datasets contain different kinds of nuisances, and the levels of nuisances are also different. Additionally, BSC is totally equivalent to SHOT, because the two methods always achieve the same results on the six datasets. On the test for varying support radii, the radius is set from 5pr to 30pr with a step of 5pr.
As we can see from Figure 3a,b, with the increase of occlusion and clutter, the repeatability of the LRF decreases on the whole. In Figure 3c, the repeatability of the LRF increases as the overlap increases. TOLDI, SHOT, BSC, and GF(σ = 0.4) are the best methods under different degrees of occlusion, clutter, or overlap. This indicates that reducing the contributions of the distant points indeed work to enhance the robustness to occlusion, clutter, and missing regions. We can also see that when the degree of occlusion, clutter, or overlap increases, the four methods are always the best. In other words, we do not need to further reduce the weights of the distant points when the degree of occlusion, clutter, or missing regions becomes larger. The weights of the distant points just need to be properly reduced. However, from the results in Table 2, we can know that if the experiments are tested on Kinect and Kinect Registration datasets, which are pretty sparse and noisy, the best methods would be GF(σ = 0.8), GF(σ = 1) and EM. Therefore, the weighting method should be designed to first reduce the effect of other nuisances (e.g., noise and keypoint localization error). Then, if occlusion, clutter, and missing regions are present, the weighting method should be further adjusted to reduce the effect of these nuisances.    ) and EM. Therefore, the weighting method should be designed to first reduce the effect of other nuisances (e.g., noise and keypoint localization error). Then, if occlusion, clutter, and missing regions are present, the weighting method should be further adjusted to reduce the effect of these nuisances. In Figure 3d, TOLDI is the best method when the support radius is 10pr and 15pr, whereas GF( 4 . 0   ) is the best method when the support radius is 20pr and 25pr. No method is always the best for all the support radii. With the increase of the support radius, the occlusion and clutter near the keypoint are introduced into the local neighborhood. Therefore, the weights of the distant points need to be further reduced. Especially, when the support radius is 30pr, GF( is the best method, which largely reduces the weights of the distant points. The occlusion and clutter that are far away from keypoint do not affect the repeatability of LRF. This explains why TOLDI, SHOT, BSC, and GF( 4 . 0   ) always achieve the best performance in Figure 3a,b when the degree of occlusion or clutter increases.

Repeatability of LRF Under Different Levels of Gaussian Noise, Point Density Variation, Shot Noise, and Keypoint Localization Error
In the experiments, the Retrieval dataset is applied. The four kinds of nuisances are respectively injected into the 6 noise-free scenes. The standard deviation of Gaussian noise is set as 0.1pr, 0.2pr, 0.3pr, 0.4pr, and 0.5pr to create noisy point clouds. The sampling rate is set as 9/10, 8/10, 7/10, 6/10, and 5/10 to simulate point density variation. The outlier ratio is set as 0.2%, 0.5%, 1%, 2%, and 5% to In Figure 3d, TOLDI is the best method when the support radius is 10pr and 15pr, whereas GF(σ = 0.4) is the best method when the support radius is 20pr and 25pr. No method is always the best for all the support radii. With the increase of the support radius, the occlusion and clutter near the keypoint are introduced into the local neighborhood. Therefore, the weights of the distant points need to be further reduced. Especially, when the support radius is 30pr, GF(σ = 0.2) is the best method, which largely reduces the weights of the distant points. The occlusion and clutter that are far away from keypoint do not affect the repeatability of LRF. This explains why TOLDI, SHOT, BSC, and GF(σ = 0.4) always achieve the best performance in Figure 3a,b when the degree of occlusion or clutter increases.

Repeatability of LRF Under Different Levels of Gaussian Noise, Point Density Variation, Shot Noise, and Keypoint Localization Error
In the experiments, the Retrieval dataset is applied. The four kinds of nuisances are respectively injected into the 6 noise-free scenes. The standard deviation of Gaussian noise is set as 0.1pr, 0.2pr, 0.3pr, 0.4pr, and 0.5pr to create noisy point clouds. The sampling rate is set as 9/10, 8/10, 7/10, 6/10, and 5/10 to simulate point density variation. The outlier ratio is set as 0.2%, 0.5%, 1%, 2%, and 5% to extract some points. These points are moved along their normal vectors by a magnitude of 0.5 × R to create outliers. The Gaussian noise with a standard deviation of 0.2pr, 0.4pr, 0.6pr, 0.8pr, and 1pr is added to the scene keypoints to simulate keypoint localization error. The experimental results are illustrated in Figure 4. to create outliers. The Gaussian noise with a standard deviation of 0.2pr, 0.4pr, 0.6pr, 0.8pr, and 1pr is added to the scene keypoints to simulate keypoint localization error. The experimental results are illustrated in Figure 4. As shown in Figure 4, for different nuisances, the best method is different. In terms of robustness to noise and keypoint localization error (Figure 4a,d), GF( 6 . 0   ), SHOT, and BSC achieve the best performance. It indicates that the contributions of distant points should be controlled properly for the two nuisances. Both contributing too many and too little will degrade the robustness of the LRF. In terms of robustness to point density variation (Figure 4b), GF( , and EM are the best methods. Therefore, the contributions of the distant points ought to be increased for point density variation. In terms of robustness to outliers (Figure 4c), GF( is the best method. This indicates that the contributions of the distant points should be weakened for outliers. This conclusion is just opposite to that drawn in reference [20]. The experiments of reference [20] were performed on the datasets with noise. Their conclusion is probably unreliable because the applied datasets are not only affected by outliers. Based on the experimental results in Sections 4.2 and 4.3, we can conclude that when the weights of the distant points are too large (e.g., EM), the LRF is susceptible to occlusion, clutter, missing regions, noise, outliers, and keypoint localization error. However, when the weights of the distant points are too small (e.g., GF( 2 . 0   )), the LRF is sensitive to noise, point density variation, and keypoint localization error. As shown in Figure 4, for different nuisances, the best method is different. In terms of robustness to noise and keypoint localization error (Figure 4a,d), GF(σ = 0.6), SHOT, and BSC achieve the best performance. It indicates that the contributions of distant points should be controlled properly for the two nuisances. Both contributing too many and too little will degrade the robustness of the LRF. In terms of robustness to point density variation (Figure 4b), GF(σ = 0.8), GF(σ = 1), and EM are the best methods. Therefore, the contributions of the distant points ought to be increased for point density variation. In terms of robustness to outliers (Figure 4c), GF(σ = 0.2) is the best method. This indicates that the contributions of the distant points should be weakened for outliers. This conclusion is just opposite to that drawn in reference [20]. The experiments of reference [20] were performed on the datasets with noise. Their conclusion is probably unreliable because the applied datasets are not only affected by outliers.
Based on the experimental results in Sections 4.2 and 4.3, we can conclude that when the weights of the distant points are too large (e.g., EM), the LRF is susceptible to occlusion, clutter, missing regions, noise, outliers, and keypoint localization error. However, when the weights of the distant points are too small (e.g., GF(σ = 0.2)), the LRF is sensitive to noise, point density variation, and keypoint localization error.

Comparison of Weights
In order to further compare the five weighting methods, we depicte the weight values of the five weighting methods, which are shown in Figure 5. Because the weights are relative quantities, we need to normalize the weights, i.e., all the weights are divided by the maximum weight for each method.

Comparison of Weights
In order to further compare the five weighting methods, we depicte the weight values of the five weighting methods, which are shown in Figure 5. Because the weights are relative quantities, we need to normalize the weights, i.e., all the weights are divided by the maximum weight for each method. From Figure 5b, we can easily know the contributions of the neighboring points with different distances to the covariance matrix in each method. The normalized weights of SHOT are absolutely the same as that of BSC. This explains why the two methods always get the same performance in the experiments. Therefore, we can also infer that the weighting method ( ) is equivalent to TOLDI. We can also see that EM and TOLDI are two relatively extreme methods. This explains why the two methods always achieve relatively poor performance on some datasets. Certainly, GF( is also an extreme method, so the method always obtains the poorest performance.

Performance Summary and Suggestions
We have presented the comprehensive performance of the LRF obtained by the five weighting methods. Several observations can be given. (1) The weighting methods lack universality to different data modalities. For example, GF( 4 . 0   ) and TOLDI achieve good performance on the Laser Scanner and LiDAR Registration datasets but poor performance on the Kinect and Kinect Registration datasets. GF( 8 . 0   ) behaves well on the Retrieval dataset but performs poorly on the Laser Scanner dataset. (2) The application scenario has little influence on the performance of the weighting method because, for different application scenarios with the same data modality, the most suitable method is basically the same. For instance, the Laser Scanner dataset is created for object recognition, and the LiDAR Registration dataset is created for 3D registration. GF( 4 . 0   ) and TOLDI obtain the best performance on both of them. The Kinect dataset is acquired for object recognition, and the Kinect Registration dataset is acquired for 3D registration. GF(  The weights calculated by the five weighting methods, including (a) weights, (b) normalized weights.
From Figure 5b, we can easily know the contributions of the neighboring points with different distances to the covariance matrix in each method. The normalized weights of SHOT are absolutely the same as that of BSC. This explains why the two methods always get the same performance in the experiments. Therefore, we can also infer that the weighting method (w q i = ( ) is equivalent to TOLDI. We can also see that EM and TOLDI are two relatively extreme methods. This explains why the two methods always achieve relatively poor performance on some datasets. Certainly, GF(σ = 0.2) is also an extreme method, so the method always obtains the poorest performance.

Performance Summary and Suggestions
We have presented the comprehensive performance of the LRF obtained by the five weighting methods. Several observations can be given. (1) The weighting methods lack universality to different data modalities. For example, GF(σ = 0.4) and TOLDI achieve good performance on the Laser Scanner and LiDAR Registration datasets but poor performance on the Kinect and Kinect Registration datasets. GF(σ = 0.8) behaves well on the Retrieval dataset but performs poorly on the Laser Scanner dataset.
(2) The application scenario has little influence on the performance of the weighting method because, for different application scenarios with the same data modality, the most suitable method is basically the same. For instance, the Laser Scanner dataset is created for object recognition, and the LiDAR Registration dataset is created for 3D registration. GF(σ = 0.4) and TOLDI obtain the best performance on both of them. The Kinect dataset is acquired for object recognition, and the Kinect Registration dataset is acquired for 3D registration. GF(σ = 0.8) and EM are the superior methods on both of them.
(3) The weighting method is always designed to reduce the contributions of the distant points for improving the robustness to shape incompleteness. However, this fails on some datasets such as the Kinect and Kinect Registration datasets. Even the occlusion, clutter, and missing regions are present, EM and GF(σ = 1) are the best methods. Because the point density variation is obvious in the two datasets, the contributions of the distant points should be enhanced, as shown in Figure 4b. (4) The weights of the distant points can be adjusted by varying the value of σ in the GF method. We can always find a proper value of σ for different data modalities to get a good performance. From this point of view, the GF is a generalized method, and therefore we suggest using this method to replace all the other methods. The value of σ can be determined according to the descriptor matching performance to get the robustness to noise and keypoint localization error, as used in many papers. Then, if shape incompleteness is present, the value of σ needs to be reduced properly to get a balanced robustness to noise, keypoint localization error, and shape incompleteness. It is worth noting that the value of σ can also be set as 0.3, 0.5, 0.7, 0.9, 1.1, and bigger values. Better performance may be obtained than that presented in the above experiments. In addition, finding a method to adaptively determine the value of σ for each keypoint is also desired. (5) No method is always suitable for different support radii. As shown in Figure 3d, when the support radius is 15pr, the repeatability of the LRF is the highest. Therefore, we suggest that the support radius is set as 15pr. (6) As shown in Figure 3a-c, with the increase of the degree of shape incompleteness, the ranking of the methods is unchanged on the whole. Hence, the weighting method should be mainly designed to get robustness to other nuisances, and then it should be properly adjusted to get the robustness to shape incompleteness. (7) Another weighting method to improve the robustness against point density variation is necessary because the ranking of the method under the nuisance is different from that under noise and keypoint localization error, as shown in Figure 4. (8) Similarly, a mechanism should be applied to detect and eliminate outliers.

Conclusions
We have conducted a comprehensive performance evaluation of five state-of-the-art weighting methods on six benchmark datasets, which refer to different data modalities and different application scenarios. The performance of the LRF is assessed under different kinds of nuisances, including occlusion, clutter, partial overlap, varying support radii, Gaussian noise, shot noise, point density variation, and keypoint localization error. Through our experiments, the five weighting methods are compared, and a qualitative assessment is provided. The performance summary and suggestions are also presented as a valuable reference for the design of a new LRF method.