A Fast Point Cloud Recognition Algorithm Based on Keypoint Pair Feature

At present, PPF-based point cloud recognition algorithms can perform better matching than competitors and be verified in the case of severe occlusion and stacking. However, including certain superfluous feature point pairs in the global model description would significantly lower the algorithm’s efficiency. As a result, this paper delves into the Point Pair Feature (PPF) algorithm and proposes a 6D pose estimation method based on Keypoint Pair Feature (K-PPF) voting. The K-PPF algorithm is based on the PPF algorithm and proposes an improved algorithm for the sampling point part. The sample points are retrieved using a combination of curvature-adaptive and grid ISS, and the angle-adaptive judgment is performed on the sampling points to extract the keypoints, therefore improving the point pair feature difference and matching accuracy. To verify the effectiveness of the method, we analyze the experimental results in scenes with different occlusion and complexity levels under the evaluation metrics of ADD-S, Recall, Precision, and Overlap rate. The results show that the algorithm in this paper reduces redundant point pairs and improves recognition efficiency and robustness compared with PPF. Compared with FPFH, CSHOT, SHOT and SI algorithms, this paper improves the recall rate by more than 12.5%.


Introduction
With the widespread usage of depth sensors, whether in pure 3D data applied to virtual reality or virtual reality fusion applied to augmented reality, it is necessary to identify and register 3D objects with different poses. As such, obtaining accurate and reliable 6D poses of objects through 3D data has received more and more attention. Over the last decade, several new forms of 3D object identification and pose estimation systems have been suggested. At present, it is mainly divided into the following categories: (1) The ICP algorithm proposed by Besl [1] registers the point cloud, but the ICP algorithm is prone to fall into the local optimal solution. As such, further scholars have studied this problem, [2][3][4] and other algorithms have improved algorithms based on ICP, which are more robust against false matches, avoid expensive nearest neighbor searches, and maintain the accuracy of the algorithm. (2) ICP is combined with other algorithms. Algorithms such as [5,6] combine ICP with other registration algorithms, which has advantages in registration accuracy, but take longer and aree not suitable for point cloud matching with less obvious features. The standard effect is not very good. (3) Feature-based matching methods, [7][8][9][10][11][12][13][14][15] aimed at feature extraction and description, and improve the accuracy of the registration algorithm through different feature extraction algorithms and descriptions, but the parameters of some algorithms should not be adjusted, and different point clouds need to adjust different parameters; the above feature methods are not robust when the point cloud is severely occluded. In order to solve this problem, Drost et al. [16] proposed a 1.
On the basis of PPF, a keypoint extraction algorithm based on grid ISS sampling combined with curvature-adaptive sampling and angle-adaptive judgment is proposed. The algorithm has higher efficiency compared to the original PPF. 2.
Several sets of experiments are compared between K-PPF, PPF and other algorithms in public datasets, and the results prove the superiority and robustness of the K-PPF algorithm.

The Proposed Method
The K-PPF algorithm process is proposed based on the traditional PPF algorithm [16] and is used to quickly complete 3D object recognition and pose estimation. As shown in Figure 1, this algorithm is divided into offline stage and online stage. 2. Several sets of experiments are compared between K-PPF, PPF and other algorithms in public datasets, and the results prove the superiority and robustness of the K-PPF algorithm.

The Proposed Method
The K-PPF algorithm process is proposed based on the traditional PPF algorithm [16] and is used to quickly complete 3D object recognition and pose estimation. As shown in Figure 1, this algorithm is divided into offline stage and online stage.  ); finally, point-to-plane ICP for further fine registration is performed, and the 6D pose after fine registration derived (i).

Offline Stage
In the offline stage, a global model description is generated for the model point cloud and stored in a hash table.. The PPF algorithm relies on the point pair feature description. Because PPF uses raster sampling, there are redundant and useless point pairs in this sampling method, and the sampling points cannot accurately describe some angular CAD models, so the algorithm's robustness is weak, and some special shapes 3D objects may be computationally expensive and poorly recognized. So the sampling method should be improved.
Our goal is to make the point pair feature vector be invariant, discriminative, and discrete. The original point pair feature is invariant and repeatable to translation and rotation. In order to improve the distinguishability and discreteness of the point pair feature, a new key point extraction method is proposed. This keypoint extraction method can improve the discreteness of point pair feature and reduce the amount of calculation. Even in areas with large changes in edges, corners or surfaces, the key points can be well-extracted. For some point clouds with obvious shape features, the method has better recognition effect. Next, the proposed keypoint extraction algorithm is introduced. As shown  hash table Quickly vote (e,f), clustering the point clouds with high votes (g,h); finally, point-to-plane ICP for further fine registration is performed, and the 6D pose after fine registration derived (i).

Offline Stage
In the offline stage, a global model description is generated for the model point cloud and stored in a hash table. The PPF algorithm relies on the point pair feature description. Because PPF uses raster sampling, there are redundant and useless point pairs in this sampling method, and the sampling points cannot accurately describe some angular CAD models, so the algorithm's robustness is weak, and some special shapes 3D objects may be computationally expensive and poorly recognized. So the sampling method should be improved.
Our goal is to make the point pair feature vector be invariant, discriminative, and discrete. The original point pair feature is invariant and repeatable to translation and rotation. In order to improve the distinguishability and discreteness of the point pair feature, a new key point extraction method is proposed. This keypoint extraction method can improve the discreteness of point pair feature and reduce the amount of calculation. Even in areas with large changes in edges, corners or surfaces, the key points can be well-extracted. For some point clouds with obvious shape features, the method has better recognition effect. Next, the proposed keypoint extraction algorithm is introduced. As shown in Figure 1a, this paper first adopts the method of grid downsampling, performs mean processing on the points in the grid, and combines ISS sampling on the basis of the grid curvature sampling, and finally perform angle-adaptive judgment on the extracted sampling points to generate the K-PPF key points. Curvature is dependent on the concavity and convexity of the model. Using the Principal Component Analysis (PCA) [28] method based on normal vector estimation, the curvature of each data point can be estimated based on normal vector estimation. The curvature estimation method is as follows: In Equation (1) [29], λ 0 describes the change of the surface along the normal vector, while λ 1 and λ 2 represent the distribution of data points on the tangent plane. The following formula is defined as the surface variation of the data point H i in the k neighborhood.
The curvature ω i of the point cloud model at the data point can be approximated as the surface variation δ at the point, that is ω i ≈ δ.
Curvature pre-search: because of the enormous number of plane features, choosing point clouds with curvatures larger than a specified value as keypoints may significantly minimize the number of duplicate point pairs. As a result, curved points can also effectively represent corners and edges of objects, improving their matching efficiency. By setting the curvature threshold ϕ 1 , a large number of plane points can be removed, and an n dimensional vector P is established to store the pre-searched point ϕ i .
Adaptive point selection: set the threshold ϕ 2 to adaptively extract the top m points with the largest curvature from P according to the point cloud number algorithm. P s is the number of point clouds P. m = P s ϕ 2 (3)

ISS Sampling
Intrinsic Shape Signature (ISS) [30] is a method for representing solid geometric shapes. The algorithm has a wealth of geometric feature information and is capable of completing high-quality point cloud registration. Suppose the point cloud p contains n points (x i , y i , z i ), i = 1, 2, · · ·, n − 1, set p i = (x i , y i , z i ), the specific process of extracting feature points would be follows: A search radius r seek is set for each query point p i . Then, calculate the Euclidean distance between the query point p i and each point p j in the neighborhood, and the covariance matrix cov(p i ) between each query point p i and all points in the neighborhood, and set the weight ω ij .
Finally, calculate all the eigenvalues λ 1 i , λ 2 i , λ 3 i of the covariance matrix cov(p i ), and sort them in descending order. Set the thresholds δ1 and δ2 and satisfy the formula (6) to be the feature points.

Angle-Adaptive Judgment
For any point P i in the point cloud P, let the normal of P i be n i , the normal of its adjacent point P ij be n ij , and define S(P i ) as the mean of the angle between the normals of the neighborhood: where m is the number of points in the neighborhood, n is the number of point cloud sampling points, j = 1, 2, . . . , m; i = 1, 2, . . . , n. The value range of S(P i ) is [0,180], where the larger the value of S(P i ), the larger the angle between n i and n ij , as shown in Figure 2a; the smaller S(P i ) means that the angle between n i and n ij is smaller, as shown in Figure 2b. We use the global mean of the normal angle as the threshold ε 1 as the final detection condition for keypoints. Therefore, the angle determination condition based on the mean of the normal angle is as follows: count the normal angle between each point and other points, if the 90% of the point normal deviations are above ε 1 , and the sample points are reserved as the keypoint output.
cos −1 n i ·n ij n i n ij (7)

Angle-Adaptive Judgment
For any point in the point cloud , let the normal of be , the normal of its adjacent point be , and define ( ) as the mean of the angle between the normals of the neighborhood: where m is the number of points in the neighborhood, n is the number of point cloud sampling points, j = 1, 2,...,m; i = 1, 2,..., n. The value range of ( ) is [0, 180], where the larger the value of ( ), the larger the angle between and , as shown in Figure 2a; the smaller ( ) means that the angle between and is smaller, as shown in Figure 2b. We use the global mean of the normal angle as the threshold as the final detection condition for keypoints. Therefore, the angle determination condition based on the mean of the normal angle is as follows: count the normal angle between each point and other points, if the 90% of the point normal deviations are above , and the sample points are reserved as the keypoint output. The results of the proposed keypoint extraction algorithm are shown in Figure 3. It can be seen that most of the sampling points are distributed at the corners with large edges and curvature change information. For some scenes with sharp edges and corners, the extraction effect is more obvious. This paper's keypoint sampling can better represent 3D objects with fewer point pairs than grid uniform sampling. The results of the proposed keypoint extraction algorithm are shown in Figure 3. It can be seen that most of the sampling points are distributed at the corners with large edges and curvature change information. For some scenes with sharp edges and corners, the extraction effect is more obvious. This paper's keypoint sampling can better represent 3D objects with fewer point pairs than grid uniform sampling. The point pair feature [16] uses 4 parameters to describe the relative position and direction of the two orientation points, as shown in Figure 1b; the point pair feature composed of the point pair ( , , ) and its normal ( , , ) is ( , , ), which is defined as a four-element vector. The point pair feature [16] uses 4 parameters to describe the relative position and direction of the two orientation points, as shown in Figure 1b; the point pair feature composed of the point pair (m 1, , m 2 ) and its normal (n 1, , n 2 ) is F(m 1, , m 2 ), which is defined as a four-element vector.
where ∠(x, y) represents the angle between the two vectors, and d 2 represents the distance between the point pairs.

Global Feature Description
Based on the reference [16], we need to use point pair features to create global feature descriptors in the offline stage. For this, we need to compute the point pair features

Point Pair Feature
The point pair feature [16] uses 4 parameters to describe the relative position and direction of the two orientation points, as shown in Figure 1b; the point pair feature composed of the point pair ( , , ) and its normal ( , , ) is ( , , ), which is defined as a four-element vector.
where ∠( , ) represents the angle between the two vectors, and ‖ ‖ represents the distance between the point pairs.

Global Feature Description
Based on the reference [16], we need to use point pair features to create global feature descriptors in the offline stage.

Online Stage
In the online stage, we randomly select a reference point from the keypoints extracted from the scene point cloud, and all other points in the scene are paired with the reference point to create point pair features. We obtain potential matches by matching these features

Online Stage
In the online stage, we randomly select a reference point from the keypoints extracted from the scene point cloud, and all other points in the scene are paired with the reference point to create point pair features. We obtain potential matches by matching these features with model features contained in the global model description hash table. We finally use a Hough-like voting scheme to vote on the pose of the object to return the best estimated pose for coarse matching.
In the Equation (10) [16]: T s→g represents the translation and rotation transformation from the reference point in the scene to the coordinate system; T m→g represents the translation and rotation transformation from the model reference point to the coordinate system; R x (α) represents the rotation around the x-axis. The positive half-axis is rotated by an angle of α.
In the online matching stage, we calculate all point pair features of the scene point cloud, find the matching model point pair and rotation angle in the hash table, and cast a vote in the corresponding position of the voting table. This scheme is similar to generalized Hough voting, as shown in Figure 1g.

Pose Clustering
To both filter out inaccurate poses and increase the accuracy of the final findings, we cluster the recovered poses so that all poses in a cluster do not deviate in translation and rotation by more than a predetermined threshold. The score of a cluster is the total of the contained posture scores, which represent the number of votes earned in the voting system. The resultant posture is calculated by averaging the poses contained in the cluster after locating the cluster with the highest score.

Fine Registration
Because the pose produced after voting clustering may not be precise enough after K-PPF coarse registration, this study employs point-to-plane ICP for correct registration on this basis.

Point-to-Plane ICP
Low [2] devised and extensively utilized the point-to-plane ICP method. The method takes the normal of point q i in the source point cloud set Q through the point p i in the target point cloud set P as the matching point, where the corresponding distance between the two points is the distance from p i to the tangent plane where q i is located. Using the distance from the point to the tangent plane as a measure for aligning two sets of point clouds speeds up ICP convergence and assures high alignment accuracy in complicated scenarios, but the normal vector of the point cloud must be determined beforehand. Figure 5 displays the algorithm's basic architecture. indicates the corresponding normal to . For the point-to-plane ICP algorithm, the error function is designed by minimizing the sum of squared distances.

Performance Evaluation Experiments
To demonstrate the advantages of K-PPF's sampling method, we test the recognition effect of various experimental scenarios. First, the ADD-S metric [31] is used as the evalu- For the point-to-plane ICP algorithm, the error function is designed by minimizing the sum of squared distances.

Performance Evaluation Experiments
To demonstrate the advantages of K-PPF's sampling method, we test the recognition effect of various experimental scenarios. First, the ADD-S metric [31] is used as the evaluation index in the UWA dataset to compare the recognition efficiency of the K-PPF algorithm compared to PPF in complex scenes, and then the algorithm in this paper is evaluated in the Redkitchen scene with rich corners. Recall and precision [32] are used as indicators to compare the algorithm in this paper with CSHOT [33], SHOT [8], SI [10], and FPFH [9]. Finally, the recognition effect of the algorithm in this paper is shown on the Kinect dataset. Through experiments in various different scenarios, the results prove the robustness and superiority of the algorithm in this paper, and their reliability is verified by the real engine dataset. The experimental environment is VS2017 and PCL1.8 [34]; the computer configuration is Intel i5-6300HQ 2.30 GHz CPU, NVIDIA GTX960m GPU, 16 GB memory Windows 10 system environment; all algorithms use OpenMP multi-core parallel acceleration.

Datesets
This paper uses thre public datasets and real engine dataset collected using realsense d435i as tests. As shown in the Figure 6, there are Cvlab's kinect dataset [35], Princeton's redkitchen dataset [36] and UWA T-rex dataset [37]. The Minolta VIVID 910 scanner created the UWA dataset, which comprises five model point clouds and 50 scene point clouds. The Kinect dataset includes six models and 16 scenes captured by the Microsoft kinect depth camera. Data from a complex point cloud with grid quality, mild occlusion, and clutter. The Redkitchen dataset comprises 60 point clouds, each of which is a 3D surface point cloud combined from 50 depth frames using TSDF volume fusion.

Comparison with Original PPF Algorithm
In this section, we will compare our method to the original PPF algorithm using various indicators in various settings.

UWA Dataset (Complex Scene)
In order to verify the efficiency of the algorithm in this paper, on the basis of the paper [38], the average nearest point distance ( − ) [31] is used as the error estimator to compare the advantages of the algorithm in this paper compared with the PPF algorithm, the estimated pose = ( , ), and the actual pose of the dataset = ( , ), to calculate the average point . 1

Comparison with Original PPF Algorithm
In this section, we will compare our method to the original PPF algorithm using various indicators in various settings.

UWA Dataset (Complex Scene)
In order to verify the efficiency of the algorithm in this paper, on the basis of the paper [38], the average nearest point distance (ADD − S) [31] is used as the error estimator to compare the advantages of the algorithm in this paper compared with the PPF algorithm, the estimated pose P = (R , T ), and the actual pose of the dataset P = (R, T), to calculate the average point ADD.
where n is the number of points in the model point cloud, and on the basis of ADD, we use the average nearest point distance (ADD − S) metric to display the area under the accuracy threshold curve (AUC).
In the formula, x 1 and x 2 are the two closest points in the model pose and the estimated pose. The AUC is calculated by changing the threshold of the average distance. The maximum threshold in this paper is set to 5 cm, and the data is trex in the UWA dataset. First, the algorithm in this paper is tested for different keypoint sampling thresholds d, and the AUC curves under different distance thresholds are obtained. It can be seen that as the threshold d increases, the time consumption increases, and the accuracy rate also increases. This is because the larger d, the more keypoints are extracted, so the effect is best when d = 0.6, but the computational cost is too high. Finally, considering d = 0.8 in this paper, the comprehensive effect is the best.
In Figures 7 and 8, accuracy is defined as the ratio of successful recognitions to the total number of ambient point clouds at various distance thresholds. The recognition effects of 50 scenarios are then assessed for PPF and K-PPF. Figure 9 depicts a portion of the experimental data. The experimental results show that the algorithm in this paper can achieve better recognition effects in complex scenes with the same time consumption and higher efficiency than the original PPF, demonstrating the recognition advantages of K-PPF in general scenes. However, the recognition advantages of K-PPF in corner-rich scenes must be verified.

Overlap Rate Calculation
One of the key indications for 3D object identification is the average closest point distance measurement. This is appropriate for situations in which the model's true stance is known, but there may also be unknown positions in the actual setting. Therefore, this paper needs to use an overlap rate indicator. Reference [39] projects the object into the depth map and defines the overlap rate by the depth deviation of the projected pixels. If the depth deviation reaches the set threshold, the point overlaps, otherwise it does not overlap. The ratio of the number of overlapping pixels to the total projected pixels is the overlap ratio.
On the basis of reference [40], this paper proposes a new method for calculating the overlap ratio. First, the kd-tree structure of the model and the scene is established, and the scene points within the neighborhood ε r of a given radius of each model point are counted (ε r = 5 × ppi), where ppi is the point cloud density. If the number of points exceeds the set threshold µ r , the model point is considered to be coincident with the scene, and then it is judged whether the point cloud is occluded. If there is no occlusion, the overlap rate is ε model /ϕ model ; if there is occlusion, output occlusion rate and occlusion rate (∂ model − ∂ scene )/ϕ model , and overlap rate is ∂ model /∂ scene . In the formula, ε model is the number of overlapping points, ϕ model is the total number of points with the model, ∂ scene is the number of overlapping points of the scene, and ϕ model is the number of overlapping points of the model.

Redkitchen Dataset (Rich Corners)
In order to verify the recognition advantages of keypoint extraction in the K-PPF algorithm in scenes with large changes in curvature angle, we selected a building scene with rich corners for testing, and the data set used 40 sets of Redkitchen large-scale point cloud data. As shown in Figure 10, the keypoints extracted by the K-PPF algorithm differ more than the PPF algorithm, and can express the shape features of the object with fewer point pair features. We show the advantages and disadvantages of the sampling method in this paper through 40 sets of experimental data. Table 1 analyzes the average reduction of the matched point pair features in each scene. It can be seen that K-PPF has an average reduction in the scene matching process compared to PPF. Therefore, the computational cost of our algorithm is greatly reduced, and the overlap rate is also improved compared with the PPF algorithm. While buildings have sharp corners and their density varies greatly, our method can detect corners stably. Therefore, in scenes with large curvature changes, the algorithm in this paper can achieve better results in less time; in normal scenes, the algorithm in this paper can also achieve better recognition results in the same time. and then it is judged whether the point cloud is occluded. If there is no occlusion, the overlap rate is / ; if there is occlusion, output occlusion rate and occlusion rate ( − )/ , and overlap rate is / . In the formula, is the number of overlapping points, is the total number of points with the model, is the number of overlapping points of the scene, and is the number of overlapping points of the model.

Redkitchen Dataset (Rich Corners)
In order to verify the recognition advantages of keypoint extraction in the K-PPF algorithm in scenes with large changes in curvature angle, we selected a building scene with rich corners for testing, and the data set used 40 sets of Redkitchen large-scale point cloud data. As shown in Figure 10, the keypoints extracted by the K-PPF algorithm differ more than the PPF algorithm, and can express the shape features of the object with fewer point pair features. We show the advantages and disadvantages of the sampling method in this paper through 40 sets of experimental data. Table 1 analyzes the average reduction of the matched point pair features in each scene. It can be seen that K-PPF has an average reduction in the scene matching process compared to PPF. Therefore, the computational cost of our algorithm is greatly reduced, and the overlap rate is also improved compared with the PPF algorithm. While buildings have sharp corners and their density varies greatly, our method can detect corners stably. Therefore, in scenes with large curvature changes, the algorithm in this paper can achieve better results in less time; in normal scenes, the algorithm in this paper can also achieve better recognition results in the same time.

Compare with Other Algorithms
To explain the benefits of the K-PPF algorithm more clearly, we compare it to CSHOT [31], SHOT [8], SI [10], and FPFH [9]. The PCL [34] open-source library serves as the experimental environment, and the data set consists of 40 Redkitchen groups. The standards of recall and precision we use are based on the transformed pose matrix. After matching with different algorithms, the generated pose matrix is compared with the initial pose of the model. If the xyz error is lower than the threshold ϕ, they are considered to match. In the formula, the algorithm determines that the correct condition is that the overlap rate is greater than 80%; that is, it is considered a successful match. It can be seen from Figure 11 below that under different thresholds, the recall rate and precision rate of the algorithm in this paper are compared with the other four algorithms. For better results, the recall rate is 12.5% higher than CSHOT and 100% higher than SI.
formed point cloud, we count the number of correct matches as well as the number of false matches. Changing the value of obtains the curve. The results are displayed as and , which are defined as: In the formula, the algorithm determines that the correct condition is that the overlap rate is greater than 80%; that is, it is considered a successful match. It can be seen from Figure 11 below that under different thresholds, the rate and rate of the algorithm in this paper are compared with the other four algorithms. For better results, the rate is 12.5% higher than CSHOT and 100% higher than SI.

Point Cloud Recognition Experiment
To verify the point cloud recognition effect of the algorithm in this paper in complex environments and different degrees of occlusion, the kinect data set of CVlab is used to test 5 sets of data, including 5 model point cloud data and 18 scene point cloud data. As shown in Figure 12, the red bounding box is drawn for the model point cloud identified in the scene point cloud, and the point cloud recognition effect is displayed in green. It can be seen from the recognition results in Table 2 that in the 5 groups of different degrees of occlusion, the average data overlap rate is above 90%.

Point Cloud Recognition Experiment
To verify the point cloud recognition effect of the algorithm in this paper in complex environments and different degrees of occlusion, the kinect data set of CVlab is used to test 5 sets of data, including 5 model point cloud data and 18 scene point cloud data. As shown in Figure 12, the red bounding box is drawn for the model point cloud identified in the scene point cloud, and the point cloud recognition effect is displayed in green. It can be seen from the recognition results in Table 2 that in the 5 groups of different degrees of occlusion, the average data overlap rate is above 90%.

Real Dataset Experiment
In order to verify the reliability of the algorithm, after preprocessing the real scene point cloud collected by the depth camera and the engine CAD model point cloud in the laboratory, a comparison between the algorithm in this paper and the original PPF algorithm is made. As shown in Figure 13, it can be seen that the initial poses of the two point clouds are quite different, and the shapes are not exactly the same. In this case, the registration effect of the algorithm is good. Table 3 displays the findings. When compared to the PPF algorithm, our method reduces redundant point pairs by 17.4%, takes 33.3% less time, and increases the overlap rate by 1.731%. time, and increases the overlap rate by 1.731%.

Conclusions
Aiming at the shortcomings of original PPF grid sampling redundant point pairs, this paper proposes a keypoint extraction algorithm based on grid ISS sampling combined with curvature-adaptive sampling and angle-adaptive judgment (K-PPF), which have the characteristics of high efficiency and strong robustness. The algorithm improves the sampling part of the original PPF and contains less duplicate point pairs. Finally, the efficiency of the method in this study is validated by comparison with other algorithms, and the program's recognition accuracy in the complex environment and occlusion process is val-

Conclusions
Aiming at the shortcomings of original PPF grid sampling redundant point pairs, this paper proposes a keypoint extraction algorithm based on grid ISS sampling combined with curvature-adaptive sampling and angle-adaptive judgment (K-PPF), which have the characteristics of high efficiency and strong robustness. The algorithm improves the sampling part of the original PPF and contains less duplicate point pairs. Finally, the efficiency of the method in this study is validated by comparison with other algorithms, and the program's recognition accuracy in the complex environment and occlusion process is validated using the CVLab dataset. In practical applications, the proposed keypoint algorithm is easy to match and can provide a good initial pose for matching. The experimental results of real datasets also prove that the algorithm in this paper has very high efficiency and can be used for laboratory virtual-real fusion experiments in the future.
Howerver, although the paper improves the real-time performance of 3D point cloud feature extraction, it still needs further improvement for large-scale and real-time application scenarios. In the future, innovations in voting strategies can be made to reduce the time consumed by algorithms.
Author Contributions: X.S. designed and performed the experiment; Z.G. and X.S. wrote and revised the paper; Q.G., H.S., X.T. and Q.C. provided revisions; All authors have read and agreed to the published version of the manuscript.