Next Article in Journal
Knowledge Distillation in Video-Based Human Action Recognition: An Intuitive Approach to Efficient and Flexible Model Training
Previous Article in Journal
Human Attention Restoration, Flow, and Creativity: A Conceptual Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Planar Projection Contour (PPC): A Novel Algorithm for Local Feature Description in Point Clouds

1
School of Mechanical and Electrical Engineering, Xi’an Polytechnic University, Xi’an 710043, China
2
School of Mechanical Engineering, Northwestern Polytechnical University, Xi’an 710060, China
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(4), 84; https://doi.org/10.3390/jimaging10040084
Submission received: 17 February 2024 / Revised: 19 March 2024 / Accepted: 26 March 2024 / Published: 29 March 2024

Abstract

:
Local feature description of point clouds is essential in 3D computer vision. However, many local feature descriptors for point clouds struggle with inadequate robustness, excessive dimensionality, and poor computational efficiency. To address these issues, we propose a novel descriptor based on Planar Projection Contours, characterized by convex packet contour information. We construct the Local Reference Frame (LRF) through covariance analysis of the query point and its neighboring points. Neighboring points are projected onto three orthogonal planes defined by the LRF. These projection points on the planes are fitted into convex hull contours and encoded as local features. These planar features are then concatenated to create the Planar Projection Contour (PPC) descriptor. We evaluated the performance of the PPC descriptor against classical descriptors using the B3R, UWAOR, and Kinect datasets. Experimental results demonstrate that the PPC descriptor achieves an accuracy exceeding 80% across all recall levels, even under high-noise and point density variation conditions, underscoring its effectiveness and robustness.

1. Introduction

Point cloud local feature descriptors are pivotal in the computer vision field [1]. They are derived by converting local geometric information into attributes such as curvature [2], eigenvalue [3], and the density of spatial points [4]. Unlike global feature descriptors, local descriptors offer enhanced robustness against occlusions and missing local information in models. Owing to these benefits, local feature descriptors have found extensive applications in object detection [5], point cloud registration [6,7], and cultural heritage restoration [8].
Recent research by both domestic and scholars has significantly advanced the study of point cloud local feature descriptors, broadly categorizing them into two types [9]. The first type encompasses descriptors independent of the Local Reference Frame (LRF), which quantify local geometric features statistically. However, this approach often overlooks the spatial location of point clouds, leading to issues like inadequate description or reduced robustness against rotation and translation. Examples of such descriptors are the spin image (SI) [10], Fast Point Feature Histograms (FPFH) [11], and 3D Shape Context (3DSC) [12].
The second category includes descriptors that rely on an LRF, transforming neighboring points’ information into a consistent LRF to achieve spatial position invariance. This method encodes the local attributes of neighboring points to represent the query point’s local feature. Descriptors in this category, such as the Signature of Histograms of Orientations (SHOT) [13], Geometric Feature Statistics Histogram (GFSH) [14], and Weighted Height Image (WHI) [15], leverage the LRF to fully decode spatial position information, offering advantages like invariance to rigid transformations, improved stability, and enhanced expressiveness. Despite these advancements, challenges remain, including limited robustness to noise and point density variations, extensive descriptor dimensions, and decreased computational efficiency.
Motivated by the aforementioned challenges, this study introduces a novel method named Planar Projection Contour (PPC), distinguished by its enhanced descriptiveness and resilience to issues like noise and variations in point density. The PPC approach begins with the construction of an LRF through the analysis of local neighboring points around a key point. Subsequently, these neighboring points are projected onto three orthogonal planes defined by the LRF. The two-dimensional projections on each plane are then modeled into convex hull contours. By concatenating these three convex hull contours, a comprehensive feature vector is formed, resulting in the generation of the PPC descriptor. The PPC’s effectiveness is rigorously evaluated against traditional descriptors across three public datasets, with the experimental outcomes indicating superior performance in tasks such as shape retrieval and alignment. This study introduces a novel local feature descriptor named PPC, designed to offer a harmonious blend of descriptiveness, robustness, and computational efficiency. Additionally, we present a comparison method specifically tailored for evaluating feature similarity based on convex hull contours, applicable to descriptors that utilize these contours as features.
The structure of this paper is organized as follows: Section 2 provides an overview of existing 3D local feature descriptors. Section 3 details the methodology behind the proposed PPC descriptor. Section 4 discusses the experimental evaluation of PPC in comparison to traditional descriptors across three public datasets. Finally, conclusions are presented in Section 5.

2. Related Work

A substantial body of research focuses on the comprehensive analysis of point cloud local feature descriptors, with each descriptor offering a distinct representation of local information. Based on the coding space of their attributes, current descriptors are categorized into two types: those operating in the 3D space domain and those in the 2D space domain.
Three-dimensional space domain descriptors characteristically partition the local vicinity of neighboring points into multiple subspaces using varied approaches to separately encode and, thus, represent features. Salti S et al. [13] introduced the SHOT descriptor, which employs a spherical bounding box to segment the local space around a key point into 32 subspaces based on radial, azimuth, and elevation directions. It then quantifies the cosine similarities between the neighboring points and the key point in each subspace. Building on the SHOT framework, the B-SHOT descriptor was developed, utilizing a method of byte-by-byte comparison for enhancement. B-SHOT [16] was generated based on the SHOT descriptor using the method of comparing individual byte sizes one by one. Zhao H et al. [17] proposed the HoPPF descriptor, segmenting the local point pairs associated with each key point into eight regions and crafting sub-features based on the distribution of these pairs within each zone. While SHOT, B-SHOT, and HoPPF leverage histogram features, their discriminatory power is somewhat curtailed by the mutual interference of adjacent histograms. Tang K et al. [18] introduced the SGC descriptor, which constructs a cube around the key point’s local support surface, subdividing it into evenly distributed grids to count geometric centroid values, vertex coordinates, and vertex numbers. Quan S et al. [19] put forward the LOVS descriptor, creating a uniformly gridded cube from the key point’s local surface and dividing it into a set of uniformly distributed voxels. Both SGC and LOVS approach the local surface construction of point clouds through cubic segmentation, which often leads to largely unoccupied meshes in edge areas, reflecting a lower spatial utilization.
Two-dimensional space domain descriptors transform 3D shapes into 2D space through projection, simplifying the 3D coordinate information of a point cloud into features like images and contours, which are then encoded. Johnson A E et al. [10] introduced the spin image (SI) descriptors, which create a 2D grayscale map by projecting neighboring points onto horizontal and vertical 2D planes using the normal vector of the key point as a reference axis. Guo Y et al. [20] developed the Rotational Projection Statistics (ROPS) descriptor, which involves continuously rotating the local surface within the LRF and projecting the surface after each rotation to calculate point density statistics. Additionally, Guo Y et al. [21] proposed the triple spin images (TriSI) descriptor, achieving a comprehensive description by generating three spin images around different axes. Yang J et al. [22] introduced the Triple Orthogonal Local Depth Images (TOLDI) descriptor, which calculates the LRF for the point cloud’s local surface using normal vectors and weighted projection vectors, effectively decoding rotation invariance and spatial information. Further, Yang J et al. [23] proposed the Rotational Contour Signatures (RCS) descriptor, characterizing the 3D shape through multi-view information obtained via rotation and by counting the 2D contours at each viewpoint. The method of mapping 3D local coordinate information into 2D simplifies descriptor computational complexity and accelerates encoding. However, this approach may result in significant loss of local information, leading to reduced robustness against noise, point density variations, and other factors.
In light of the analyzed strengths and weaknesses of 2D and 3D space domain descriptors, this work proposes the PPC descriptor, which enhances robustness by encoding convex hull contours of 2D projection points from three orthogonal views. This strategy not only preserves high descriptiveness but also simplifies the encoding of point cloud local information, addressing the common robustness issues found in many 2D space domain descriptors.

3. Construction of PPC

3.1. Overall Construction Process

The process of constructing the PPC descriptor unfolded in several key steps. Initially, the LRF was determined by analyzing the covariance between the query point and its neighboring points. Subsequent to this, the neighboring points were projected onto the LRF’s three orthogonal planes. The culmination of this process involved fitting the 2D projection points on each plane into convex hull contours. The PPC descriptor was then formulated through the systematic concatenation of these contours. A visual representation of the PPC construction workflow is provided in Figure 1.

3.2. Construction of LRF

The LRF played a crucial role in embedding spatial position information into the local feature representation of a point cloud. This spatial information, once anchored to the LRF, became invariant to rotation and translation, significantly enhancing the descriptor’s robustness and utility. The method used to establish the LRF included precise calculations for the X-axis, Y-axis, and Z-axis components.
The Z-axis was computed as the normal vector of key point P, which could be obtained via covariance on the neighboring points Pi of P [13].
To determine the the X-axis, the neighboring points Pi were projected on the tangent plane defined by the Z-axis. For each projected point, a projection vector Vi was computed, which is shown in Formula (1).
V i = P P i ( P P i Z ) Z
where P denotes the key point, and Pi denote neighboring points. Z denotes the Z-axis vector.
We assigned weights W i 1 according to the Euclidean distances from Pi to P. We also assigned weights W i 2 based on the projection distances from Pi to the tangent plane of the Z-axis. Then, the weighted projection vectors were summed and normalized to obtain the X-axis, which is shown in Formulas (2) and (3).
X = i = 1 K W i 1 W i 2 V i / i = 1 K W i 1 W i 2 V i
W i 1 = ( r p i p ) 2 W i 2 = ( P P i Z ) 2
where r denotes search radius of the key point. K denotes the number of neighboring points.
The Y-axis was determined by calculating the cross-product between the Z-axis and the X-axis. This method ensured that the Y-axis was orthogonal to both the Z-axis and X-axis, thereby completing the orthogonal basis of the LRF and establishing a robust coordinate system for representing the point cloud’s local geometric features.

3.3. Construction of PPC

In Section 3.1, the construction of the LRF segregated the local support plane, created by the key point and its neighboring points, into three orthogonal planes: the X-Y plane, the Y-Z plane, and the X-Z plane. This division facilitated a comprehensive analysis of the point cloud’s local geometry from multiple perspectives, as depicted in Figure 2.
The projection points on the three planes were obtained by projecting the neighboring points Pi onto the X-Y plane, Y-Z plane, and X-Z plane, respectively. This process was mathematically represented in Formula (4), which provided a systematic approach for capturing the local geometric characteristics of the point cloud from different orientations, ensuring a thorough and multidimensional feature extraction.
J x y = ( x , y ) | ( x , y , z ) P i J y z = ( y , z ) | ( x , y , z ) P i J x z = ( x , z ) | ( x , y , z ) P i
where Jxy, Jyz, and Jxz denote the set of projected points of Pi in the X-Y plane, Y-Z plane, and X-Z plane, respectively.
The sets Jxy, Jyz, and Jxz contained a number of points equal to the original count of neighboring points, resulting from the transformation of 3D coordinate information into 2D information through projection. This transformation, while simplifying the data, did not remove the inherent vulnerabilities to noise, variations in point density, missing information, and other perturbations common to both 3D and 2D coordinate systems. To mitigate these vulnerabilities, this study employed the convex hull algorithm to transform the projected 2D point sets Jxy, Jyz, and Jxz into convex hull contours [24], as elaborated in Formula (5). This approach effectively summarized the spatial distribution of points while enhancing the robustness of the descriptor against the aforementioned factors.
C x y = C O N V E X { p 1 , p 2 , p n | p J x y } C y z = C O N V E X { p 1 , p 2 , p n | p J y z } C x z = C O N V E X { p 1 , p 2 , p n | p J x z }
where Cxy, Cyz, and Cxz denote the convex hull contour of Jxy, Jyz, and Jxz, respectively. P1, P2, … Pn denote the edge points of the set of projected points. CONVEX denotes the convex hull algorithm.
The adoption of convex hull contours for representing the 2D projected point set offers significant benefits, which are outlined as follows:
(1)
Utilizing convex hull contours allowed for the internal characteristics of the point set to be disregarded, substantially simplifying the representation of the 2D point set’s geometric information. This simplification led to an improvement in computational efficiency by focusing on the external boundary of the point distribution, as demonstrated in Figure 3a.
(2)
Convex hull contours exhibited increased stability when faced with noise interference, in contrast to raw coordinate information. By encapsulating the outermost points, the convex hull effectively minimized the impact of outliers or noise within the data, ensuring a more consistent representation, as illustrated in Figure 3b.
(3)
The representation via convex hull contours proved to be more resilient to variations in point density. Unlike methods that rely on the detailed arrangement of points, the convex hull approach maintained a consistent outline, regardless of the density of points within the contour. This robustness was critical for ensuring reliable feature extraction across datasets with varying point densities, as shown in Figure 3c.
Finally, the Cxy, Cyz, and Cxz of the three planes were encoded as the local features of P, forming the PPC descriptor.
D = C x y , C y z , C x z

3.4. Feature Matching of PPC

Ideally, the local features of corresponding points between the model point cloud and the scene point cloud should match perfectly. However, in practice, differences in the local features of these corresponding points frequently arose due to factors such as noise interference and missing local information. These discrepancies can challenge the process of accurately matching and recognizing patterns within and between point clouds, underscoring the importance of developing robust feature descriptors capable of mitigating the effects of such imperfections. Hence, it was essential to obtain the optimal feature matching degree by utilizing descriptors to extract the local features of the point cloud, thereby establishing pairs of points based on feature matching. This approach aimed to maximize the similarity between corresponding features, even in the presence of noise and incomplete data, facilitating accurate point cloud registration and recognition tasks. Given a certain pair of feature points Ti and Tj, the local features Di and Dj were obtained based on PPC. Then, the matching degree of the convex hull contour of both corresponding to the three planes was obtained by accumulating the matching degree of the convex hull contour of the corresponding three planes. The details are shown in Formulas (7) and (8).
M 1 = S x y Δ / ( S x y i + S x y j S x y Δ ) M 2 = S y z Δ / ( S y z i + S y z j S y z Δ ) M 3 = S x z Δ / ( S x z i + S x z j S x z Δ )
M ( D i , D j ) = i = 1 3 M i
where M1, M2, and M3 denote the feature matching values corresponding to the convex hull contours, respectively. S x y Δ , S y z Δ , and S x z Δ denote the overlap area of the convex hull contours. S x y i , S y z i , and S x z i are the areas of the X-Y plane, Y-Z plane, and X-Z plane convex hull contours of the feature point Ti. S x y j , S y z j , and S x z j are the areas of convex hull contours of the X-Y plane, Y-Z plane, and X-Z plane of the feature point Tj. M ( D i , D j ) denotes the total feature matching of descriptor Di and Dj.
To demonstrate the resilience of convex hull contour features against noise and variations in point density, an experiment was conducted where a pair of feature points—one from the model point cloud and one from the scene point cloud—was selected. The experiment tested the matching degree of planar convex hull contours subjected to Gaussian noise and downsampling, respectively. This procedure aimed to highlight the robustness of these features under conditions that typically challenge the integrity of point cloud data. The results of this testing are illustrated in Figure 4, providing visual evidence of the effectiveness of convex hull contours in maintaining reliable feature matching in less-than-ideal data quality scenarios.

4. Performance Testing of PPC

In this section, the effectiveness of the PPC descriptor is evaluated through a series of tests. Initially, the dataset and evaluation criteria for the experiment are carefully selected. Following this, an analysis is conducted of how key parameters of the PPC influence the descriptor’s performance, utilizing the established criteria. Subsequently, the resilience of PPC and several traditional descriptors, namely TOLDI [20], LOVS [23], SHOT [13], B-SHOT [16], and SGC [22], to noise and variations in point density is assessed using the chosen dataset. Additionally, the computational efficiency of these descriptors is examined. All experiments conducted in this study leverage the capabilities of the open3d [25] point cloud data processing library, ensuring a standardized and reproducible framework for evaluation.

4.1. Datasets and Standards

The datasets used in this experiment include the B3R dataset [25], the UWAOR dataset [26], and the Kinect dataset [27]. A partial dataset is shown in Figure 5.
The RP curve, a recognized benchmark in descriptor performance analysis, offers insights into the trade-off between recall (the ability to retrieve relevant instances) and precision (the accuracy of the retrieved instances). The methodology for calculating the RP curve is detailed below, following the protocols outlined in prior studies [17].
The process begins with the model point cloud, the scene point cloud, and the transformation matrix that defines the relationship between them. From the model point cloud, 1000 points are randomly selected as model key points. For each model key point, the point within the scene point cloud that is closest in terms of Euclidean distance is identified as its corresponding scene key point, forming a pair of key points. The PPC features for each key point pair are then extracted and compared. The model key point features are matched against all features in the scene point cloud to find the nearest and the second nearest feature matches based on feature similarity.
A threshold is applied to the ratio of the distance to the nearest feature vs. the distance to the next nearest feature. If this ratio falls below the set threshold, the model and scene key points are considered a matching pair M. A correct match TM is identified if the Euclidean distance between the matched points is below a certain criterion, indicating a successful match; otherwise, it is classified as a false match FM. By varying the threshold, a series of data points is generated, from which the RP curve is constructed.
Recall is defined as the proportion of true positive matches out of all positive instances in the data, while 1 − Precision reflects the rate of false positive matches in relation to all retrieved instances. Mathematically, these metrics provide a comprehensive overview of a descriptor’s performance across different levels of stringency in feature matching, highlighting the PPC’s ability to identify and correctly match point cloud features under various conditions.
R e c a l l = l e n g t h ( T M ) l e n g t h ( P a i r )
1 P r e c i s i o n = l e n g t h ( F M ) l e n g t h ( M )
The support radius r is a key parameter involved in defining the performance of the PPC, which determines the information of the neighborhood points, thus affecting the convex hull contours. In the following, r is analyzed in detail. Based on the B3R dataset, after adding the 0.1 mr (mr denotes mesh resolution) Gaussian noise to the scene point cloud and 1/2 downsampling, the performances of the PPC with r = 10 mr, 15 mr, 20 mr, 25 mr, and 30 mr are tested, respectively. The results are shown in Figure 6.
Figure 6 reveals that the performance of the Planar Projection Contour (PPC) descriptor is suboptimal at support radii of r = 10 mr and r = 15 mr. This observation suggests that smaller values of r result in fewer neighboring points being included within the descriptor’s calculation range. Consequently, this scarcity of neighborhood points leads to diminished robustness in the convex hull contours derived from these points. As the support radius r increases, the descriptor incorporates a larger pool of neighboring points, enhancing the stability of the convex hull contours. This expansion directly contributes to improved feature matching capabilities of the PPC descriptor. Considering this analysis, a support radius of r = 10 mr is identified as the optimal setting for constructing the PPC, striking a balance between the inclusivity of neighborhood information and computational efficiency.
In Section 4.2, Section 4.3 and Section 4.4 comparative tests are conducted between PPC and other established descriptors, including SGC, LOVS, SHOT, B-SHOT, and TOLDI. These comparisons aim to highlight the distinctive advantages and performance characteristics of PPC in relation to its peers. The specific parameters utilized for each descriptor in the comparative analysis are detailed in Table 1, providing a clear framework for understanding the experimental setup and the basis for performance evaluation.

4.2. Robustness Testing against Gaussian Noise

To assess the Planar Projection Contour (PPC) descriptor’s resilience against Gaussian noise, this study utilizes the B3R dataset to conduct feature matching performance tests on the scene point cloud under various conditions: without noise and with Gaussian noise with standard deviations of 0.1 mr, 0.3 mr, and 0.5 mr. These results are then benchmarked against established descriptors such as SGC [18], LOVS [19], SHOT [13], B-SHOT 16, and TOLDI [22]. The corresponding Recall vs. 1 − Precision (RP) curves generated from these experiments are displayed in Figure 7a–d. Further, the robustness of PPC to Gaussian noise is examined using the UWAOR and Kinect datasets, with a Gaussian noise of 0.3 mr applied to the scene point clouds.
Analysis of Figure 7a–d reveals that under a low-noise scenario (0.1 mr), descriptors such as SGC, PPC, LOVS, and TOLDI demonstrate robust feature matching capabilities. Conversely, SHOT and B-SHOT exhibit a marked decline in performance as noise levels escalate. While the effectiveness of SHOT and B-SHOT significantly diminishes with increased Gaussian noise, the remaining descriptors experience only marginal performance reductions, maintaining commendable feature matching capabilities. Among these, PPC ranks second, slightly trailing behind SGC. Additionally, as illustrated in Figure 7e–f, PPC continues to exhibit superior feature matching performance on both the UWAOR and Kinect datasets under conditions of added noise. Collectively, these findings underscore PPC’s substantial robustness in the face of noise interference, highlighting its potential applicability in diverse 3D computer vision tasks under varying environmental conditions.

4.3. Robustness Test for Point Density Variation

The robustness of the Planar Projection Contour (PPC) descriptor against variations in point density was evaluated through a series of tests. These tests involved assessing the feature matching performance of the scene point cloud under different downsampling levels (1/2, 1/4, and 1/8) using the B3R dataset, and the results were compared against those of established descriptors such as SGC, LOVS, SHOT, B-SHOT, and TOLDI. The experimentally generated Recall vs. 1 − Precision (RP) curves are displayed in Figure 8a–c. Further analysis was conducted on the UWAOR and Kinect datasets with a 1/4 downsampling rate, and the corresponding RP curves are depicted in Figure 8d–e. A comprehensive test to gauge PPC’s resilience to both Gaussian noise and point density variations was also performed by applying 1/4 downsampling to the scene point cloud from the B3R dataset, followed by the addition of 0.3 pr Gaussian noise. The RP curves from this experiment are illustrated in Figure 8f.
Observations from Figure 8a–c indicate that PPC consistently outperforms other descriptors across various downsampling levels on the B3R dataset, closely followed by LOVS and SGC. TOLDI’s performance was found to be slightly inferior to that of SGC, while SHOT and B-SHOT exhibited greater performance fluctuations. Additionally, PPC maintained the highest feature matching performance on the UWAOR and Kinect datasets, as evident from Figure 8d–e, suggesting that the convex hull contour’s feature representation offers considerable advantages in handling point density variations. Furthermore, Figure 8f demonstrates that PPC’s performance even remains superior under the combined influence of Gaussian noise and point density variations, affirming its robustness against these common challenges.
In summary, PPC’s architecture, centered around the convex hull contour feature representation, provides exceptional resilience to point density variations, establishing its robustness and effectiveness for various 3D computer vision applications.

4.4. PPC Calculation Efficiency Test

To assess the computational efficiency of the Planar Projection Contour (PPC) descriptor, it was benchmarked against other established descriptors, including SGC [18], LOVS [19], SHOT [13], B-SHOT [16], and TOLDI [22]. This evaluation focused on measuring the total time required for representing the local features of 1000 key points across varying descriptor radius parameters, with radii including 10 mr, 15 mr, 20 mr, 25 mr, and 30 mr. The efficiency of each descriptor was quantified based on the total time consumed for feature representation, with the findings illustrated in Figure 9.
The analysis revealed that all descriptors exhibit high computational speed at smaller radii. However, as the radius r increases, the complexity involved in establishing the Local Reference Frame (LRF) and the extended dimensions of the descriptor result in a notable increase in computation time for TOLDI. Similarly, LOVS experiences a significant surge in computation time. In contrast, the increases in computation time for SGC, SHOT, and B-SHOT are more gradual. When compared to SGC, SHOT, and B-SHOT, PPC demonstrates superior performance, showcasing its enhanced computational efficiency.
This comparison highlights PPC’s advantage in terms of computational efficiency, especially as the complexity of the descriptor’s calculation increases with larger radii. PPC’s ability to maintain a lower computation time while ensuring accurate and robust feature representation makes it an attractive option for applications requiring efficient processing of point cloud data.

4.5. Application to 3D Matching

The effectiveness of the Planar Projection Contour (PPC) descriptor was further validated through 3D matching tests using the dataset mentioned earlier. These tests adhered to the matching process outlined in [23], with a notable distinction: the point cloud description task was conducted using the PPC descriptor, and the method for comparing descriptor similarity was the one proposed in this study. The dataset used for testing included a diverse range of point cloud types, spanning from complete to localized formations and varying from high to low quality. The outcomes of these matching tests are visually documented in Figure 10.
Analysis of the experimental results reveals that the point cloud matching techniques employing the PPC descriptor successfully matched pairs of point cloud slices across all tested scenarios. Specifically, the PPC descriptor demonstrated its ability to form robust feature matching connections, leading to superior matching results in point clouds of both low and high quality. This performance underscores the PPC’s highly descriptive nature, highlighting its effectiveness in accurately capturing and representing the nuanced geometric information present in point clouds. Such attributes make the PPC an invaluable tool for 3D matching tasks, especially in applications where precision and reliability are paramount.

5. Conclusions

In this study, we introduce the Planar Projection Contour (PPC) descriptor, a novel local feature descriptor for point clouds. This descriptor enhances the robustness and descriptiveness of feature matching by leveraging Local Reference Frames (LRFs), convex hull contour extraction, and a unique method for assessing the matching degree of PPC descriptors. The process encompasses three primary steps:
(1)
For a given key point, its nearest neighbors are identified. The Z-axis is derived through a weighted covariance analysis based on the spatial relationship between these neighboring points and the key point. The X-axis is then determined by the sum of weighted projections of the neighborhood points onto a plane, leading to the construction of the LRF.
(2)
The neighborhood points are projected onto three orthogonal planes defined by the LRF, representing the local surface interaction between the key point and its neighbors. These 2D projection points are then modeled into convex hull contours, which succinctly capture the essential geometric characteristics of the local point cloud structure.
(3)
The feature matching process involves extracting the convex hull contours from the three orthogonal planes and computing the overlapping areas of corresponding PPC contours. The matching degree is determined by accumulating these areas, with the highest accumulation signifying the optimal feature match between PPC descriptors.
This methodology demonstrates the PPC descriptor’s ability to effectively match point cloud features across varying conditions, such as noise interference and point density variations. The PPC descriptor’s design, focusing on the geometric fidelity of the point cloud data and its computational efficiency, showcases its potential as a powerful tool for applications in 3D computer vision and point cloud analysis.

Author Contributions

Conceptualization, W.T. and Y.L.; Methodology, W.T. and Y.C.; Validation, L.Z. and R.W.; Writing—original draft, W.T. and Y.L.; Writing—review and editing, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded by the National Natural Science Foundation of China under Grant 52105559, the Aviation Science Foundation under Grant 2022Z050111001, the Xi’an Science and Technology program under Grant 23GXFW0022, Key Research and Development Program of Shaanxi under Grant 2024GX-YBXM-278, and the Graduate Innovation Foundation of Xi’an Polytechnic University under Grant chx2023005.

Data Availability Statement

The original data presented in the study are openly available in https://graphics.stanford.edu/data/3Dscanrep/ (accessed on 27 March 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shi, W.; Dai, B.; Shen, W.; Sun, Y.; Zhao, K.; Zhang, Y. Automatic estimation of dairy cow body condition score based on attention-guided 3D point cloud feature extraction. Comput. Electron. Agric. 2023, 206, 107666. [Google Scholar] [CrossRef]
  2. Tian, M.; Lu, J.; Zhang, L. Curvature Estimation on Point Cloud Using an Indicator Function. In Proceedings of the HPCCT & BDAI 2020: 2020 4th High Performance Computing and Cluster Technologies Conference & 2020 3rd International Conference on Big Data and Artificial Intelligence, Qingdao, China, 3–6 July 2020; pp. 24–28. [Google Scholar]
  3. Xiang, M.; Ye, H.; Yang, B.; Cao, F. Multi-space and detail-supplemented attention network for point cloud completion. Appl. Intell. 2023, 53, 14971–14985. [Google Scholar] [CrossRef]
  4. Cao, R.; Zhang, Y.; Liu, X.; Zhao, Z. 3D building roof reconstruction from airborne LiDAR point clouds: A framework based on a spatial database. Int. J. Geogr. Inf. Sci. 2017, 31, 1359–1380. [Google Scholar] [CrossRef]
  5. Sun, B.; Li, W.; Liu, H.; Wang, P.; Gao, S.; Feng, P. Mathematical Method for Lidar-based Obstacle Detection of Intelligent Vehicle. IAENG Int. J. Comput. Sci. 2021, 48, 1–9. [Google Scholar]
  6. Pei, H.; Zhou, W.; Zhang, P.; Luo, M. A review of point set registration: From fundamental algorithms to geometric quality inspection of aviation complex parts. J. Adv. Manuf. Sci. Technol. 2023, 3, 2023012. [Google Scholar] [CrossRef]
  7. Gao, J.; Zhang, Y.; Liu, Z.; Li, S. HDRNet: High-Dimensional Regression Network for Point Cloud Registration. Comput. Graph. Forum J. Eur. Assoc. Comput. Graph. 2023, 42, 33–46. [Google Scholar] [CrossRef]
  8. Andriasyan, M.; Moyano, J.; Nieto-Julián, J.E.; Antón, D. From Point Cloud Data to Building Information Modelling: An Automatic Parametric Workflow for Heritage. Remote Sens. 2020, 12, 1094. [Google Scholar] [CrossRef]
  9. Lai, B.; Liu, W.; Wang, C.; Fan, X.; Lin, Y.; Bian, X.; Wu, S.; Cheng, M.; Li, J. 2D3D-MVPNet: Learning cross-domain feature descriptors for 2D-3D matching based on multi-view projections of point clouds. Appl. Intell. 2022, 52, 14178–14193. [Google Scholar] [CrossRef]
  10. Johnson, A.E.; Hebert, M. Using Spin Images for Efficient Object Recognition in Cluttered 3D Scenes. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 433–449. [Google Scholar] [CrossRef]
  11. Rusu, B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the IEEE International Conference on Robotics & Automation, Kobe, Japan, 12–17 May 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 3212–3217. [Google Scholar]
  12. Frome, A.; Huber, D.; Kolluri, R.; Bülow, T.; Malik, J. Recognizing objects in range data using regional point descriptors. In Proceedings of the European Conference on Computer Vision, Prague, Czech Republic, 11–14 May 2004; pp. 224–237. [Google Scholar]
  13. Salti, S.; Tombari, F.; Stefano, L.D. SHOT: Unique signatures of histograms for surface and texture description. Comput. Vis. Image Underst. 2014, 125, 251–264. [Google Scholar] [CrossRef]
  14. Hao, L.; Wang, H. Geometric feature statistics histogram for both real-valued and binary feature representations of 3D local shape. Image Vis. Comput. 2022, 117, 0262–8856. [Google Scholar] [CrossRef]
  15. Sun, T.; Liu, G.; Liu, S.; Meng, F.; Zeng, L.; Li, R. An efficient and compact 3D local descriptor based on the weighted height image. Inf. Sci. 2020, 520, 209–231. [Google Scholar] [CrossRef]
  16. Prakhya, S.M.; Liu, B.; Lin, W. B-SHOT: A binary feature descriptor for fast and efficient keypoint matching on 3D point clouds. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots & Systems, Hamburg, Germany, 28 September–2 October 2015; pp. 1929–1934. [Google Scholar]
  17. Zhao, H.; Tang, M.; Ding, H. HoPPF: A novel local surface descriptor for 3D object recognition. Pattern Recognit. 2020, 103, 0031–3203. [Google Scholar] [CrossRef]
  18. Tang, K.; Peng, S.; Chen, X. Signature of geometric centroids for 3D local shape description and partial shape matching. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; pp. 311–326. [Google Scholar]
  19. Quan, S.; Ma, J.; Hu, F.; Fang, B.; Ma, T. Local voxelized structure for 3D binary feature representation and robust registration of point clouds from low-cost sensors. Inf. Sci. 2018, 444, 153–171. [Google Scholar] [CrossRef]
  20. Guo, Y.; Sohel, F.; Bennamoun, M.; Lu, M.; Wan, J. Rotational projection statistics for 3D local surface description and object recognition. Int. J. Comput. Vis. 2013, 105, 63–86. [Google Scholar] [CrossRef]
  21. Guo, Y.; Sohel, F.; Bennamoun, M.; Wan, J.; Lu, M. A novel local surface feature for 3D object recognition under clutter and occlusion. Inf. Sci. 2015, 293, 196–213. [Google Scholar] [CrossRef]
  22. Yang, J.; Zhang, Q.; Xiao, Y.; Cao, Z. TOLDI: An effective and robust approach for 3D local shape description. Pattern Recognit. 2017, 65, 175–187. [Google Scholar] [CrossRef]
  23. Yang, J.; Zhang, Q.; Xian, K.; Xiao, Y.; Cao, Z. Rotational contour signatures for both real-valued and binary feature representations of 3D local shape. Comput. Vis. Image Underst. 2017, 160, 133–147. [Google Scholar] [CrossRef]
  24. Pang, J.; Zhang, J.; Li, Y.; Sun, W. A marker-less assembly stage recognition method based on segmented projection contour. Adv. Eng. Inform. 2020, 46, 101149. [Google Scholar] [CrossRef]
  25. Tombari, F.; Salti, S.; Stefano, L.D. Unique signatures of histograms for local surface description. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
  26. Mian, A.; Bennamoun, M.; Owens, R. On the repeatability and quality of keypoints for local feature-based 3d object retrieval from cluttered scenes. Int. J. Comput. Vis. 2010, 89, 348–361. [Google Scholar] [CrossRef]
  27. Tombari, F.; Salti, S.; Di Stefano, L. Performance evaluation of 3D keypoint detectors. Int. J. Comput. Vis. 2013, 102, 198–220. [Google Scholar] [CrossRef]
Figure 1. PPC descriptor construction flowchart.((a), Object and key point. (b), LRF was constructed based on key Point and neighboring points. (c), Projection points of neighboring points onto three planes. (d), Contours of neighboring points. (e), Concatenation of convex hull contours).
Figure 1. PPC descriptor construction flowchart.((a), Object and key point. (b), LRF was constructed based on key Point and neighboring points. (c), Projection points of neighboring points onto three planes. (d), Contours of neighboring points. (e), Concatenation of convex hull contours).
Jimaging 10 00084 g001
Figure 2. Orthogonal views of three planes. (Light red, light blue, and light green planes represent the X-Z, X-Z, and Y-Z planes, respectively).
Figure 2. Orthogonal views of three planes. (Light red, light blue, and light green planes represent the X-Z, X-Z, and Y-Z planes, respectively).
Jimaging 10 00084 g002
Figure 3. Convex hull contour comparison schematic.((a), Original convex hull contour. (b), Convex hull contour after adding Gaussian noise. (c), Convex hull contour after adding down sampling).
Figure 3. Convex hull contour comparison schematic.((a), Original convex hull contour. (b), Convex hull contour after adding Gaussian noise. (c), Convex hull contour after adding down sampling).
Jimaging 10 00084 g003
Figure 4. Add noise or downsampled convex matching.((a,b) The contours and convex hulls under original condition, with added Gaussian noise, and after down sampling. (c), Comparison of the overlapping areas between the original convex hull contours and the convex hull contours after adding Gaussian noise, as well as after down sampling).
Figure 4. Add noise or downsampled convex matching.((a,b) The contours and convex hulls under original condition, with added Gaussian noise, and after down sampling. (c), Comparison of the overlapping areas between the original convex hull contours and the convex hull contours after adding Gaussian noise, as well as after down sampling).
Jimaging 10 00084 g004
Figure 5. Example of point cloud dataset.((a), B3R dataset. (b), UWA dataset. (c), Kinect dataset).
Figure 5. Example of point cloud dataset.((a), B3R dataset. (b), UWA dataset. (c), Kinect dataset).
Jimaging 10 00084 g005
Figure 6. Support radius r parameter selection.
Figure 6. Support radius r parameter selection.
Jimaging 10 00084 g006
Figure 7. Robustness testing against Gaussian noise. ((a),B3R dataset without Gaussian noise. (b), B3R dataset with 0.1 mr Gaussian noise. (c), B3R dataset with 0.3 mr Gaussian noise. (d), B3R dataset with 0.5 mr Gaussian noise. (e), UWA dataset with 0.3 mr Gaussian noise. (f), Kinect dataset with 0.3 mr Gaussian noise).
Figure 7. Robustness testing against Gaussian noise. ((a),B3R dataset without Gaussian noise. (b), B3R dataset with 0.1 mr Gaussian noise. (c), B3R dataset with 0.3 mr Gaussian noise. (d), B3R dataset with 0.5 mr Gaussian noise. (e), UWA dataset with 0.3 mr Gaussian noise. (f), Kinect dataset with 0.3 mr Gaussian noise).
Jimaging 10 00084 g007
Figure 8. Robustness testing against point density variations. ((a), B3R dataset with 1/2 mesh decimation. (b), B3R dataset with 1/4 mesh decimation. (c), B3R dataset with 1/8 mesh decimation. (d), UWA dataset with 1/4 mesh decimation. (e), Kinect dataset with 1/4 mesh decimation. (f), B3R dataset with 0.3 mr Gaussian noise and 1/4 mesh decimation).
Figure 8. Robustness testing against point density variations. ((a), B3R dataset with 1/2 mesh decimation. (b), B3R dataset with 1/4 mesh decimation. (c), B3R dataset with 1/8 mesh decimation. (d), UWA dataset with 1/4 mesh decimation. (e), Kinect dataset with 1/4 mesh decimation. (f), B3R dataset with 0.3 mr Gaussian noise and 1/4 mesh decimation).
Jimaging 10 00084 g008
Figure 9. Descriptors’ computational efficiency tests.
Figure 9. Descriptors’ computational efficiency tests.
Jimaging 10 00084 g009
Figure 10. Example of 3D matching results. (The model point cloud and the scene point cloud are represented by red and blue respectively, and the green line represents the feature matching relationship between them).
Figure 10. Example of 3D matching results. (The model point cloud and the scene point cloud are represented by red and blue respectively, and the green line represents the feature matching relationship between them).
Jimaging 10 00084 g010
Table 1. Descriptor specific parameters.
Table 1. Descriptor specific parameters.
DescriptorRadiusParametersDimensionType
SHOT20 mr32 × 11352Float
SGC20 mr8 × 8 × 8 × 21024Float
LOVS20 mr9 × 9 × 9729Binary
B-SHOT20 mr32 × 11352Binary
TOLDI20 mr20 × 20 × 31200Float
PPC20 mr1 × 33Convex
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, W.; Lv, Y.; Chen, Y.; Zheng, L.; Wang, R. Advanced Planar Projection Contour (PPC): A Novel Algorithm for Local Feature Description in Point Clouds. J. Imaging 2024, 10, 84. https://doi.org/10.3390/jimaging10040084

AMA Style

Tang W, Lv Y, Chen Y, Zheng L, Wang R. Advanced Planar Projection Contour (PPC): A Novel Algorithm for Local Feature Description in Point Clouds. Journal of Imaging. 2024; 10(4):84. https://doi.org/10.3390/jimaging10040084

Chicago/Turabian Style

Tang, Wenbin, Yinghao Lv, Yongdang Chen, Linqing Zheng, and Runxiao Wang. 2024. "Advanced Planar Projection Contour (PPC): A Novel Algorithm for Local Feature Description in Point Clouds" Journal of Imaging 10, no. 4: 84. https://doi.org/10.3390/jimaging10040084

APA Style

Tang, W., Lv, Y., Chen, Y., Zheng, L., & Wang, R. (2024). Advanced Planar Projection Contour (PPC): A Novel Algorithm for Local Feature Description in Point Clouds. Journal of Imaging, 10(4), 84. https://doi.org/10.3390/jimaging10040084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop