Next Article in Journal
PIMA: Parameter-Shared Intelligent Media Analytics Framework for Low Resource Languages
Next Article in Special Issue
CasOmniMVS: Cascade Omnidirectional Depth Estimation with Dynamic Spherical Sweeping
Previous Article in Journal
Evaluation of the Effects of a 50 Hz Electric Field on Brain Tissue by Immunohistochemical Method, and on Blood Tissue by Biochemical, Physiological and Comet Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CA-CGNet: Component-Aware Capsule Graph Neural Network for Non-Rigid Shape Correspondence

Beijing Key Lab of Petroleum Data Mining, Department of Computer Science and Technology, China University of Petroleum, Beijing 102249, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(5), 3261; https://doi.org/10.3390/app13053261
Submission received: 19 January 2023 / Revised: 10 February 2023 / Accepted: 2 March 2023 / Published: 3 March 2023
(This article belongs to the Special Issue Applications of Computer Vision in 3D Perception)

Abstract

:
3D non-rigid shape correspondence is significant but challenging in computer graphics, computer vision, and related fields. Although some deep neural networks have achieved encouraging results in shape correspondence, due to the complexity of the local deformation of non-rigid shapes, the ability of these networks to identify the spatial changes of objects is still insufficient. In this paper, we design a Component-aware Capsule Graph Network (CA-CGNet) to further address the features of embedding space based on the component constraints. Specifically, the dynamic clustering strategy is used to classify the features of patches produced by over-segmentation in order to further reduce noise interference. Moreover, aiming at the problem that existing routing ignores the embedding relationship between capsules, we propose a component-aware capsule graph routing to fully describe the relationship between capsules, which regards capsules as nodes in the graph network and constrains nodes through component information. Then, a knowledge distillation strategy is introduced to improve the convergence speed of the network by decreasing the parameters while maintaining accuracy. Finally, a component pair constraint is added to the functional map, and the component-based semantic loss function is proposed, which can compute isomeric in both direct and symmetric directions. The experimental results show that CA-CGNet improves by 10.21% compared with other methods, indicating the accuracy, generalization, and efficiency of our method on the FAUST, SCAPE, TOSCA, and KIDS datasets.

1. Introduction

With the continuous development of computer 3D imaging technology, shape correspondence has become an important research direction in computer vision, which is closely related to object recognition, 3D reconstruction, image retrieval, image analysis, and other problems. The shape correspondence problem can be summarized as identifying homologous points of two or more shapes, where a large number of variables are needed to define a dense map for a non-rigid object due to its complex changes such as deformation, distortion, and extension. Although this problem has made some breakthroughs [1,2,3,4], finding dense shape correspondence is still very challenging.
Traditional non-rigid shape correspondence methods are generally established by measuring the similarity of point descriptors between two shapes, such as spectral geometry [5] or diffusion distance [6]. However, these methods rely heavily on initialization, which falls easily into local optimization with slow matching speed and low matching accuracy. Faced with large amounts of data and dense correspondence, scholars have further proposed to obtain the features of each point on the manifold surface through neural networks [7,8,9]. Unfortunately, these techniques directly regress the dense relationship of all input points, severely over-parameterizing the deformation and leading to poor generalization. Recent efforts mainly focus on functional map frameworks [10,11,12,13] that converts point-to-point correspondence between models into a linear transformation in function space, mapping a point on the source model to multiple points on the target model to solve ambiguity problems. Despite significant progress in this area, these methods ignored the hierarchical geometric relations of the associated semantic parts. When the differences between the two input models are large, the extracted non-rigid shape descriptors lack sensitivity to specific information such as orientation and position, leading to unreliable correspondence results.
Based on the above limitation, we propose a novel capsule graph neural network based on component constraint for the 3D non-rigid shape correspondence, called CA-CGNet, to further capture descriptors with powerful representation abilities. The entire framework of our method is given in Figure 1. The over-segmentation method is used to segment the mesh into several patches, which are clustered to form components using a dynamic clustering algorithm. In order to further improve the quality of feature acquisition, we design a novel graph routing algorithm to learn the relationship between the capsules in the same layer through a component-aware graph neural network (CA-GNN). Then, the knowledge distillation strategy is used to improve the convergence speed of the network by reducing the parameters while maintaining accuracy. Finally, a component pair constraint is added to the functional map with a component-based semantic loss function, which makes the shape correspondence more accurate and improves the robustness of the network.
Our main contributions can be summarized as follows:
  • We design a novel network structure called CA-CGNet, which enhances the expressive ability of the network to represent object spatial pose and orientation changes by adding local mesh pose details to achieve high-precision shape correspondence.
  • To reduce the interference of noisy patch assignments, we propose a dynamic clustering algorithm to cluster the over-segmented meshes dynamically according to feature similarity to form components. By adding a component constraint to the functional maps and integrating component-based semantic constraint loss into the regularization term, the accuracy of shape correspondence is further improved.
  • The component-aware graph routing treats capsules as nodes in a graph neural network by adding component constraints to obtain more accurate relationships between capsules. In addition, the knowledge distillation strategy is used to reduce the number of parameters while maintaining network performance.
  • Experiments on four challenging datasets show qualitatively and quantitatively that the CA-CGNet has stronger robustness and better generalization. The ablation study demonstrates that component pair constraint, component-aware graph routing, and knowledge distillation strategy have a great improvement in network performance.

2. Related Work

2.1. Non-Rigid Shape Correspondence

Various work has been done in the direction of shape correspondence. Before deep learning methods were commonly applied to point descriptor learning, many scholars conducted in-depth research on handcrafted feature descriptors for characterizing the geometric information of shapes. Tombari et al. [14] proposed the signature of histograms of orientations (SHOT) descriptor with rotation and translation invariance by counting the topological features adjacent to points in the local coordinate system and saving them in the histogram. Average geodesic distance (AGD) [15] was the attitude insensitive descriptor, which could effectively measure the overall topological characteristics of multi-resolution models. Sun et al. [16] extracted all key points in different proportions and combined them with the thermal diffusion equation to develop a heat kernel signature (HKS) with multi-scale characteristics. In [17], the wave kernel signature (WKS) used the Schrodinger equation to acquire the evolution process of particles on the shape surface to obtain the features of the vertex on the object surface. To explore posture more effectively, Zuffi et al. [18] proposed to use graphical nodes corresponding to body parts so that these parts could be independently translated and rotated in 3D to represent different body shapes and to capture pose-dependent shape variations. This method was similar to our approach but only applied to human models.
Recently, researchers have tried to use deep learning methods to further process shape descriptors to obtain high-quality geometric features of 3D non-rigid surfaces [19,20]. Cyclic-FM [21] was based on a cyclic mapping between metric spaces for self-supervised dense correspondence mapping between non-isometric shapes. Marin et al. [22] extracted stable landmarks over human bodies, relying entirely on the geometric properties in the spectral domain. The MGCN [23] advanced a multi-scale graph convolutional network, which transformed wavelet energy decomposition signature (WEDS) to a more discriminative descriptor. Amor et al. [24] made use of deep residual neural networks to calculate minimizing paths between deformations and, thus, between shapes for the alignment of 3D shapes under complex topology-preserving transformations. Although these methods have achieved good results, they did not make full use of local pose information to maintain the geometrical relationship of the manifold. In addition, when the model resolution decreases, the discriminative power of the network decreases significantly.

2.2. Capsule Graph Neural Network

Capsule network was an improved strategy for CNN proposed by Hinton et al. [25], which used vectors to represent the mutual relations between features. We find the outcomes and limitations of several approaches based on a capsule graph neural network in Table 1. The basis of the capsule network was the routing process, which passed the input vector to the upper capsule through the protocol. To effectively preserve the variability and correlation between features during the routing process, capsule graph neural networks [26] extracted node features in the form of capsules to capture high-quality node embeddings at the graph level. In [27], Caps-GNN was adopted to learn graph properties for encoding underlying characteristics. CapsGNNEM [28] utilized EM routing to obtain graph attributes from the node features extracted by the graph neural network. Some other research [29,30] represented nodes as a group of node-level capsules, jointly learning node embedding and extracting salient features of corresponding nodes through heterogeneous factors of the capsules. Later in [31], the method was proposed to model the relationship between concept capsules of the same layer through a graph network with an external storage matrix. In our work, we propose a novel routing algorithm that leverages component constraints to transform primary capsules into high-quality graph embedding, which better maintains the local posture of the model.

2.3. Correspondence with Functional Maps

Different from point-to-point correspondence, the concept of the functional map was introduced in [32], which represented the mapping between shapes as a small matrix and transmitted information by encoding the relationship between the basic functions of shapes. Inspired by that, some scholars have proposed several learning methods for optimizing functional maps [10,11,33]. With the development of deep learning, a neural network was used in the functional map framework to generate more accurate correspondence. Litany et al. [12] proposed a deep neural network architecture named FMNet by providing a dense corresponding representation of linear operators through paradigm transformation and a structured prediction model. SURFMNet introduced in [13] was to use pure geometric scalars to directly establish correspondence in the 3D shape set without any prior information. In [34], Deep Geometric Functional Maps was a feature extraction network based on a functional map representation that learned directly from the shape geometry with a new regularized map extraction layer and loss to enforce the structural properties. Donati et al. [35] generalized the functional map framework to the conformal maps between tangent vector fields, which reflected the complex structure of the surface to maintain orientation-aware results but naturally restricted to differentials of conformal mappings. The key difference between the above approach and our proposed approach is that we embed component constraints in the functional maps framework, focusing on the local area feature information of the mesh.

3. Proposed Method

In this section, we present a component-aware capsule graph neural network for shape correspondence. In Section 3.1, we introduce the whole network architecture of CA-CGNet. Section 3.2 illustrates the process of turning input shapes into component pairs. The component-aware graph routing is given in Section 3.3. We introduce the functional map of component pair constraints in Section 3.4 and the semantic regularization term in Section 3.5.

3.1. CA-CGNet

In this work, we design a deep neural network named the “Component-aware Capsule Graph Neural Network” (CA-CGNet) for shape correspondence based on the functional maps framework. We believe that the capsule graph network allows a powerful understanding of the mesh’s positions and directions features. Since existing routing algorithms do not fully consider the potential features between capsules, a component-aware graph capsule network is proposed to further extract deeper semantic information by building graph relationships between capsules. Then, in order to get more precise shape correspondence, the functional map with component constraint and the loss function with semantic constraint as the regularization terms are designed. The basic pipeline can be described by the following steps: firstly, the mesh is over-segmented to form several patches, and the patches with similar features are aggregated together to form components through a dynamic clustering algorithm. Secondly, in order to fully learn the relationships between capsules, we consider capsules as nodes of the graph neural network and further enhance the capsules through a component-aware graph routing algorithm with a knowledge distillation strategy to increase the convergence speed by decreasing the number of parameters while maintaining network accuracy. Finally, the component-constrained functional map is represented with Equation (11), and the loss function with semantic constraint is derived from Equation (17). The entire network architecture is given in Figure 2 in detail.

3.2. Component Extraction

A 3D model usually contains a large number of faces, which can generate a huge amount of computation if the entire mesh model is segmented directly. Therefore, the mesh sub-module is constructed using the method proposed by Golovinskiy et al. [36] to over-segment the 3D mesh into several small patches. Since the number of vertices contained in each patch is inconsistent, Principal Components Analysis (PCA) is used to reduce the dimensionality to make the number of vertices contained in each patch equal. After that, the Average Geodesic Distance (AGD) feature of each patch is calculated, and the sub-model is aggregated into components with dynamic clustering. Finally, the improved Hausdorff distance is used for component matching to extract the component pairs.

3.2.1. Dynamic Clustering

Mesh segmentation is equivalent to the clustering task of a 3D surface. To further reduce noise interference, we use dynamic clustering strategy to perform more accurate part-to-whole classification of the acquired features. In the DC part of Figure 2, small circles with different colors represent different types of features, and the circles with red borders are used as the initial feature of clustering. The features are clustered according to the correlation, and the highly correlated features are clustered together. Generally speaking, dynamic clustering includes three steps, namely, calculating the feature correlation matrix, selecting the set of features with the largest difference as the initial feature, and clustering the remaining features.
In a 3D mesh, the properties of vertices are represented by their feature matrices. By calculating the Manhattan distance between the feature matrices of different vertices, we can obtain the degree of correlation between these features. Specifically, the degree of correlation R between feature X and Y is calculated with
R ( X , Y ) = σ i = 1 n x i y i
where X = x 1 , x 2 , , x n and Y = y 1 , y 2 , , y n are both n-dimensional vectors, σ represents the sigmoid activation function.
In the correlation matrix R R N × N , the degree of correlation between features decreases with the increase of the coefficient. By choosing the coefficient with the largest value in R , we can obtain the row coordinate r 1 and column coordinate r 2 of this value, which represent the sequence numbers of the two groups of features with the farthest correlation. Then, these two features are added to set G and respectively used as the initial features for each cluster.
The value of the i-th row in matrix R represents the degree of correlation between the i-th type feature and other types of features. The cluster of the remaining features is expressed as follows:
r i = arg min r j G R r j , ε , ε r j
where r j represents the sequence number of the initial two types of features, and ε represents the sequence number of the remaining features. Thus, these features can dynamically be divided into two clusters. Repeating the above steps on each cluster, we will repartition each cluster into two new sets of feature clusters. After executing this process three times, the over-segmented patches are clustered into several types.

3.2.2. Component Matching

The Hausdorff distance [37] reflects the gap of two sets by measuring the distance between two nonempty subsets. Using Hausdorff distance for model matching can simplify the calculation, but it is frequently interfered with by noise. In order to build the corresponding relationship between the components of the two models, this paper adopts the improved Hausdorff distance to match the components.
For two sets A = a 1 , a 2 , , a i and B = b 1 , b 2 , , b j , the bidirectional Hausdorff distance between these two sets is defined as follows:
H ( A , B ) = max { h ( A , B ) , h ( B , A ) }
where h ( A , B ) denotes the unidirectional Hausdorff distance from set A to set B, h ( B , A ) denotes the unidirectional Hausdorff distance from set B to set A. The improved unidirectional Hausdorff distance is written as:
h ( A , B ) = 1 N a A min b B a b
where N represents the number of elements in set A, · represents the distance between two elements in different sets.
We first select component A in the manifold M and calculate its Hausdorff distance to each component in the manifold N to obtain the distance set D = d 1 , d 2 , , d k . Then we choose the minimum distance d min in the set to form a component pair and delete the two components from the original set. We repeat the steps for the remaining components in turn to construct a pair of components.

3.3. Component-Aware Graph Routing

Considering the CapsNet can capture the part-whole relationship to cope with ambiguity [25], we introduce the capsule network to enhance feature expression ability. The features extracted by GNN are replicated in eight copies, the primary capsules are obtained after Conv1d+BN+ReLU, and the pose features of the mesh are further extracted to compose the latent capsules. Among them, routing is the basis of the capsule network, which transmits the information from the previous capsule layer to the next one through the protocol. To further improve the quality of feature acquisition, we propose a component-aware graph routing, which learns the relationship between capsules in the same layer through CA-GNN.

3.3.1. Multi-Layer Attention Graph Routing

We extract features from different CA-GNN layers and fuse the capsule features through a multi-layer attention mechanism. The capsule is regarded as each node in the graph, and the adjacency matrix A is used to represent the relationship between nodes in CA-GNN, which can be expressed as:
A ˜ = D 1 2 A D 1 2 + E
where A ˜ is the normalized N × N adjacency matrix, N is the number of capsules, D is the degree matrix, and E is the identity matrix.
The capsule vectors can be regarded as the probability of certain attributes, which is used to measure the semantic similarity of capsules. Since the capsule vector is an implicit expression, it is difficult to directly calculate the relationship between capsules. Thus, we used Wasserstein Distance to measure the distance between the distribution of capsule properties. The Component-aware Wasserstein Distance is defined as:
W ( V i , V j ) = inf Ω R m × R m v i v j 2 + λ X v i X v j 2 Ω d v i , d v j 1 2
where X v i and X v j are the component category to which the point belongs, Ω is an element of set Π V i , V j , and λ is the coefficient and is set to be 0.005. Let a i j = W ( V i , V j ) , we can convert component-aware Wasserstein Distance to adjacency between capsules inside. In this case, a i j may be negative, so the adjacency matrix A needs to be normalized, which is denoted as
A ˜ = e A i j e A i + E
Considering that the contribution of different levels of graph embedding is inconsistent, we introduce an attention mechanism to combine these different levels of graph embedding to get the final output result as shown in Figure 3. The procedure of the multi-layer attention module can be written as:
w j = σ H 1 G A P , H 1 G M P H 2 G A P , H 2 G M P H 3 G A P , H 3 G M P
where σ ( · ) represents the softmax function, and ⊕ represents the feature concat.
Each capsule is regarded as a center node, and the relationship between capsules is measured by distance with the performed normalization. Then, the multi-layer attention mechanism is introduced to transform the low-level capsule into the high-level capsule using component graph routing. Algorithm 1 gives the complete algorithm of Component-aware Graph Routing.
Algorithm 1: The Algorithm of the Component-aware Graph Routing
Applsci 13 03261 i001

3.3.2. Knowledge Distillation Strategy

We introduce a knowledge distillation strategy into component-aware graph routing, which can transfer knowledge from a complex deep teacher network to a simple student model. Since the node features and the associations between nodes extracted by graph neural networks play an important role in embedding features, we capture the geometric information of the latent space by extracting the features of the CA-GNN layer and using the graph information extracted from the teacher network to train the student network. Therefore, the student can be further guided by matching the node structure extracted by itself and the node structure embedded by the teacher. As shown in Figure 4, the original component-aware graph routing is defined as the teacher network, and the pruned component-aware graph routing is defined as the student network. The teacher network deploys six CA-GNN modules after the primary capsule to extract features with more semantic information. In contrast, the student network only contains three CA-GNN modules, which greatly reduces the parameters to improve the convergence speed of the network.
In order to measure local detail dependence in low dimensional space, we conduct inter-layer knowledge distillation between CA-GNN layers. The feature vectors of teacher network I T are obtained through two CA-GNN layers, and the feature vectors of student network I S are obtained through one CA-GNN layer. Then, the loss of inter-layer knowledge distillation is defined as:
L i n t e r = log ( I S ) log log ( I S ) I T
After the Multi-layer attention module integrates the features of each CA-GNN layer, the latent capsule containing global information is obtained. Let O T and O S be the feature vectors of the teacher network and student network, the output layer knowledge distillation loss over all the capsules is as follows:
L o u t = 1 N k = 1 N j : ( i , j ) ε O S i j log O S i j O T i j
where ε represents edges in a graph network, k donates the number of capsules.

3.4. Functional Maps with Component Constraint

We propose a component constrained functional maps, which provides a local–global commutative preservation for shape correspondence between manifold P and Q. Despite traditional functional maps providing the simplicity and efficiency of shape correspondence, its estimation pipeline solely enforces the commutativity of global operators. This method is to search matching points across the whole manifold, which leads to a large extent to some inaccurate matching. To alleviate this issue, we propose an available segmentation of shapes for relation constraints. Then, the optimal functional map C in the least square sense can be expressed as:
C o p t = arg min C C A P B Q 2 + γ 1 Λ P C C Λ Q 2 + γ 2 C X P X Q C 2
where A P and B Q are the basis matrix consisting of the basis function coefficients of the source model P and the target model Q, Λ P and Λ Q are the diagonal matrix of Laplacian-iBeltrami eigenvalues, X P and X Q are the Hausdorff distance matrix, and γ 1 and γ 2 are the weights.
The first term of Equation (11) calculates the deviation of the point-wise descriptor coefficient matrix based on SHOT descriptors of manifold P and Q. For a connected smooth compact manifold P with K vertices, an index function f : P R can be constructed by calculating the average value of the point and its neighborhood coordinates, which can be expressed as:
f ( x ) = i m i ϕ i
where m i = f , ϕ i P , ϕ i i = 0 , 1 , , k 1 is the set of the eigenfunctions for the orthogonal basis of f, and f is the corresponding discretization of the smooth function f. Let F : f ( P ) f ( Q ) be a linear operator to calculate correspondence between manifolds P and Q. By computing a set of descriptor functions ϕ i i 0 and ϕ j j 0 on shapes P and Q, the function F can be represented as:
F ( f ) = j i m i c j i ϕ j
where c j i = F ϕ i , ϕ j N , and then admit a matrix representation C = c j i . To determine the matrix C , we truncate the eigenfunctions after the first d coefficients to significantly approximate matrix C . Assuming to be given a set of coefficients z i = f , ϕ i and g j = s , ϕ j , the matrices A P and B Q can be stored by these Fourier coefficients.
The second term of Equation (11) is a regularization item that constrains the descriptor by reinforcing its overall structure properties. With the Laplacian matrix L , the set of eigenvalues λ i are then defined as:
λ i = L M 1
where M is the K × K diagonal mass matrix. Thus, the first k eigenvalues of Laplacian–Beltrami operator on model P and Q compose the diagonal matrices Λ P = d i a g ( λ i ) P and Λ Q = d i a g ( λ j ) Q .
The third term of Equation (11) penalizes the local commutativity through the component constraints. Each manifold is segmented into several types of components, and Hausdorff distance is used to construct component pairs. Hausdorff distance is a measure of the similarity between two sets of points, so we can calculate the difference between components. The distance matrix X P and X Q can be expressed as:
X P = h C P , C Q X Q = h C Q , C P
where C P and C Q are the features of component pairs. Thus, each set of component pairs can be used to impose local constraints on functional maps.

3.5. Semantic Regularization

In order to enhance the robustness of function mapping, we transfer the component network discussed in Section 3.2 to the feature space and further develop the semantic regularization term. By penalizing the difference between the weights of each node on each component pair and the weights of corresponding nodes to regularize the classification model, we define a semantically constrained loss function to preserve the classification in the embedded feature space, which is formulated as:
L c = 1 n j n i x P i x Q i 2
We integrate this semantic constrained loss function into the final loss function, then the overall spectral loss function L of the corresponding shape is defined as follows:
L = μ 1 L c + μ 2 L g
where μ 1 and μ 2 are the coefficient, and L g denotes the difference between the mapping matrix C and the corresponding ground-truth matrix [38]. In our experiments, μ 1 and μ 2 are both set to be 1. By adding the semantic regularization term, the spectral loss function of shape corresponding learning network is minimized; thus, the constraints on the corresponding functions are increased during the training process.

4. Experiments and Evaluation

In this section, we verify the performance of 3D non-rigid shape correspondence by comparing CA-CGNet with some classical methods. This experiment runs on GeForce GTX 1080Ti CPU with 32GB memory. ADAM optimization algorithm with parameter set as β 1 = 0.9 , β 2 = 0.999 , ε = 10 8 is used, and the learning rate is changed by polynomial decay strategy, where the initial learning rate is 10 3 , the end learning rate is 10 5 , and the batch size is 1.

4.1. Dataset

Four types of non-rigid 3D datasets, namely FAUST, SCAPE, TOSCA, and KIDS, are used for experiments. In the experiment, the first 80 meshes in the original FAUST dataset are trained, and the last 20 meshes are used to test the model. And its vertices are sampled to 4096 vertices to verify the robustness of the experiment. For other datasets, the vertices are sampled to 4096 to facilitate comparison experiments.
FAUST dataset has models with the same triangulation and resolution, and each model contains 6890 vertices. There are 100 models in the data set, and every 10 models is a group representing different poses of the same research object. The shapes in the dataset have strong non-isometric deformation, and the ground truth correspondence of vertices between all shapes is known.
SCAPE dataset is a data-driven approach to building a human shape model across the shape and pose variations, using a static scan and a labeled motion capture sequence of humans to artificially generate high-quality animated surface models of moving humans with realistic muscle deformations.
TOSCA dataset consists of many different types of 3D graphics, with significant differences between each set of models. Shapes in the same group are compatible, meaning that all shapes have the same mesh resolution and connectivity.
KIDS dataset is a synthetic non-isometric dataset containing two sets of human models with different poses, which have the same number of vertices and contain ground truth.

4.2. Component Matching

For each 3D mesh, over-segmentation is used to divide it into many different regions, and then dynamic clustering is used to merge regions to form components. The selection of the number of clusters k not only affects the amount of computation but also affects the subsequent matching effect. Therefore, we use different cluster numbers with k = 4 , 5 , 6 , 7 , 8 , 9 , 10 for experimental tests. Table 2 shows the identification errors of experiments on human and non-human models with a different number of clusters. Through the experiment, it can be found that the number of clusters with k = 8 can achieve the minimum identification error on the human model and centaur model, and the clustering amount of k = 5 can achieve the minimum identification error on other non-human models. Therefore, in the subsequent experiments, we adopted the dynamic classification strategy with k = 8 on the human model and centaur model and k = 5 on other non-human models.
Figure 5 shows the component formation results of the human model on four different datasets, where different colored body parts represent different types of components. In order to further verify the robustness of the proposed method, we conduct dynamic clustering on non-human models, as shown in Figure 6. It shows that our method achieves equally effective clustering results on these non-human models.
After the components are formed, the component of different models is matched to obtain a pair of components. Figure 7 visualizes the matching results of different models, where the left side shows the matching results of components with a different pose of the same subject, and the right side shows the matching results of components with different subjects. Experiments demonstrate that our method achieves good results in different model component matching, which proves the robustness of our method.

4.3. Correspondence Results

In this section, we evaluate the experimental results of the proposed algorithm and other classical algorithms on four non-rigid datasets. We test our model on the original FAUST dataset, and Figure 8 shows the correspondence results, where intra-pair represents different poses of the same subject and inter-pair represents different subjects.
The average error (AE) of our method compared with other state-of-the-art methods on the original FAUST dataset is shown in Table 3, including the average error of intra-pair and the average error of inter-pair. On the intra-pairs, our method achieves 1.85 cm, which outperforms ResNet-LDDMM significantly by 4.15%. On the inter-pairs, our method achieves 2.37 cm, which even outperforms the state-of-the-art method by 9.20%. Our method has the smallest average error in inter-pair and also performs well in intra-pair. Furthermore, SP [18] uses distributed inference to predict the model in the model initialization phase and combines local-based human pose inference with graph form human body shape models, the average error is smaller in the intra-pair. However, our method is more concerned with the generalization of shape matching, the combined average error is smaller than SP.
In order to verify the robustness of the proposed method, we resample the original FAUST dataset to 4096 vertices for testing. Figure 9 visualizes the matching results of some methods on the remeshed FAUST dataset. It shows that our method achieves the best matching results in deformable shape correspondence.
Furthermore, the correspondence performance of several methods is compared on the FAUST dataset and the remeshed FAUST dataset by using the Princeton benchmark protocol [39], as shown in Figure 10. With the increase of the geodesic error threshold, the matching accuracy is improved gradually. On the original FAUST dataset, our method is slightly improved compared with FMNet, SURFMNet, and MGCN, while the corresponding accuracy of our method is significantly improved compared with other methods on the remeshed FAUST dataset. Our approach achieves good matching for meshes of different resolutions, which indicates that it has strong robustness.
We further test on SCAPE, TOSCA, and KIDS datasets to verify the generalization of our approach. For a more intuitive comparison of experimental results, Figure 11 illustrates the geodesic errors of these methods on remeshed SCAPE, TOSCA, and KIDS datasets. In order to clearly demonstrate the corresponding effect, we visualize the correspondence results of different methods on these three datasets, as shown in Figure 12. Our method shows more accurate matching results, which indicates that our method has better generalization.
Considering the large gap between the animal model and the human model, the proposed method is tested on the animal model in the TOSCA dataset to verify the adaptability of the method. Figure 13 shows the corresponding results of reference shapes and deformable shapes on the non-human models. By contrast with FMNet, SURFMNet, and MGCN, our method is closer to ground truth in corresponding details.
In Table 4, we show the average geodesic error of several methods on four resampled datasets. The bold represents the minimum error in each column. In the remeshed FAUST dataset, the error of our method is 78.72% less than FMNet, 56.60% less than SURFMNet, and 47.43% less than MGCN. In remeshed SCAPE, it is 48.88% less than FMNet, 45.42% less than SURFMNet, and 27.71% less than MGCN. In the remeshed KIDS dataset, it is 62.32% less than FMNet, 65.66% less than SURFMNet, and 57.74% less than MGCN. In remeshed TOSCA, it is 74.65% less than FMNet, 76.16% less than SURFMNet, and 71.87% less than MGCN. It can be found that our model achieves better results than those of the compared methods, which demonstrates that our model is more robust on non-rigid 3D models with lower resolution.

4.4. Ablation Study

4.4.1. Effectiveness of the Component Pair Constraint

In Section 3.2, we introduce the component pair generation strategy. The purpose of this section is to study whether component pair constraints affect network correspondence results. Figure 14 visualizes the geodesic error on the original FAUST dataset. The correspondence error on the basis of the geodesic distance to the ground truth is demonstrated with a color scale of 0 to 0.1. As shown in this figure, our model achieves a lower geodesic error on the original FAUST dataset by using a component pair constraint, which proves the effectiveness of the component pair constraint.

4.4.2. Effectiveness of the Component-Aware Graph Routing

In Section 3.3, component-aware graph routing is proposed to compute the primary capsule to transfer the input vector to the latent capsule. To test the effectiveness of the routing algorithm, we compare the proposed algorithm with dynamic routing. Table 5 shows the average geodesic error of different routing algorithms on the TOSCA dataset. The bold represents the minimum error in each column. As shown in Table 5, the average geodesic error of the dynamic routing algorithm is 0.1276, and the average geodesic error of our method without knowledge distillation is 0.1213. Our method improves the correspondence performance with an average error of 0.1187, which is 6.97% less than the dynamic routing and 2.14% less than our method without knowledge distillation. Our model achieves smaller average geodesic errors on different types of meshes, providing slightly better results on the TOSCA dataset. It illustrates that the routing strategy in this paper can capture more adequate pose information compared to dynamic routing, resulting in significant improvement in shape correspondence.
In order to fully capture the features of the potential space, inter-layer distillation and output-layer distillation are used to transfer the model’s knowledge. To verify the effectiveness of the knowledge distillation strategy, we train the CA-GCNet and the CA-GCNet without a knowledge distillation strategy separately to acquire the network accuracy, as shown in Figure 15. The accuracy curves illustrate that the prediction accuracy of our proposed CA-CGNet is higher than that of the CA-CGNet without the knowledge distillation strategy, which indicates that the student network can learn more detailed features from the teacher network.

5. Discussion

As shown in Figure 14, the functional maps with component pair constraint can capture the local pose details of 3D non-rigid models more effectively. Hence, functional maps with component pair constraints are able to use deep features with local and global structural information for shape correspondence. To reduce the effect of noise interference, the dynamic clustering strategy is used to dynamically merge the patches formed by over-segmentation, capturing the potential mutual influence of information between the local patch and the global model. Then, component pairs are obtained by matching components on different models using the Hausdorff distance, which leads to high-level feature representation and corresponding accuracy by using component pair-constrained functional maps.
Inter-layer distillation and output-layer distillation are designed to constrain the student network to reduce the network parameters and speed up the gradient descent during the training process. The average geodesic error shown in Table 5 illustrates the efficiency of our component-aware graph routing for improving the corresponding accuracy of 3D models with low resolution.

6. Conclusions

Based on the functional maps framework, this paper proposes a component-aware shape correspondence network that can effectively deal with non-rigid deformation. Taking capsules as nodes in the graph, the semantic information between capsules is further extracted by the component-aware graph routing, and the model is simplified by a knowledge distillation strategy. By adding the component constraint to the functional map and taking the component-based semantic loss as a regularization term, our method can learn more representative features on 3D meshes. Experiments on four challenging datasets show qualitatively and quantitatively that our model has stronger robustness and better generalization ability. The proposed framework also can provide more effective solutions in biological computing, human pose estimation, and medical image processing. One limitation of dynamic clustering is that it is difficult to cluster patches if the feature difference is hardly remarkable, which we hope to address in a future study. On the other hand, the determination of the number of clusters requires a large number of iterative experiments and is computationally intensive. In future work, we will adopt an optimization algorithm to determine the number of clusters to reduce computational expenditure.

Author Contributions

Conceptualization, Y.L. and M.C.; methodology, Y.L.; software, M.C.; validation, Y.L.; formal analysis, M.C.; investigation, Y.L.; resources, M.C.; data curation, Y.L.; writing—original draft preparation, M.C.; writing—review and editing, Y.L.; visualization, M.C.; supervision, Y.L.; project administration, Y.L.; funding acquisition, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NSFC 61972353, NSF IIS-1816511, OAC-1910469 and Strategic Cooperation Technology Projects of CNPC and CUPB: ZLZX2020-05.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jain, V.; Zhang, H.; Van Kaick, O. Non-rigid spectral correspondence of triangle meshes. Int. J. Shape Model. 2007, 13, 101–124. [Google Scholar] [CrossRef]
  2. Jian, B.; Vemuri, B.C. Robust point set registration using gaussian mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1633–1645. [Google Scholar] [CrossRef] [PubMed]
  3. Ma, J.; Zhao, J.; Tian, J.; Yuille, A.L.; Tu, Z. Robust point matching via vector field consensus. IEEE Trans. Image Process. 2014, 23, 1706–1721. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ma, J.; Zhao, J.; Yuille, A.L. Non-rigid point set registration by preserving global and local structures. IEEE Trans. Image Process. 2015, 25, 53–64. [Google Scholar] [PubMed] [Green Version]
  5. Bronstein, A.M.; Bronstein, M.M.; Kimmel, R. Generalized multidimensional scaling: A framework for isometry-invariant partial surface matching. Proc. Natl. Acad. Sci. USA 2006, 103, 1168–1172. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Coifman, R.R.; Lafon, S.; Lee, A.B.; Maggioni, M.; Nadler, B.; Warner, F.; Zucker, S.W. Geometric diffusions as a tool for harmonic analysis and structure definition of data: Diffusion maps. Proc. Natl. Acad. Sci. USA 2005, 102, 7426–7431. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Litman, R.; Bronstein, A.M. Learning spectral descriptors for deformable shape correspondence. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 171–180. [Google Scholar] [CrossRef] [PubMed]
  8. Groueix, T.; Fisher, M.; Kim, V.G.; Russell, B.C.; Aubry, M. 3d-coded: 3d correspondences by deep deformation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 230–246. [Google Scholar]
  9. Wang, W.; Ceylan, D.; Mech, R.; Neumann, U. 3dn: 3d deformation network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1038–1046. [Google Scholar]
  10. Kovnatsky, A.; Bronstein, M.M.; Bresson, X.; Vandergheynst, P. Functional correspondence by matrix completion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 905–914. [Google Scholar]
  11. Nogneng, D.; Melzi, S.; Rodola, E.; Castellani, U.; Bronstein, M.; Ovsjanikov, M. Improved functional mappings via product preservation. In Computer Graphics Forum; Blackwell: Oxford, UK, 2018. [Google Scholar]
  12. Litany, O.; Remez, T.; Rodola, E.; Bronstein, A.; Bronstein, M. Deep functional maps: Structured prediction for dense shape correspondence. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5659–5667. [Google Scholar]
  13. Roufosse, J.M.; Sharma, A.; Ovsjanikov, M. Unsupervised deep learning for structured shape matching. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1617–1627. [Google Scholar]
  14. Tombari, F.; Salti, S.; Stefano, L.D. Unique signatures of histograms for local surface description. In Proceedings of the European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; Springer: Cham, Switzerland, 2010; pp. 356–369. [Google Scholar]
  15. Hilaga, M.; Shinagawa, Y.; Kohmura, T.; Kunii, T.L. Topology matching for fully automatic similarity estimation of 3D shapes. In Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 12–17 August 2001; pp. 203–212. [Google Scholar]
  16. Sun, J.; Ovsjanikov, M.; Guibas, L. A concise and provably informative multi-scale signature based on heat diffusion. In Computer Graphics Forum; Blackwell: Oxford, UK, 2009. [Google Scholar]
  17. Aubry, M.; Schlickewei, U.; Cremers, D. The wave kernel signature: A quantum mechanical approach to shape analysis. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; IEEE: New York, NY, USA, 2011; pp. 1626–1633. [Google Scholar]
  18. Zuffi, S.; Black, M.J. The stitched puppet: A graphical model of 3d human shape and pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3537–3546. [Google Scholar]
  19. Guo, J.; Wang, H.; Cheng, Z.; Zhang, X.; Yan, D.M. Learning local shape descriptors for computing non-rigid dense correspondence. Comput. Vis. Media 2020, 6, 95–112. [Google Scholar] [CrossRef] [Green Version]
  20. Halimi, O.; Litany, O.; Rodola, E.; Bronstein, A.M.; Kimmel, R. Unsupervised learning of dense shape correspondence. In Proceedings of the PIEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4370–4379. [Google Scholar]
  21. Ginzburg, D.; Raviv, D. Cyclic functional mapping: Self-supervised correspondence between non-isometric deformable shapes. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 36–52. [Google Scholar]
  22. Marin, R.; Melzi, S.; Rodola, E.; Castellani, U. Farm: Functional automatic registration method for 3d human bodies. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2020; Volume 39, pp. 160–173. [Google Scholar]
  23. Wang, Y.; Ren, J.; Yan, D.M.; Guo, J.; Zhang, X.; Wonka, P. Mgcn: Descriptor learning using multiscale gcns. ACM Trans. Graph. (TOG) 2020, 39, 1–122. [Google Scholar] [CrossRef]
  24. Amor, B.B.; Arguillère, S.; Shao, L. ResNet-LDDMM: Advancing the LDDMM Framework Using Deep Residual Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 3707–3720. [Google Scholar] [CrossRef]
  25. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic routing between capsules. Adv. Neural Inf. Process. Syst. 2017, 30, 3859–3869. [Google Scholar]
  26. Xinyi, Z.; Chen, L. Capsule graph neural network. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  27. Li, J.; Li, S.; Zhao, W.X.; He, G.; Wei, Z.; Yuan, N.J.; Wen, J.R. Knowledge-enhanced personalized review generation with capsule graph neural network. In Proceedings of the 29th ACM International Conference on Information & Knowledge Management, Galway, Ireland, 19–23 October 2020; pp. 735–744. [Google Scholar]
  28. Lei, Y.; Zhang, J. Capsule Graph Neural Networks with EM Routing. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management (CIKM ’21), Gold Coast, QL, Australia, 1–5 November 2021; Association for Computing Machinery: New York, NY, USA, 2021; pp. 3191–3195. [Google Scholar] [CrossRef]
  29. Yang, R.; Dai, W.; Li, C.; Zou, J.; Xiong, H. NCGNN: Node-Level Capsule Graph Neural Network for Semisupervised Classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 3179306. [Google Scholar] [CrossRef] [PubMed]
  30. Yang, J.; Zhao, P.; Rong, Y.; Yan, C.; Li, C.; Ma, H.; Huang, J. Hierarchical graph capsule network. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 10603–10611. [Google Scholar]
  31. He, Z.; Li, W.; Yan, Y. Modeling knowledge proficiency using multi-hierarchical capsule graph neural network. Appl. Intell. 2022, 52, 7230–7247. [Google Scholar] [CrossRef]
  32. Ovsjanikov, M.; Ben-Chen, M.; Solomon, J.; Butscher, A.; Guibas, L. Functional maps: A flexible representation of maps between shapes. ACM Trans. Graph. (ToG) 2012, 31, 1–11. [Google Scholar] [CrossRef]
  33. Corman, E.; Ovsjanikov, M.; Chambolle, A. Supervised descriptor learning for non-rigid shape matching. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 283–298. [Google Scholar]
  34. Donati, N.; Sharma, A.; Ovsjanikov, M. Deep geometric functional maps: Robust feature learning for shape correspondence. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8592–8601. [Google Scholar]
  35. Donati, N.; Corman, E.; Melzi, S.; Ovsjanikov, M. Complex functional maps: A conformal link between tangent bundles. In Proceedings of the Computer Graphics Forum, Eugene, OR, USA, 30 August–1 September 2022; Wiley Online Library: Hoboken, NJ, USA, 2022; Volume 41, pp. 317–334. [Google Scholar]
  36. Golovinskiy, A.; Funkhouser, T. Consistent segmentation of 3D models. Comput. Graph. 2009, 33, 262–269. [Google Scholar] [CrossRef] [Green Version]
  37. Huttenlocher, D.P.; Klanderman, G.A.; Rucklidge, W.J. Comparing images using the Hausdorff distance. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 850–863. [Google Scholar] [CrossRef] [Green Version]
  38. Lian, Y.; Gu, D.; Hua, J. SORCNet: Robust non-rigid shape correspondence with enhanced descriptors by Shared Optimized Res-CapsuleNet. Vis. Comput. 2022, 39, 749–763. [Google Scholar] [CrossRef]
  39. Kim, V.G.; Lipman, Y.; Funkhouser, T. Blended intrinsic maps. ACM Trans. Graph. (TOG) 2011, 30, 1–12. [Google Scholar]
Figure 1. The pipeline of component-aware capsule graph neural network framework.
Figure 1. The pipeline of component-aware capsule graph neural network framework.
Applsci 13 03261 g001
Figure 2. The architecture of our proposed network CA-CGNet. For the mesh component, dynamic clustering (DC) is applied to form the patches generated by over-segmentation. To further improve the quality of feature acquisition, the component-aware graph routing algorithm is constructed to learn the relationship between the capsules in the same layer through a component-aware graph neural network (CA-GNN). The component constrained functional map is represented by Equation (11) and the component-based semantic loss is derived from Equation (17).
Figure 2. The architecture of our proposed network CA-CGNet. For the mesh component, dynamic clustering (DC) is applied to form the patches generated by over-segmentation. To further improve the quality of feature acquisition, the component-aware graph routing algorithm is constructed to learn the relationship between the capsules in the same layer through a component-aware graph neural network (CA-GNN). The component constrained functional map is represented by Equation (11) and the component-based semantic loss is derived from Equation (17).
Applsci 13 03261 g002
Figure 3. The Multi-layer Attention Module. GAP means global average pooling, while GMP refers to global maximum pooling. FC is the fully connected layer. w j means the weight for each layer.
Figure 3. The Multi-layer Attention Module. GAP means global average pooling, while GMP refers to global maximum pooling. FC is the fully connected layer. w j means the weight for each layer.
Applsci 13 03261 g003
Figure 4. Our knowledge distillation strategy in Component-aware Graph Routing. CA-GNN means the Component-aware Graph Neural Network. F T and F S are the features of the teacher network and of the student network, respectively. MLAM is the Multi-layer Attention Module.
Figure 4. Our knowledge distillation strategy in Component-aware Graph Routing. CA-GNN means the Component-aware Graph Neural Network. F T and F S are the features of the teacher network and of the student network, respectively. MLAM is the Multi-layer Attention Module.
Applsci 13 03261 g004
Figure 5. Visualization of the components of the human model in FAUST, SCAPE, TOSCA, and KIDS datasets.
Figure 5. Visualization of the components of the human model in FAUST, SCAPE, TOSCA, and KIDS datasets.
Applsci 13 03261 g005
Figure 6. Clustering results of the proposed method on the animal model in the TOSCA dataset. (a) cat, (b) centaur, (c) dog, and (d) horse.
Figure 6. Clustering results of the proposed method on the animal model in the TOSCA dataset. (a) cat, (b) centaur, (c) dog, and (d) horse.
Applsci 13 03261 g006
Figure 7. Component matching results of different models. (a) is the component matching result of the same subject with different poses, and (b) is the component matching result of different subjects.
Figure 7. Component matching results of different models. (a) is the component matching result of the same subject with different poses, and (b) is the component matching result of different subjects.
Applsci 13 03261 g007
Figure 8. Correspondence results on the original FAUST dataset. The four columns on the left are intra-pairs, and the four columns on the right are inter-pairs. In all matched pairs, the left side is the reference shape and the right side is the matching result of our method.
Figure 8. Correspondence results on the original FAUST dataset. The four columns on the left are intra-pairs, and the four columns on the right are inter-pairs. In all matched pairs, the left side is the reference shape and the right side is the matching result of our method.
Applsci 13 03261 g008
Figure 9. Correspondence results on the remeshed FAUST dataset. The first column shows the reference shape, the second column shows ground truth, and columns three through six show the corresponding results of different networks.
Figure 9. Correspondence results on the remeshed FAUST dataset. The first column shows the reference shape, the second column shows ground truth, and columns three through six show the corresponding results of different networks.
Applsci 13 03261 g009
Figure 10. Quantitative correspondence performance of different methods on the original FAUST dataset and the remeshed FAUST dataset. (a,b) are the evaluation on the original FAUST dataset, (c,d) are the evaluation on the remeshed FAUST dateset.
Figure 10. Quantitative correspondence performance of different methods on the original FAUST dataset and the remeshed FAUST dataset. (a,b) are the evaluation on the original FAUST dataset, (c,d) are the evaluation on the remeshed FAUST dateset.
Applsci 13 03261 g010
Figure 11. Quantitative correspondence performance of proposed methods on three remeshed datasets. (a) is the evaluation on the remeshed SCAPE dataset, (b) is the evaluation on the remeshed TOSCA dataset, (c) is the evaluation on the remeshed KIDS dataset.
Figure 11. Quantitative correspondence performance of proposed methods on three remeshed datasets. (a) is the evaluation on the remeshed SCAPE dataset, (b) is the evaluation on the remeshed TOSCA dataset, (c) is the evaluation on the remeshed KIDS dataset.
Applsci 13 03261 g011
Figure 12. Visualization of the corresponding results of our method and other algorithms on the human model in remeshed SCAPE, TOSCA, and KIDS datasets. (a) SCAPE. (b) and (c) TOSCA. (d,e) KIDS. Corresponding points show the same color.
Figure 12. Visualization of the corresponding results of our method and other algorithms on the human model in remeshed SCAPE, TOSCA, and KIDS datasets. (a) SCAPE. (b) and (c) TOSCA. (d,e) KIDS. Corresponding points show the same color.
Applsci 13 03261 g012
Figure 13. Correspondence results on the non-human model in remeshed TOSCA datasets.
Figure 13. Correspondence results on the non-human model in remeshed TOSCA datasets.
Applsci 13 03261 g013
Figure 14. Visualization of geodesic error of several methods on original FAUST dataset.
Figure 14. Visualization of geodesic error of several methods on original FAUST dataset.
Applsci 13 03261 g014
Figure 15. The accuracy of our CA-CGNet and CA-CGNet without knowledge distillation in 10,000 iterations.
Figure 15. The accuracy of our CA-CGNet and CA-CGNet without knowledge distillation in 10,000 iterations.
Applsci 13 03261 g015
Table 1. A comparative study for Capsule Graph Neural Network in open literature.
Table 1. A comparative study for Capsule Graph Neural Network in open literature.
ReferenceRouting MethodsOutcomesLimitations
CapsGNN [26] (2018)Dynamic routingHigher accuracy rate compared to traditional methodsIrrelevant messages from multi-hop neighborhoods has not been restrained
Caps-GNN [27] (2020)Dynamic routingHigher inference in personalized preferenceExternal knowledge has not been considered
HGCN [30] (2021)Nonlinear functionMore effectively capturing the heterogeneous factors under each node.The over-smoothing issue over graph is ignored
CapsGNNEM [28] (2021)EM routingHigher graph classification compared to standard methodsStructural information of the graph has not been considered
NCGNN [29] (2022)Dynamic routingAdaptively identifying a subset of crucial node-level capsulesUnable to preserve structure information of lower-level parts
Caps-HAGKT [31] (2022)Capsule routingExtracting the latent knowledge structure between levelsAutomatic modeling of the complex knowledge structure of the knowledge capsule at same layer is insufficient
OursComponent-aware graph routingUsing component constraints to solve problem of model posture details and low resolutionPerformance can be improved with optimizing selection of the number of components
Table 2. Identification error with a different number of clusters on four datasets. Error is measured as the average Hausdorff distance between the source model and the target model (cm).
Table 2. Identification error with a different number of clusters on four datasets. Error is measured as the average Hausdorff distance between the source model and the target model (cm).
Clustering NumberFAUSTSCAPEKIDSTOSCA
DavidMichaelVictoriaCatCentaurDogHorseWolf
40.06590.08720.08140.06520.07960.06840.03690.04270.03580.03740.0335
50.05860.06970.07050.06130.06860.05290.02540.02910.02160.01950.0208
60.04030.05740.05920.04860.04730.03510.02630.03140.02280.01990.0219
70.02580.02610.02430.01860.01950.02170.02690.03160.02310.02070.0231
80.01930.02010.01890.01740.01810.02030.02710.03200.02350.02140.0233
90.02060.02110.01930.01760.01840.02070.02770.03250.02370.02230.0238
100.02140.02190.01970.01920.01870.02100.02810.03260.02430.02250.0241
Table 3. Average errors of different methods on the original FAUST dataset. The error is expressed as the distance between the mapping point and the ground truth (cm).
Table 3. Average errors of different methods on the original FAUST dataset. The error is expressed as the distance between the mapping point and the ground truth (cm).
MethodIntra AEInter AEAverage
FMNet [12]2.444.833.635
Cyclic-FM [21]2.124.073.095
SP [18]1.573.132.350
3D-CODED [8]1.982.882.430
FARM [22]2.814.123.465
SURFMNet [13]1.733.632.680
MGCN [23]2.513.653.080
ResNet-LDDMM [24]1.932.612.270
Ours1.852.372.110
Table 4. The average geodesic error (cm) with some methods on remeshed FAUST, SCAPE, KIDS, and TOSCA datasets. Average error is normalized to eliminate the effects of shape scale transformations.
Table 4. The average geodesic error (cm) with some methods on remeshed FAUST, SCAPE, KIDS, and TOSCA datasets. Average error is normalized to eliminate the effects of shape scale transformations.
MethodFAUSTSCAPEKIDSTOSCA
DavidMichaelVictoriaCatCentaurDogHorseWolf
FMNet0.36010.37090.34020.34130.37120.31000.37520.35740.37450.35220.0360
SURFMNet0.39520.38950.37270.16730.34850.35720.40840.36590.37920.34960.0545
MGCN0.32110.27550.27520.13810.26310.27870.31530.30150.33980.30450.1779
Ours0.13570.08250.05960.07260.19020.08640.10990.07410.11580.09130.0184
Table 5. Comparison of average geodesic error (cm) on remeshed TOSCA dataset using different routing strategy.
Table 5. Comparison of average geodesic error (cm) on remeshed TOSCA dataset using different routing strategy.
DatasetsDynamic
Routing
Ours without
Knowledge Distillation
Ours
david0.11530.11930.1186
michael0.12950.11360.1129
victoria0.06440.06120.0598
cat0.10580.09490.0941
centaur0.18240.16980.1605
dog0.16250.15790.1572
horse0.11370.11820.1174
wolf0.14710.13540.1287
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lian, Y.; Chen, M. CA-CGNet: Component-Aware Capsule Graph Neural Network for Non-Rigid Shape Correspondence. Appl. Sci. 2023, 13, 3261. https://doi.org/10.3390/app13053261

AMA Style

Lian Y, Chen M. CA-CGNet: Component-Aware Capsule Graph Neural Network for Non-Rigid Shape Correspondence. Applied Sciences. 2023; 13(5):3261. https://doi.org/10.3390/app13053261

Chicago/Turabian Style

Lian, Yuanfeng, and Mengqi Chen. 2023. "CA-CGNet: Component-Aware Capsule Graph Neural Network for Non-Rigid Shape Correspondence" Applied Sciences 13, no. 5: 3261. https://doi.org/10.3390/app13053261

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop