Next Article in Journal
The Relationship between the Facial Expression of People in University Campus and Host-City Variables
Previous Article in Journal
Electricity Market Challenges of Photovoltaic and Energy Storage Technologies in the European Union: Regulatory Challenges and Responses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Graph Convolution Using Heat Kernel for Attributed Graph Clustering

1
Institute of Microelectronics of the Chinese Academy of Sciences, Beijing 100029, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(4), 1473; https://doi.org/10.3390/app10041473
Submission received: 13 January 2020 / Revised: 15 February 2020 / Accepted: 17 February 2020 / Published: 21 February 2020
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:

Featured Application

We propose a novel model to perform attributed graph clustering, which exploits heat kernel to enhance the performance of graph convolution and adopts adaptive architecture to work on different graph datasets. The model proposed in this paper can be deployed to a product recommendation system, where users with specific preferences can be classified precisely and recommended satisfactory products. It can be applied to citation networks to analyze the categories of different articles without prior knowledge. It can be deployed into business forecasting, where the proposed model can identify the operating situation of enterprises significantly by analyzing their business data and investment relationships jointly.

Abstract

Attributed graphs contain a lot of node features and structural relationships, and how to utilize their inherent information sufficiently to improve graph clustering performance has attracted much attention. Although existing advanced methods exploit graph convolution to capture the global structure of an attributed graph and achieve obvious improvements for clustering results, they cannot determine the optimal neighborhood that reflects the relevant information of connected nodes in a graph. To address this limitation, we propose a novel adaptive graph convolution using a heat kernel model for attributed graph clustering (AGCHK), which exploits the similarity among nodes under heat diffusion to flexibly restrict the neighborhood of the center node and enforce the graph smoothness. Additionally, we take the Davies–Bouldin index (DBI) instead of the intra-cluster distance individually as the selection criterion to adaptively determine the order of graph convolution. The clustering results of AGCHK on three benchmark datasets—Cora, Citeseer, and Pubmed—are all more than 1% higher than the current advanced model AGC, and 12% on the Wiki dataset especially, which obtains a state-of-the-art result in the task of attributed graph clustering.

Graphical Abstract

1. Introduction

With the rapid developments of social networks, communication networks, biological networks, and other applications in various fields, the scale of graph data grows sharply. Attributed graphs as the basic data representation contain a large number of node attributes and connection relationships, but it is difficult to perform node classification due to the massive nodes and complicated topology. How to mine the inherent information hidden in the attributed graph without the prior knowledge is a challenging task.
Attributed graph clustering aims to group nodes into different clusters by exploiting node attributes and graph structures sufficiently, where relevant nodes are assigned to the same cluster and the difference between clusters is maximized. By leveraging node attributes and structural information, attributed graph clustering is mainly summarized in three main categories as follows.
Attribute-based clustering such as spectral clustering only respects node attributes and performs clustering on the similarity matrix with node attributes directly. In comparison, structure-based clustering just explores vertex connectivity by manipulating the adjacency matrix of the attributed graph, e.g., random walk [1], Laplacian eigenmaps [2]. However, the above two kinds of graph clustering methods lack incorporated node features and a mining topological structure.
In recent years, generative models based on graph convolutional network (GCN [3]) has been widely studied for attributed graph clustering, where GCN [3] updates the node representation by incorporating neighbor node features to construct graph embedding. Classic generative models include variational graph auto-encoder (VGAE) [4], marginalized graph autoencoder (MGAE) [5], and adversarially regularized graph autoencoder (ARGE) [6]. These methods have been demonstrated as being very practical for performing clustering by unifying graph structures and attributing information. However, graph clustering via GCN [3], such as VGAE [4] or MGAE [5], simply employs shadow two-layer or three-layer graph convolution respectively, which only leverages two or three-hop neighborhood features, and it is hard to capture global structural information. On the contrast, stacking too many layers may lead to over-smoothing and complex computation.
Zhang et al. [7] adopt adaptive graph convolution (AGC) as a low-pass graph filter to make node features smoother. Nevertheless, AGC [7] might not determine the appropriate neighborhood that reflects the relevant information of connected nodes represented in graph structures.
To overcome the above difficulty, we propose an adaptive graph convolution model using heat kernel (AGCHK) to obtain an appropriate neighborhood of the center node and enhance the ability to capture graph smoothness. Our contributions can be summarized as follows:
  • We replace the weak linear low-pass filter in standard AGC [7] by heat kernel to enhance the low-pass characteristics of the graph filter.
  • We leverage the scaling parameter to restrict the neighborhood of the center node, which is flexible to exploit distant-distance nodes while excluding some irrelevant close-distance nodes.
  • We choose the Davies–Bouldin index (DBI) as the criterion to evaluate the cluster quality, which can exactly determine the order of adaptive graph convolution.
  • Experimental results show that AGCHK is obviously superior to other compared methods in the task of attribute graph clustering on benchmark datasets such as Cora, Citeseer, Pubmed, and Wiki.

2. Preliminary

2.1. Problem Formalization

2.1.1. Graph Definition

An undirected graph is represented as G = { V ,   E ,   X } , where V = { v i } i = 1 N consists of a set of nodes with | V | = N . A graph signal x   : V is a real-valued function on the nodes regarded as a vector x N , where x i is the value of x at the i t h node. E is a set of edges, and A = { a i , j } N × N is the adjacency matrix with a i , j = a j , i representing the connection relationships between node v i and node v j . The graph Laplacian matrix is denoted as = D A , where D is the diagonal degree matrix with D i , i = j A i , j , and the normalized graph Laplacian is defined as L = I N D 1 2 A D 1 2 . Since L is a real symmetric positive semidefinite matrix, it can be eigendecomposed as L = U Λ U T where U = [ u 1 , u 2 , , u N ] N × N is a complete set of orthonormal eigenvectors known as graph Fourier modes and Λ = diag { λ i } i = 1 N are ordered real nonnegative eigenvalues associated with { u i } i = 1 N , which are identified as the frequencies of the graph.

2.1.2. Goal

Given an attributed graph G , graph clustering aims to partition nodes V into m disjoint clusters C = { c i } i = 1 m so that nodes within the same cluster are more likely to have similar features and be close to each other, while nodes distributed in different clusters have dissimilar features and are distant to each other.

2.2. Graph Convolution

Here, we briefly review the notions and evolutions of graph convolution. The graph Fourier transform of a graph signal x is defined as x ^ = U T x , and the inverse graph Fourier transform is x = U x ^ [8]. According to the convolution theorem [8], the graph convolution operator represented as G is defined as
x G f = U ( ( U T f ) ( U T x ) ) ,
where f denotes the graph convolution kernel in the spatial domain and is the element-wise Hadamard product.
Spectral CNN [9] replaces U T f by a diagonal matrix g θ so that the Hadamard product can be written as matrix multiplication,
y = x G f = U g θ U T x ,
where L is the normalized graph Laplacian matrix and g θ = diag ( { θ i } i = 1 N ) defined in the spectral domain denotes the frequency response function of the graph filter g θ ( L ) .
However, there are two limitations with the above convolution kernel: (i) it does not have the spatial localization and (ii) it is computationally expensive [10]. To circumvent these issues, g θ is well-approximated by a Chebyshev polynomial in ChebyNet [10], which is defined as
g θ ( Λ ) = p = 0 P 1 θ p Λ p ,
where P is a hyper-parameter and θ p N is the vector of polynomial coefficients. Correspondingly, the graph convolution of ChebyNet [10] is defined as
y = U g θ U T x = U ( p = 0 P 1 θ p Λ p ) U T x = p = 0 P 1 θ p L p x .
GCN [3] only considers the first-order polynomial by setting P = 2 and θ = θ 0 = θ 1 ; thus, the simplified graph convolution is defined as
y = U g θ U T x = θ U ( I Λ ) U T x = θ ( I L ) x ,
where GCN [3] constrains the number of parameters further to address overfitting and accelerate computation.
To strengthen the low-pass performance of the graph filter and improve graph smoothness, AGC [7] modifies the frequency response function of GCN [3] as
y = U g θ U T x = θ U ( I 1 2 Λ ) U T x = θ ( I 1 2 L ) x .

3. Clustering via Adaptive Graph Convolution Using Heat Kernel

Our basic assumption is that connected nodes tend to have similar features or same labels i.e., the graph smoothness. Although previous graph convolution methods gain success by capturing the smoothness of connected nodes, we still need a new methodology to enhance the low-pass performance of graph filter. In this section, we prove that graph smoothness is associated with the eigenvalues of the normalized graph Laplacian matrix, this is, it is relevant to the frequency of graph signal. Next, we analyze previous graph filters from the perspective of signal frequency response and propose AGCHK to circumvent the existing problem of previous filters.

3.1. Motivation

A graph signal x can be linearly represented by the bases of the spectral domain, which are denoted as x = α 1 u 1 + α 2 u 2 + + α N u N , where { α i } i = 1 N is the coefficient of the eigenvector. The smoothness of a basis vector u q corresponding to the graph smoothness is measured by the Laplacian–Beltrami operator Ω ( · ) [11], i.e.,
Ω ( u q ) = 1 2 ( v i , v j ) E a i j u q ( i ) d i u q ( j ) d j 2 2 = u q T L u q = λ q ,
where u q ( i ) means the i t h element of the basis signal u q , d i denotes the degree of node v i , and a i j represents the connected weight between nodes v i and v j . Equation (7) verifies that the basis signals associated with smaller eigenvalues (lower frequencies) are smoother concerning graph structures. Hence, to make the graph signal smoother, graph filters should highlight low-frequency signals by assigning larger weights to low-frequency basis signals, acting as a low-pass filter.
Based on the above principle, we analyze the weakness of previous graph filters in Section 2.2. The frequency response function in ChebyNet [10] is g ( λ q ) = p = 0 P 1 λ q p , which assigns weight λ i p to u i u i T . Since λ q is in ascending order, i.e., 0 λ 1 p λ 2 p λ N p , ChebyNet [10] assigns higher importance to high-frequency basis signals and discounts graph smoothness. GCN [3] defines graph convolution based on the frequency function g ( λ q ) = 1 λ q [12], which cannot perform better low-pass characteristics as g ( λ q ) is negative for 1 < λ q 2 . The frequency response function in AGC [7] is g ( λ q ) = 1 1 2 λ q , which exploits a linear low-pass filter to suppress the high-frequency signal; however, it might not highlight low-frequency signals adequately and cannot determine the appropriate neighborhood that reflects the relevant information of smoothness.

3.2. Adaptive Graph Convolution Using Heat Kernel

3.2.1. Graph Convolution Using Heat Kernel

We propose a novel graph convolution model based on heat kernel to enhance low-frequency signals and suppress high-frequency signals, which can capture the smoothness of node features or labels sufficiently [13]. The heat kernel is defined as
f ( λ q ) = e s λ q ,
where the scaling parameter s > 0 determines the range of heat diffusion. The frequency response function based on heat kernel is defined as g θ = p = 0 P 1 θ p Λ s p [13], where Λ s = d i a g { e s λ q } q = 1 N , and graph convolution via heat kernel denotes
y = U g θ U T x = U ( p = 0 P 1 θ p Λ s p ) U T x = p = 0 P 1 θ p e p s L x .
To preserve the locality of graph convolution based on heat kernel and keep the tradeoff between low-pass performance and computational complexity, the frequency response function g ( λ q ) = 1 + e s λ q is obtained by setting P = 2 , and the associated graph filter is
y = ( I + e s L ) x .
The above graph filter can preserve more low-frequency signals by assigning the weight e s λ q to the basis signal u i u i T , and e s λ q decreases exponentially as λ q increases. Furthermore, for datasets with different node connection structures, we can dynamically regulate the allocation strategy of various basis signal weights by adjusting the scaling coefficient s of the heat kernel.
From the perspective of heat diffusion, ( e s L ) i j reflects the similarity metric, which captures the amount of energy from node v i to node v j , and a target node v j can be regarded as the neighboring node of the center node v i when the similarity metric ( e s L ) i j between nodes v i and v j is higher than the threshold ε , which offers us a more flexible approach to define neighboring nodes of the center node and accelerates the computation by setting the threshold ε . Figure 1 depicted by the Graph Signal Processing toolbox [14] illustrates that the range of heat diffusion controlled by the scaling parameter s becomes larger as s increases. Different from the graph convolution methods in Section 2.2, which constrain neighboring nodes via the shortest path distance, graph convolution based on heat kernel leverages a continuous manner to determine the neighborhood, which can utilize high-order neighbor nodes sufficiently and discard some irrelevant low-order neighbor nodes [13].
Compared with the linear frequency response function g ( λ q ) = 1 1 2 λ q proposed in AGC [7], we leverage the exponential frequency function g ( λ q ) = 1 + e s λ q based on heat kernel, which highlights low-frequency signals exponentially to make the graph smoother. Figure 2 illustrates g ( λ q ) = 1 + e s λ q assigns larger weights to low-frequency basis signals compared with g ( λ q ) = 1 1 2 λ q so that connected nodes have more similar features, which indicates that graph convolution based on heat kernel might perform graph clustering better.

3.2.2. K-Order Adaptive Graph Convolution

First-order graph convolution is not sufficient to capture graph smoothness, since it updates the center node through aggregating a 1-hop neighborhood only, which might not suitable for large and sparse graphs. To make full use of node features and global structural information, we exploit k -order graph convolution, which is defined as
X ¯ = ( I + e s L ) k X ,
where k > 0 is the order of graph convolution. Figure 2 shows that g ( λ q ) = ( I + e s λ q ) k becomes more low-pass as k increases; that is, the filtered graph will be smoother. In particular, we normalize the node features at each iteration ( l 1 or   l 2 ), which scales the input vector to the unit norm.

3.2.3. Cluster Evaluation Index

According to Equation (11), the representation of connected nodes will be more similar as k increases, and we perform spectral clustering on node features X ¯ filtered by graph convolution in each iteration. In detail, the pairwise similarity between nodes is measured as S = 1 2 ( | K | + | K | T ) [15], where K = X ¯ X ¯ T denotes a linear kernel. Since K is symmetric and nonnegative, learning the eigenvectors of K is equivalent to computing the left singular vectors of X via singular value decomposition (SVD). Thus, we perform k -means on the left singular vectors associated with the m largest eigenvalues of X directly to obtain cluster partitions. Moreover, we do multiple spectral clustering in each iteration to preserve stable cluster partitions.
As the iteration goes on, the features of connected nodes will become more and more similar; this is, the intra-cluster distance is getting smaller. Figure 3 illustrates that nodes with different labels are mixed together when iteration number k is smaller, while node features will be over-smoothing when iteration number k is larger. To determine an appropriate iteration number k comprehensively, we adopt Davies–Bouldin index (DBI) [16] as the criterion to evaluate cluster quality, which is defined as
D B I ( C k ) = 1 m i = 1 m max i j d i + d j r i j .
Denote by m the number of clustering partitions, d i the average distance between each point of a cluster and the cluster centroid i , and   r i j the distance between cluster centroids i and j . The score is defined as the average similarity measure of each cluster with its most similar cluster, and clusters that are farther apart and less dispersed will result in a better (smaller) score. To stop iterating in time, we choose k corresponding to the first local minimum of D B I ( C k ) as the most appropriate iteration number. More intuitively, considering d _ D B I ( k ) = D B I ( C k + 1 ) D B I ( C k ) , we stop iterating immediately once d _ D B I ( k ) > 0 as the iteration number k increases and obtain the final cluster partition C k . By leveraging the above strategy, AGCHK is able to capture the representation of graphs with different structures adaptively and avoid over-smoothi

3.2.4. Architecture and Algorithm

Based on the design and analysis of the previous chapter, we summarize the proposed AGCHK algorithm as follows. Firstly, we construct a low-pass filter based on the heat kernel, which is defined as ( I + e s L ) k , where k = 1 . Next, we exploit the low-pass filter to filter the original graph signals to obtain a smoother graph representation. Then, we can obtain the left singular vectors U of X ¯ k by the singular value decomposition and perform k -means on U 10 times, getting the cluster partition C k . Finally, we calculate the DBI of the cluster partition C k ; if the DBI decreases, we set k = k + 1 and continue the above loop, else, we stop the loop, and the final cluster partition is C k . Figure 4 is the architecture of the novel AGCHK model. Algorithm 1 describes in detail the algorithm of obtaining the better cluster partition by performing AGCHK. We utilize the low-pass filter based on heat kernel in the adaptive graph convolution architecture to enhance the smoothness of the graph representation, which can adaptively determine the order of graph convolution and does not require training parameters, unlike the other methods based on graph neural network [4,5,6].
Algorithm 1 AGCHK
Input: Node features X , adjacency matrix A , and maximum iteration number m a x _ i t e r .
Output: Cluster partition C .
1.
Initialize k = 0 ,   r e p = 0 , R E P = 10 , Compute the normalized graph Laplacian L = I N D 1 2 A D 1 2 .
2.
Compute graph filter I + e s L .
3.
repeat
4.
  Set k = k + 1 .
5.
  Calculate   k -order graph convolution by Equation (11) and obtain filtered features X ¯ k .
6.
   X ¯ k = n o r m a l i z e ( X ¯ k ) .
7.
  Obtain the left singular vectors U   of   X ¯ k by SVD.
8.
  repeat
9.
    Set r e p = r e p + 1
10.
    Perform k -means on U and obtain clustering partition C k .
11.
    Calculate D B I ( C k ) [ r e p ] by Equation (12).
12.
     until   r e p =   R E P
13.
  Compute the mean of R E P partition scores D B I ( C k ) [ 0 : r e p ] .
14.
until   d _ D B I ( k ) > 0 or t > m a x _ i t e r
15.
return final clustering partition C k .

3.3. Algorithm Time Complexity

According to Algorithm 1, denote by D the number of node features, m the number of clusters, N the number of nodes, and r e p the number of clustering in each iteration. The graph filter   I + e s L can be efficiently approximated by Chebyshev polynomials without requiring the eigendecomposition of the graph Laplacian matrix [17], and the computational complexity is O ( P | E | ) , where P is the order of Chebyshev polynomial and | E | is the number of edges. Such a linear complexity makes methods based on heat kernel applicable to large-scale networks [3]. After k iterations, the time complexity of calculating Equation (11) k times is   O ( P | E | + N D k ) , performing spectral clustering on X k is O ( N 2 D k + N 2 m k ) , and computing D B I ( C k ) is O ( 1 m N 2 D k + m 2 k ) . Note that for a spare A, m D ,   m N 2 , the overall time complexity of AGCHK is O ( P | E | + N D k + N 2 D k ) . AGCHK is more time-efficient than the clustering methods based on graph neural networks, since AGCHK does not need to train parameters.

4. Experiments

4.1. Datasets

To verify the effectiveness and benefit of the proposed AGCHK for attributed graph clustering, we conduct experiments on four benchmark datasets. The dataset details are demonstrated in Table 1. Cora, Citeseer, and Pubmed [4] are citation networks whose nodes represent documents and edges are citation links. Wiki [17] is a webpage network, whose nodes are webpages and edges are link relations. The node features of Cora and Citeseer are binary word vectors, and the node features of Pubmed and Wiki are computed by the term frequency–inverse document frequency.

4.2. Baselines and Evaluation Metrics

To highlight the performance of AGCHK, we choose the same benchmark methods as AGC [7].
  • Methods that only exploit node features: classic spectral clustering methods such as k -means and spectral-f, which perform clustering on the similarity matrix constructed by node features directly.
  • Clustering methods that only handle graph structures: spectral-g clustering on the graph adjacency matrix simply, Deepwalk [2], and graph neural networks for graph representations (DNGR) [18].
  • Attributed graph clustering methods utilize graph structures and graph node features jointly: AGC [7], which does not need to train graph neural network parameters, Graph neural network methods based on autoencoder such as graph autoencoder (GAE) and graph variational autoencoder (VGAE) [4], marginalized graph autoencoder (MGAE) [5], adversarially regularized graph autoencoder (ARGE), and variational graph autoencoder (ARVGE) [6].
To measure the ability of the model comprehensively, we adopt the following three cluster evaluation indexes [19]: graph clustering accuracy (Acc), normalized mutual information (NMI), and macro F1-score (F1).

4.3. Parameter Settings

For AGCHK, we set the maximum number of iterations m a x _ i t e r to 20. According to Section 3.2.1, the range of heat diffusion becomes larger as the scaling parameter s increases. Cora and Citeseer might have similar parameter settings since they have the close node and edge sizes, and the parameter s might be smaller due to their lower quantity. Pubmed has more nodes and edges; thus, its parameter s might be larger. Since Wiki has more edges and fewer nodes, we set s very small so that AGCHK can leverage a fine structure. In view of the above analysis, for Cora, s = 3.3 and ε = 10 4 , for Citeseer, s = 2 and ε = 10 5 , for Pubmed, s = 8 and ε = 10 5 , and for Wiki, s = 0.5 and ε = 10 5 .
For other baseline methods, we keep the same parameter settings as the original papers. For AGC [7], we set m a x _ i t e r to 60. For Deepwalk [2], denote by 10 the number of random walks, 128 the number of latent dimensions of each node, and 80 the path length of each random walk. For DNGR [18], the autoencoder is constructed by three layers, where the hidden layers are 512 neurons and 256 neurons respectively. For GAE and VGAE [4], we construct encoders with 32-neuron hidden and 16-neuron hidden layers respectively, and we use the Adam optimizer to train the encoders for 200 iterations with a learning rate 0.01. For MGAE [5], denote by 0.4 the degree of corruption level p , 3 the number of layers, and 10 5 the coefficient λ . For ARGE and ARVGE [6], we build encoders with 32-neuron and 16-neuron hidden layers respectively, and their discriminators consist of two hidden layers, which are composed of 16 neurons and 64 neurons, respectively. On Cora, Citeseer, and Wiki, we use Adam optimizer to train ARGE and ARVGE [6] for 200 iterations, where their encoder and discriminator learning rate are both 0.001. On Pubmed, we train ARGE and ARVGE [6] for 2000 iterations, where the learning rates of the encoder and the discriminator are 0.001 and 0.008, respectively.

4.4. Result Analysis

To obtain stable clustering results, we repeat 10 experiments for each method, and the average clustering results are shown in Table 2. Since autoencoder-based methods make use of node features and graph structures jointly, they perform better than classic spectral clustering methods that exploit node features or graph structures independently. Unfortunately, these autoencoder-based methods only utilize the 2-hop or 3-hop neighborhoods of the center node to update node representation, which is insufficient to capture global structures.
AGC [7] achieves better clustering results since it exploits adaptive graph convolution to select k-hop neighbors to aggregate information and update the central node representation. However, for Wiki that is more densely connected than other datasets, AGC [7] cannot perform well because of its weaker ability to capture smoothness.
As for our AGCHK model, it outperforms all the baseline methods, achieving state-of-the-art results on the four datasets, especially on Wiki. AGCHK implements such a smooth adaptive graph convolution by enhancing low-frequency basic filters and discounting high-frequency basic filters that it makes the representation of connected nodes smoother. Furthermore, the scaling parameter s of the heat kernel is flexible to adjust the diffusion range to suit different applications and different networks. Especially for Wiki, AGCHK achieves its superiority significantly because it sets the scaling parameter s smaller to leverage fine graph structure information.
To verify the reliability of the proposed cluster evaluation index d _ D B I ( k ) > 0 , we plot d _ D B I ( k ) and the clustering performance w.r.t. k on Cora and Wiki respectively in Figure 4. The intra-cluster distance decreases and the inter-cluster distance increases in the early iteration, which is corresponding to d _ D B I ( k ) < 0 . However, too many iterations will make node features over-smoothing, which leads to the inter-cluster distance rises. The cluster evaluation d _ D B I ( k ) takes intra and inter-cluster distance comprehensively, and it can select the most appropriate k to stop iterating, as shown in Figure 5. One can see that the Acc, NMI, and F1 scores evaluating clustering performance are the best or close to the best when d _ D B I ( k ) is greater than zero the first time, which demonstrates the validity of the proposed selection evaluation index d _ D B I ( k ) > 0 . The selected iteration number k for Cora, Citeseer, Pubmed, and Wiki is 9, 9, 13, and 8 respectively, which are all values that are lower than the respective k values on these datasets in AGC [7]—12, 55, 60, and 8.
AGCHK performs very stable on benchmark datasets, and the standard deviations of Acc, NMI, and F1 are 0.18%, 0.35%, and 0.24% on Cora, 0.00%, 0.00%, and 0.00% on Citeseer, 0.01%, 0.01%, and 0.01% on Pubmed, and 0.05%, 0.03%, and 0.02% on Wiki, which are more stable than AGC [7].
We compare the time efficiency of several baseline methods as Table 3, and the best score is in bold, while the second best score is underlined. We can see that the running time of AGCHK and AGC are comparable, while AGCHK is more than three times faster than the other methods. Since AGCHK does not need to train parameters, it is more efficient than the baselines based on graph neural networks.

4.5. Influence of Hyper-Parameters s and ε

To demonstrate the flexibility of the scaling parameter s and the threshold   ε , we perform experiments on Cora as shown in Figure 6. For the fixed threshold ε , Acc scores first increase and then decrease as the scaling parameter s rises, i.e., sufficient neighborhood information cannot be employed when s is small. On the contrast, the range of heat diffusion is large and different clusters cannot be well distinguished when s is comparatively large.
For the fixed threshold s , good clustering results can be obtained with s in a large range when ε is small, as many neighbors with weak relationships are excluded. However, as ε becomes larger, a large number of neighbor nodes with strong relationships are dropped, resulting in poor clustering. Briefly, by setting a smaller threshold ε , noise nodes will be ignored and computation can be accelerated.

5. Conclusions

In this paper, we improve AGC [7] by utilizing heat kernel instead of the original weak linear kernel, which makes the low-pass performance of the graph filter better. The scaling parameter s of heat kernel can adjust the range of heat diffusion effectively so that it is suitable for various datasets with different node and edge sizes. Besides, we redesign the clustering criterion to achieve the best clustering results in fewer k -order graph convolution. Our proposed method AGCHK has reached advanced levels in all four baseline datasets.

Author Contributions

Conceptualization, D.Z.; methodology, D.Z. formal analysis, D.Z. and X.M.; validation, D.Z. and X.M.; visualization, D.Z.; data curation, D.Z. and X.M., writing—original draft preparation, D.Z.; writing—review and editing, D.Z., R.D. and S.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Strategic Priority Research Program of Chinese Academy of Sciences, grant number XDC02070600.

Acknowledgments

We would like to thank Boyue Wang, Shaojie Li, the editor, and the anonymous reviewers for their insightful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Newman, M. Finding community structure in networks using the eigenvectors of matrices. Phys. Rev. E 2006, 74, 036104. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  3. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  4. Kipf, T.N.; Welling, M. Variational graph auto-encoders. In Proceedings of the NIPS Workshop on Bayesian Deep Learning, Barcelona, Spain, 10 December 2016. [Google Scholar]
  5. Wang, C.; Pan, S.; Long, G.; Zhu, X.; Jiang, J. Mgae: Marginalized graph autoencoder for graph clustering. In Proceedings of the 2017 ACM Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 889–898. [Google Scholar]
  6. Pan, S.; Hu, R.; Long, G.; Jiang, J.; Yao, L.; Zhang, C. Adversarially regularized graph autoencoder for graph embedding. In Proceedings of the Twenty-Seventh International Joint Conferences on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 2609–2615. [Google Scholar]
  7. Zhang, X.; Liu, H.; Li, Q.; Wu, X.-M. Attributed graph clustering via adaptive graph convolution. In Proceedings of the International Joint Conferences on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 4327–4333. [Google Scholar]
  8. Shuman, D.I.; Narang, S.K.; Frossard, P.; Ortega, A.; Vandergheynst, P. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Process. Mag. 2013, 30, 83–98. [Google Scholar] [CrossRef] [Green Version]
  9. Bruna, J.; Zaremba, W.; Szlam, A.; Lecun, Y. Spectral networks and locally connected networks on graphs. In Proceedings of the 2nd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  10. Defferrard, M.; Perraudin, N.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016; pp. 3837–3845. [Google Scholar]
  11. Chung, F.R.; Graham, F.C. Spectral Graph Theory; American Mathematical Society: Providence, RI, USA, 1997. [Google Scholar]
  12. Li, Q.; Wu, X.; Liu, H.; Zhang, X.; Guan, Z. Label efficient semi-supervised learning via graph filtering. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9582–9591. [Google Scholar]
  13. Xu, B.; Shen, H.; Cao, Q.; Cen, K.; Cheng, X. Graph convolutional networks using heat kernel for semi-supervised learning. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 1928–1934. [Google Scholar]
  14. Perraudin, N.; Paratte, J.; Shuman, D.; Martin, L.; Kalofolias, V.; Vandergheynst, P.; Hammond, D.K. Gspbox: A toolbox for signal processing on graphs. arXiv 2014, arXiv:1408.5781. [Google Scholar]
  15. Nikolentzos, G.; Meladianos, P.; Vazirgiannis, M. Matching node embeddings for graph similarity. In Proceedings of the Thirty-First Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 2429–2435. [Google Scholar]
  16. Davies, D.L.; Bouldin, D.W. A Cluster Separation Measure. IEEE Trans. Pattern Anal. Mach. Intell. 1979, 1, 224–227. [Google Scholar] [CrossRef] [PubMed]
  17. Yang, C.; Liu, Z.; Zhao, D.; Sun, M.; Chang, E.Y. Network representation learning with rich text information. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 2111–2117. [Google Scholar]
  18. Cao, S.; Lu, W.; Xu, Q. Deep neural networks for learning graph representations. In Proceedings of the Thirtieth Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1145–1152. [Google Scholar]
  19. Aggarwal, C.C.; Reddy, C.K. Data Clustering: Algorithms and Applications; CRC Press: Boca Raton, FL, USA, 2014. [Google Scholar]
Figure 1. Heat diffusion comparison as s increases on an example graph. The center node is represented in yellow, and the similarity between the neighbor nodes and the center node decreases as the node color becomes darker. The range of heat diffusion increases from a small scale (a) to a large scale (b).
Figure 1. Heat diffusion comparison as s increases on an example graph. The center node is represented in yellow, and the similarity between the neighbor nodes and the center node decreases as the node color becomes darker. The range of heat diffusion increases from a small scale (a) to a large scale (b).
Applsci 10 01473 g001
Figure 2. The linear low-pass filters in adaptive graph convolution (AGC) [7] and heat kernel comparison.
Figure 2. The linear low-pass filters in adaptive graph convolution (AGC) [7] and heat kernel comparison.
Applsci 10 01473 g002
Figure 3. Spectral clustering visualization of k -order graph convolution for dataset Cora. Colors of different nodes represent various labels, and the representation of nodes with the same label is closer as the value of k increases. Low-order graph convolution makes node features indistinguishable, while high-order graph convolution might cause node features over-smoothing.
Figure 3. Spectral clustering visualization of k -order graph convolution for dataset Cora. Colors of different nodes represent various labels, and the representation of nodes with the same label is closer as the value of k increases. Low-order graph convolution makes node features indistinguishable, while high-order graph convolution might cause node features over-smoothing.
Applsci 10 01473 g003
Figure 4. The architecture of proposed adaptive graph convolution using heat kernel model for attributed graph clustering (AGCHK).
Figure 4. The architecture of proposed adaptive graph convolution using heat kernel model for attributed graph clustering (AGCHK).
Applsci 10 01473 g004
Figure 5. d _ D B I ( k ) , i n t r a ( k ) , i n t e r ( k ) and clustering performance Acc, NMI, and F1 w.r.t. k on datasets Cora and Wiki.
Figure 5. d _ D B I ( k ) , i n t r a ( k ) , i n t e r ( k ) and clustering performance Acc, NMI, and F1 w.r.t. k on datasets Cora and Wiki.
Applsci 10 01473 g005
Figure 6. The Acc scores of clustering performance using the different scaling parameter s and threshold ε on Cora.
Figure 6. The Acc scores of clustering performance using the different scaling parameter s and threshold ε on Cora.
Applsci 10 01473 g006
Table 1. Statistics of datasets.
Table 1. Statistics of datasets.
DatasetNodesEdgesFeaturesClasses
Cora2708542914337
Citeseer3327473237036
Pubmed19,71744,3385003
Wiki240517,981497317
Table 2. Clustering performance. The best score is in bold, while the second best score is underlined. DNGR: graph neural networks for graph representations, GAE: graph autoencoder, VGAE: variational graph auto-encoder, MGAE: marginalized graph autoencoder, ARGE: adversarially regularized graph autoencoder, ARVGE: variational graph autoencoder, AGC: adaptive graph convolution.
Table 2. Clustering performance. The best score is in bold, while the second best score is underlined. DNGR: graph neural networks for graph representations, GAE: graph autoencoder, VGAE: variational graph auto-encoder, MGAE: marginalized graph autoencoder, ARGE: adversarially regularized graph autoencoder, ARVGE: variational graph autoencoder, AGC: adaptive graph convolution.
MethodsInputCoraCiteseerPubmedWiki
Acc%NMI%F1%Acc%NMI%F1%Acc%NMI%F1%Acc%NMI%F1%
k -meansFeature34.6516.7325.4238.4917.0230.4757.3229.1257.3533.3730.2024.51
Spectral-fFeature36.2615.0925.6446.2321.1933.7059.9132.5558.6141.2843.9925.20
Spectral-gGraph34.1919.4930.1725.9111.8429.4839.743.4651.9723.5819.2817.21
DeepwalkGraph46.7431.7538.0636.159.6626.7061.8616.7147.0638.4632.3825.38
DNGRGraph49.2437.2937.2932.5918.0244.1945.3515.3817.9037.5835.8525.38
GAEBoth53.2540.6941.9741.2618.3429.1364.0822.9749.2617.3311.9315.35
VGAEBoth55.9538.4541.5044.3822.7131.8865.4825.0950.9528.6730.2820.49
MGAEBoth63.4345.5738.0163.5639.7539.4943.888.1641.9850.1447.9739.20
ARGEBoth64.0044.9061.9057.3035.0054.6059.1223.1758.4141.4039.5038.27
ARVGEBoth63.8045.0062.7054.4026.1052.9058.2220.6223.0441.5540.0137.80
AGCBoth68.9253.6865.6167.0041.1362.4869.7831.5968.7247.6545.2840.36
AGCHKBoth70.56 ± 0.1855.44 ±   0.35 67.09 ±   0.24 68.35 ±   0.00 42.25 ±   0.00 63.89 ±   0.00 70.82 ±   0.01 32.40 ±   0.01 69.95 ±   0.01 60.13 ±   0.05 55.47 ±   0.03 46.27 ±   0.02
Table 3. The running time of related methods performing on benchmark datasets. For other baselines, we follow the parameter settings in the original papers. Hardware configuration: Intel(R) Core(TM) i7-7700K CPU, 16.0 GB RAM and no GPU.
Table 3. The running time of related methods performing on benchmark datasets. For other baselines, we follow the parameter settings in the original papers. Hardware configuration: Intel(R) Core(TM) i7-7700K CPU, 16.0 GB RAM and no GPU.
MethodsCoraCiteseerPubmedWiki
GAE38.72 s57.95 s2265.55 s-
VGAE41.66 s61.34 s2433.76 s-
ARGE48.49 s68.59 s2021.94 s-
ARVGE43.16 s62.33 s1850.21 s-
AGC17.46 s111.04 s151.71 s22.09 s
AGCHK8.23 s22.92 s854.55 s23.64 s

Share and Cite

MDPI and ACS Style

Zhu, D.; Chen, S.; Ma, X.; Du, R. Adaptive Graph Convolution Using Heat Kernel for Attributed Graph Clustering. Appl. Sci. 2020, 10, 1473. https://doi.org/10.3390/app10041473

AMA Style

Zhu D, Chen S, Ma X, Du R. Adaptive Graph Convolution Using Heat Kernel for Attributed Graph Clustering. Applied Sciences. 2020; 10(4):1473. https://doi.org/10.3390/app10041473

Chicago/Turabian Style

Zhu, Danyang, Shudong Chen, Xiuhui Ma, and Rong Du. 2020. "Adaptive Graph Convolution Using Heat Kernel for Attributed Graph Clustering" Applied Sciences 10, no. 4: 1473. https://doi.org/10.3390/app10041473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop