Next Article in Journal
Realistic Aspects of Simulation Models for Fake News Epidemics over Social Networks
Previous Article in Journal
Characterization of the Digital Identity of Chilean University Students Considering Their Personal Learning Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dirichlet Process Prior for Student’s t Graph Variational Autoencoders

1
College of Computer Science and Technology, Jilin University, Changchun 130012, China
2
Key Laboratory of Symbolic Computation and Knowledge Engineer, Ministry of Education, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Future Internet 2021, 13(3), 75; https://doi.org/10.3390/fi13030075
Submission received: 10 February 2021 / Revised: 6 March 2021 / Accepted: 11 March 2021 / Published: 16 March 2021

Abstract

:
Graph variational auto-encoder (GVAE) is a model that combines neural networks and Bayes methods, capable of deeper exploring the influential latent features of graph reconstruction. However, several pieces of research based on GVAE employ a plain prior distribution for latent variables, for instance, standard normal distribution (N(0,1)). Although this kind of simple distribution has the advantage of convenient calculation, it will also make latent variables contain relatively little helpful information. The lack of adequate expression of nodes will inevitably affect the process of generating graphs, which will eventually lead to the discovery of only external relations and the neglect of some complex internal correlations. In this paper, we present a novel prior distribution for GVAE, called Dirichlet process (DP) construction for Student’s t (St) distribution. The DP allows the latent variables to adapt their complexity during learning and then cooperates with heavy-tailed St distribution to approach sufficient node representation. Experimental results show that this method can achieve a relatively better performance against the baselines.

1. Introduction

Variational autoencoders (VAEs), graph variational autoencoders (GVAEs) and their variants [1,2,3], the deep generative models that have emerged in recent years, integrate the auto-encoder and variational inference, not only keep the excellent expression of the neural network but solve the randomness of probabilistic latent variables in the generation process. At present, the development of VAEs in the probability models has two primary directions. The first is to design a more expressive variational posterior, the other one is to replace the standard normal distribution prior of VAEs with learning more intricate prior [4,5,6,7,8]. Although works in the first direction have made some signs of progress in various kinds of learning tasks, delving into learning more about whether the variational posterior is valid is still interesting. Reference [9] proved that even with a very strong posterior, VAEs always can not match the learned posterior with N(0, I) prior. So we focus on learning a prior with abundant semantics for GVAEs.
The performance of GVAEs is mainly dependent on the validity of their data representation. Learning the representation of graph data, also known as network representation learning (NRL) [10,11,12], encodes nodes in the network and denotes them in a unified low-dimensional latent space. Hence, one way to improve GVAEs prior is to use more flexible probability distribution modeling latent variables. Gaussian distribution [13,14,15,16] is the most widely used probability distribution of latent variables in VAE models [17,18,19]. This kind of phenomenon is because of the well-known “central-limit theorem” [20]. If the sample size is large enough, the distribution of the samples’ mean is approximately Gaussian, irrelevant to the distribution of samples in the population. The mean value of the samples’ mean distribution does not have to be zero, it depends on the value range of samples. More specifically, if there are a set of non-negative samples, their mean value is also non-negative. In addition, Gaussian distribution is the natural selection in the case of known mean and variance owing to its information entropy is the largest under this condition. However, it also has some disadvantages, especially its “empirical rule”, which makes up about 99.7 % of the numerical values distributed within three standard deviations centered on the mean. When the latent space dimension increases, its ability to capture more information will decline. Nevertheless, the choice of heavy-tailed Student’s t (St) distribution [21,22,23,24] can significantly remedy this defect. For the reason that it can adaptively model the networks according to the degree of freedom (dof) parameter. The dof parameter controls the curve shape changes of the St distribution. The smaller the dof, the flatter the St distribution curve. Conversely, the larger the dof, the curve is closer to the Gaussian distribution with the same mean and scale parameters. Because most of the real-world networks are not random, the degree distribution in different networks more or less has the power-law property [25,26]. For instance, in the American airline network, nodes are airports, and the links between them are airlines. The nodes such as John F. Kennedy International Airport and Los Angeles International Airport are hubs with the power-law property. However, there are just several hubs in a network, and most nodes have a handful of links [27]. In addition, each latent vector is a node representation, so that we can fit the structure of real-world networks by applying the more general St distribution to model node representations. Concretely, ensuring the variational posterior of latent representation is close to a St distribution with zero mean and one scale, while dof remains unchanged [4]. They solve the problem that Gaussian prior never rejects outliers in the modeling. Furthermore, they deal with data with image characteristics, which is different from this paper.
As we probe into the data with St distribution, it is a pivotal point that how to statistically model data at an applicable complexity. If finite models used for analysis, we must specify the number of latent features in advance, but it is unknown beforehand and needs to learn from the data. In the paper citation network, latent features are represented by the keywords of a paper. If the latent dimension is small, the latent features are general keywords, such as data mining and the uncertainty of artificial intelligence. If the latent dimension is large, the latent features may be more specific keywords, such as anomaly detection, Bayesian network. By adding latent dimensions, the topic related to the paper can be dug out; then we can find neglected correlation relationship of some papers. However, if the latent dimension is too large, some useless latent features will appear, such as Chebyshev’s inequality. Papers that apply the same inequality are regarded as relevant, but this decision is inaccurate in most cases.
The Bayesian nonparametric approach [28,29,30,31] gives us a way to solve this problem, which assumes that latent structures grow with data so that an infinite number of latent features can exist in the model. Under this condition, the determination of model complexity is incorporated into the data analysis process without setting manually based on expert knowledge or supervision information. The Dirichlet Process (DP) [32,33,34] is a typical Bayesian nonparametric method, which defines a binary matrix and each row of the matrix represents a node representation, each dimension captures a specific aspect of nodes. DP, as a prior of St distribution, can find possible features of all nodes in networks and also help discover important missing information or links.
In this paper, we propose DP-St prior for GVAE. In detail, we use St distribution as the probability distribution of a latent variable and solve the problem of limited detection ability of the traditional Gaussian distribution. At the same time, we apply the DP to explore more prior knowledge of nodes in a network, solve the problem that a fixed capacity can not precisely capture the complexity of the available networks well. We integrate DP and St distribution to construct an unbounded latent representation with sufficient expression ability, longer tails, and a deeper dimension to provide tremendous flexibility for latent variables. Our contributions are summarized as follows:
  • We proposed DP-St prior, DP is the prior of St distribution and combined it with GVAE for network representation learning. The DP-St prior is a heavy-tailed nonparametric process, which can adapt to the dimension of latent variables, generate manifold prior knowledge for GVAE;
  • We illustrate that the adjustment of the heavy-tailed level of St distribution is conducive to filter out useless latent features and retain the most representative ones. It can help detect missing links in the network and accurately model the real-world networks.
  • In the experiments, the improvement of link prediction and semi-supervised node classification tasks are demonstrated. In particular, our model can effectively improve the prediction accuracy of link prediction in low-dimensional latent variables caused by St distribution modeling. It can also obtain the optimal prediction results in the semi-supervised node classification task when the labeled training data is small.

2. Related Work

Exploring the proper form of latent representation is an essential task for many representation learning applications. Moreover, we have seen that learning a prior of latent variables is indispensable to the VAE models. From the theoretical perspective, some recent studies demonstrate the necessity of learning VAE prior. Reference [35] highlighted the significance of the probability distribution of latent space by analyzing some variants of variational evidence lower bound objective (ELBO). Since it is necessary to complete the VAE optimization process by maximizing the ELBO, and this process is equivalent to minimizing the KL divergence between the variational posterior and prior of latent variables. Furthermore, Reference [36] illustrates the negative impact of applying a simple prior for VAE models. That is to say, it will result in few active latent variables, and finally produce over-regularized models with unpromising latent representation, which is not conducive to the tasks of presentation learning.
Considering the importance of latent representation for carrying out various learning tasks, some models improved VAE from the perspective of the Gaussian distribution of latent variables, which is commonly used in most VAE models and their variants. Reference [37] elaborate that applying the nonparametric process to a neural network can automatically ascertain some critical parameters in the model, such as model capacity. They use the stick-breaking process to generate latent variables, and the Stick-Breaking VAE model is constructed by various parameterization methods. This model shows more discriminative ability of latent representation than Gaussian VAEs. Reference [4] replace the Gaussian distribution of latent variables in the traditional VAE models with a higher error-tolerant St distribution and apply the standard St distribution as a prior, and demonstrate that heavy-tailed latent variables can aggrandize their flexibility when the data contains outliers. Nonetheless, only aggrandizing the tail thickness is far from enough to make the latent variables adapt to the dimension with sufficient flexibility and obtain ample prior knowledge to represent the data. Nonetheless, only enlarging the tail thickness is far from enough to make the latent variables adapt to the dimension with sufficient flexibility and obtain ample prior knowledge to represent the data. All the works mentioned above are used to deal with tasks related to image data, while we intend to do the presentation learning tasks on network data.
There are also a few works that employ the DP prior. In [38], they focus on the task of various classification through learning logistic-regression methods, and they illustrate that the DP is helpful to learn the similarity between different classification tasks. The DP-GP-LVM (Dirichlet Process-Gaussian Process-Latent Variable Models) [39] algorithm presents a Bayesian nonparametric latent variable model, apply the Gaussian process prior and DP prior at the same time. More precisely, they are used for producing latent variables and learning structure, respectively. However, neither of these works has considered the application of DP prior to latent variables. Our model takes full advantage of DP as a nonparametric method in the adaptive dimension and facilitates the learning of appropriate representation for latent variables.

3. Background

3.1. Dirichlet Process

In the network representation learning, it is pivotal to learn an appropriate representation for each node, which is demonstrated by a binary vector, every dimension of it indicates a latent feature the node may possess. The Bayesian nonparametric method is conducive to determining the number of latent features. The Dirichlet process (DP), as a widely used Bayesian nonparametric method, places a prior probability distribution over the sparse latent feature matrix, denoted Z D P ( α ) , whose number of rows is finite and columns are unbounded. The DP is defined as:
Z n G G , G D P ( α , G 0 ) ,
where n = 1 , 2 , . . . , N for N nodes in a network. The node representation Z n is drawn from a prior probability distribution G. G is sampled from the DP with a base distribution G 0 and a positive scale parameter α .
In addition, the product of DP is a discrete distribution. Hence, we can apply the stick-breaking process to represent it more intuitively. The stick-breaking process is used to generate infinite-dimensional discrete variables. This procedure requires sampling a vector from a base distribution B e t a ( 1 , α ) over [ 0 , 1 ] for each node, and all elements of the vector sum to one. Then, Reference [40] takes this process as the basis for the stick-breaking generation, and completes the DP with stick-breaking construction by arranging feature weights in descending order:
b ( k ) B e t a ( 1 , β ) , μ ( k ) = b ( k ) l = 1 k 1 ( 1 b ( l 1 ) ) ,
where k = 1 , 2 , . . . , K , β is the hyperparameter of Beta distribution, b ( k ) is the kth sample from the Beta distribution, μ ( k ) is the kth weight from the DP process, and μ ( 1 ) > μ ( 2 ) > . . . > μ ( K ) indicates the descending ordering of μ 1 , μ 2 , . . . , μ K .

3.2. Student’s t Distribution

In the real world, many observed phenomena do not follow the frequently-used Gaussian distribution, and a more permissive distribution should be used to describe them. The Student’s t distribution (St) [41], as the most frequently used heavy-tailed distribution, differs significantly from the Gaussian distribution, the thickness of tails is modulated by the dof parameter υ . The smaller the dof parameter υ , the flatter the St distribution, meaning the lower the middle of the curve, the higher the tails rise. In Figure 1 and Figure 2, we compare the St distribution with other similar probability distributions. As the dof υ = 1 , the curve is the Cauchy distribution, and as the dof υ = , the curve is the Gaussian distribution. Overall, the lower dof υ in distribution will result in the more permissive of outliers. The probability density function (pdf) of the multivariate St distribution is:
f ( x = Γ ( υ + n 2 ) Γ ( υ 2 ) ( π υ ) n d e t ( Σ ) [ 1 + ( x μ ) T Σ 1 ( x μ ) υ ] υ + n 2 ,
where random variable x R n , with mean parameter μ R n , scale matrix Σ R n × n , dof parameter υ , the diagonal elements of Σ are σ 2 , and n is the dimension of x.
We can use the reparameterization trick to reparameterize two samples u N ( 0 , I ) , w G a m m a ( u 2 , 1 2 ) by the transformation T ( u , w ) = u υ w , samples T ( u , w ) S t ( 0 , I , υ ) of the standard St distribution can be obtained from that [42], and then the linear transformation t = μ + σ T ( u , w ) gives us t S t ( μ , σ , υ ) .

4. Method

4.1. DP-St GVAE

In this section, we will introduce the heavy-tailed priors based on Bayesian nonparametric methods in detail, which is DP construction for St distribution priors. Utilizing the augment of the prior distribution complexity, latent variables can be stimulated to learn more features, and more expressive latent representation can be obtained to reduce the reconstruction error.
Considering the St distribution as the prior distribution of latent variables, the bilateral tail is thicker, which can cover more latent features with higher probability, compared with the most commonly used Gaussian distribution. As the features increase, we have to allow higher dimensional latent representation to pertain to the nodes. Therefore, we hope to use DP to generate the mean parameter μ and scale parameter σ of the St distribution. This method allows us to infer which factors have significant effects, and the dimensions of latent variables. For the last parameter, dof parameter υ > 0 controls the tail shape of the pdf curve of St distribution, with the increase of dof υ , the curve of the latent variables changes gradually from the heavy-tailed Cauchy distribution to the Gaussian distribution, which makes us find the curve that fits the observed data most in the procedure of learning the dof parameter.
In Figure 3, we show the framework of the DP-St GVAE model in detail. First, we assume that the observed data, adjacency matrix A R N × N and attribute matrix X R N × C , are used as the inputs of encoder to generate the latent variables Z R N × K , where the N observations come from a graph undirected attributed graph G ( V , E , X ) . V is the node set { v 1 , v 2 , . . . , v N } , and E is a unweighted edge set, e i j = 1 means there is a undirected link between node v i and v j . We put a St distribution over the latent space, and place the outputs of GVAE’s encoder over the mean, scale, and dof parameters of St distribution:
q ϕ ( Z A , X ) = S t ( f 1 , ϕ ( A , X ) , f 2 , ϕ ( A , X ) , f 3 , ϕ ( A , X ) ) ,
where f 1 , ϕ , f 2 , ϕ , f 3 , ϕ denote the neural networks with parameter ϕ .
Second, we replace the standard Gaussian prior with a DP-St distribution. The DP is composed of a discrete base distribution G 0 and a concentration parameter. To explicitly illustrate the base distribution G 0 , the stick-breaking process is a good choice. It generates unbounded sticks b n k [ 0 , 1 ] independently from a beta distribution:
p ( b n k β ) = B e t a ( b n k 1 , β ) .
From the procedure above, we obtain an infinite dimensional vector with components π k ( b ) = b k l = 1 k 1 ( 1 b l 1 ) , which is used as the initialization of mean and scale parameters in St prior of each node, and the dof parameter is initialed from a Uniform distribution:
p ( Z n β ) = S t ( Z n π 1 , n ( β ) , π 2 , n ( β ) , U n i ( α ) ) .
In this way, we generate a latent representation of infinite dimensions for each node, and the DP allows latent variables to adapt complexity in the learning process. Moreover, it enables almost all nodes to seek out the most powerful features, reduced the error caused by the traditional allocation mode of weak discrimination. Finally, we make full use of the information inside the latent representation, and reconstruct a graph through the function g ( · ) with parameter θ :
p θ ( A Z ) = B e r n ( A g θ ( Z ) ) .

4.2. Variational Inference for DP Prior

We find that there is no direct method to deal with the posterior distribution of latent variables in the condition of DP-St prior. Variational inference opens up our train of thought and leads to a deterministic way for the approximate calculation of the likelihoods and posteriors. We denote the observed adjacency matrix A and attribute X as W = { A , X } , the hyperparameter p s i , the log posterior distribution of latent variable Z is:
l o g p ( Z W , ψ ) = l o g p ( W , Z ψ ) l o g p ( W ψ ) .
However, we don’t compute this posterior distribution directly due to the calculation of normalizing constant, and through some transformation, we can obtain the log marginal probability of the observation W:
l o g p ( W ψ ) = l o g p ( W , Z ψ ) d Z .
This equation is difficult to compute because of the dependency relationship between the latent variables and observed data.
In this work, we use the mean-field variational method, converting to an optimization problem from handling the posterior distribution problem. On this basis, the Kullback-Leibler (KL) divergence between posterior distribution q ϕ ( Z W ) and true posterior p ( Z ) need to be optimized in training:
K L [ q ϕ ( Z W ) p ( Z ) ] = E q [ l o g q ϕ ( Z W ) ] E q [ l o g p ( Z ) ] + l o g P ( W ) .
The upper bound of KL divergence between variational posterior q ( z ) = S t ( μ 1 , Σ 1 , υ 1 ) and St prior p ( z ) = S t ( μ 2 , Σ 2 , υ 2 ) is [43]:
K L [ q ( z ) p ( z ) ] 1 2 ln d e t Σ 2 d e t Σ 1 + n 2 ln υ 2 υ 1 + ln Γ ( υ 1 + n 2 ) ln Γ ( υ 1 2 ) ln Γ ( υ 2 + n 2 ) + ln Γ ( υ 2 2 ) υ 1 + n 2 [ ψ ( υ 1 + n 2 ) ψ ( υ 1 2 ) ] + υ 2 + n 2 ln ( 1 + 1 υ 2 t r ( Σ 2 1 Σ ˜ 1 ) + 1 υ 2 t r [ Σ 2 1 ( μ 1 μ 2 ) ( μ 1 μ 2 ) T ] ) ,
where ψ ( · ) is digamma function, Σ ˜ 1 = υ 1 υ 1 2 Σ 1 .
Minimizing the Equation (10) can be translated into maximizing the lower bound of the log-likelihood l o g p ( W ) . We finally achieve the objective function:
L ( μ , σ , υ , ϕ ) = E q ϕ ( Z A , X ) [ l o g p ( A Z ) ] K L [ q ϕ ( Z A , X ) p μ , σ , υ ( Z ) ] .

5. Experiments

We report the performance of our DP-St GVAE on two tasks—unsupervised link prediction and semi-supervised node classification. Our model has been put into effect on three benchmark citation datasets (Cora, Citeseer, and PubMed) [44], and the datasets are described in Table 1 in detail. Each node in these datasets represents literature, and has a unique attribute vector, the connection between literature indicates that there is a reference relationship between them.
In all experiments, the models are realized by TensorFlow, and ADAM is used for training. The encoder is composed of GCN [45] with parameter ϕ , and we let function g ( Z ) use activation function tanh for the transformation Z Z T . In our DP-St GVAE model, the sticking-breaking hyperparameter β is set to 10, and the Uniform distribution hyperparameter α is set to 10.

5.1. Link Prediction

To assess our model’s ability to perform link prediction tasks, we compared it with two algorithms: GVAE [3] and standard tGVAE [4]. Our model improves the problem that the standard St distribution as a prior can not fully discover the latent information for a large portion of nodes in networks, and in our experiments, it is applied to the network data, whose encoder and decoder are changed to use the same settings as our method, then we denoted it as St GVAE.
For this task, we will follow the settings of GVAE and make the encoder use GCN of two layers to conduct four groups of comparative experiments, the dimensions of latent space are 8, 16, 32, and 64, and the corresponding GCN hidden layer sizes are 16, 32, 64, and 128 respectively. All experiments use a learning rate of 0.01. When the learning rate is 0.1, it is difficult for the model to converge in training. When the learning rate is 0.001, the model converges very slowly, and its performance in learning tasks is lower than that in the learning rate of 0.01. The validation set and the test set contain 5% and 10% of the network edges. AUC (area under ROC curve) and AP (average accuracy) are used as the assessment criteria for the correct classification of edge types in the networks.
We show link prediction performance on three citation datasets in Table 2, Table 3 and Table 4, and our model is almost better than the contrastive approaches. As we can see from the experimental results, with the augment of the latent space dimension, the predictive performances of GVAE and St GVAE models are increasing, and our model presents a trend like a parabola. This is because our model can automatically select the most appropriate latent dimension, and the latent features under this dimension are conducive to the reconstruction of network structure. Moreover, there are St GVAE with a large predicted improvement span in the low-dimension interval, and GVAE with a uniform rising trend about the dimension. When the dimension of latent space is low, such as 8, the performance of St GVAE is the weakest among the three models, because, in the case of fewer latent features, the use of heavy-tailed St distribution can, to some extent, obscure the differences between various features. This indicates the importance of constructing the St distribution with DP because DP can help find the most critical features for each node among these few latent features. This ranking of importance can digest the disadvantage of heavy-tailed distribution in low-dimensional latent space.
Now, to understand the latent representation Z, we visualize its posterior probability, q ϕ ( Z ) , for GVAE and DP-St GVAE models on Cora and Citeseer datasets. With regard to the node selection for visualization, we first randomly select ten nodes with the most and the least number of latent features with high probability respectively, and then randomly choose 10 nodes in the remaining node-set. In addition, the dimensions of latent variables in the DP-St GVAE model are automatically determined, and for the convenience of comparison, we select similar dimensions for the GVAE model. From Figure 4a and Figure 5a, we can see that many nodes have multiple latent features with high probability, and the latent representation contains rich information and complex correlation. However, in Figure 4b and Figure 5b, the latent representation has higher sparsity, and the latent features of each node are more explicit. In the same dimension of latent space, the latent representation generated by GVAE will bring more disturbing information and severely restrict the nodes to find the most representative latent features. After adding a DP-St prior, the various information can be screened, and those useless features that have a negative impact on the target task can be discarded to produce appropriate latent representation.

5.2. Semi-Supervised Node Classification

For the semi-supervised node classification task, the goal is to make the label prediction for nodes without labels. So, in this task, we add the cross-entropy term of labels to the objective function (12). We make use of the random split partition method [46], randomly dividing all nodes in each dataset into three subsets, training set, validation set, and test set. The number of known labels has been separated into three groups, with 5, 10, and 20 labels per class. For this experiment, we compare our model with the GVAE and St GVAE models used in the link prediction task and added several related works in this task—Graph attention network (GAT) [47], and Graph convolutional neural network (GCN) [45].
Table 5, Table 6 and Table 7, respectively, show the results on the Cora, Citeseer, and PubMed datasets. The experimental results illustrate that our model achieves the optimal classification result on the PubMed dataset, because there is heavy-tailed degree distribution, and our model can precisely grasp the hidden information and obtain a more obvious accuracy improvement. On the Cora and Citesser datasets, our model is able to achieve more competitive results when the number of labels is known to be smaller. When the number of known labels is tiny, we can see that the node classification performance of GVAE is less than 50 % on the Citeseer dataset, suggesting a high error rate in the generated node representation. Furthermore, it is obvious that the St GVAE model improves the prediction accuracy by replacing the Gaussian distribution with the St distribution, indicating that the traditional Gaussian distribution leaves out some essential information for nodes, and the St distribution can avoid this problem. Finally, when we construct a DP prior for the St distribution, each node can capture the most important hidden information, which provides convenience for discovering the correlation between nodes, and thus the affiliation of nodes.

6. Conclusions

In this work, we propose a DP construction for Student’s t distribution prior, and combine it with GVAE. We solve the inaccuracy problem of modeling real-world networks by manually setting latent dimensions through the adaptive complexity of Bayesian nonparametric methods. In addition, the advantages and disadvantages of applying Gaussian distribution in traditional VAEs are analyzed at length, and we have illustrated that the application of Student’s t distribution can effectively improve the limitation of effective latent features. Here, our DP-St GVAE is used to the network structure data, and the combination of the infinite-dimensional nonparametric generation process and heavy-tailed distribution plays an important role in network presentation learning tasks. In the future, our work will proceed, and there are two primary directions: (i) applying the DP-St to deep neural networks so that the problem of future research will break the limitation of graph structure data; (ii) using other Bayesian nonparametric methods for further improving the prediction accuracy of network representation learning tasks.

Author Contributions

Conceptualization, Y.Z.; methodology, Y.Z.; software, Y.Z.; validation, Y.Z. and J.H.; formal analysis, Y.Z. and J.H.; investigation, Y.Z. and J.H.; resources, Y.Z. and J.H.; data curation, Y.Z.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.Z. and J.H.; visualization, Y.Z.; supervision, J.H.; project administration, J.H.; funding acquisition, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grant number 61572226, 61876069; Jilin Province Key Scientific and Technological Research and Development project under grant numbers 20180201067GX, 20180201044GX; Jilin Province Natural Science Foundation under grant No. 20200201036JC.

Data Availability Statement

Data is contained within the article. The data presented in this study are available in [Revisiting Semi-Supervised Learning with Graph Embeddings. Zhilin Yang, William W. Cohen, Ruslan Salakhutdinov. ICML 2016.], the computer code is available at [https://github.com/kimiyoung/planetoid (accessed on 9 February 2021)].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  2. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  3. Kipf, T.N.; Welling, M. Variational graph auto-encoders. arXiv 2016, arXiv:1611.07308. [Google Scholar]
  4. Abiri, N.; Ohlsson, M. Variational auto-encoders with Student’s t-prior. arXiv 2020, arXiv:2004.02581. [Google Scholar]
  5. Atanov, A.; Struminsky, K.; Welling, M.; Ashukha, A.; Vetrov, D. The deep weight prior. In Proceedings of the 7th International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  6. Bauer, M.; Mnih, A. Resampled priors for variational autoencoders. In Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics (AISTATS 2019), Naha, Okinawa, Japan, 16–18 April 2019; pp. 66–75. [Google Scholar]
  7. Klushyn, A.; Chen, N.; Kurle, R.; Cseke, B.; van der Smagt, P. Learning hierarchical priors in VAEs. arXiv 2019, arXiv:1905.04982. [Google Scholar]
  8. Tomczak, J.; Welling, M. VAE with a VampPrior. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS 2018), Canary Islands, Spain, 9–11 April 2018; pp. 1214–1223. [Google Scholar]
  9. Rosca, M.; Lakshminarayanan, B.; Mohamed, S. Distribution matching in variational inference. arXiv 2018, arXiv:1802.06847. [Google Scholar]
  10. Wang, Y.; Yao, Y.; Tong, H.; Xu, F.; Lu, J. A brief review of network embedding. J. Big Data Min. Anal. 2020, 2, 35–47. [Google Scholar] [CrossRef] [Green Version]
  11. Hamilton, W.L.; Ying, R.; Leskovec, J. Representation learning on graphs: Methods and applications. arXiv 2017, arXiv:1709.05584. [Google Scholar]
  12. Zhang, D.; Yin, J.; Zhu, X.; Zhang, C. Network representation learning: A survey. IEEE Trans. Big Data 2018, 6, 3–28. [Google Scholar] [CrossRef] [Green Version]
  13. Goodman, N.R. Statistical analysis based on a certain multivariate complex Gaussian distribution (an introduction). Ann. Math. Stat. 1963, 34, 152–177. [Google Scholar] [CrossRef]
  14. Novey, M.; Adali, T.; Roy, A. A complex generalized Gaussian distribution—Characterization, generation, and estimation. IEEE Trans. Signal Process. 2009, 58, 1427–1433. [Google Scholar] [CrossRef]
  15. Bazi, Y.; Bruzzone, L.; Melgani, F. Image thresholding based on the EM algorithm and the generalized Gaussian distribution. Pattern Recognit 2007, 40, 619–634. [Google Scholar] [CrossRef]
  16. Arulampalam, M.S.; Maskell, S.; Gordon, N.; Clapp, T. A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Signal Process 2002, 50, 174–188. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, L.; Schwing, A.G.; Lazebnik, S. Diverse and accurate image description using a variational auto-encoder with an additive Gaussian encoding space. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NeurIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5758–5768. [Google Scholar]
  18. Khan, R.A.; Kleinsteuber, M. Epitomic Variational Graph Autoencoder. arXiv 2020, arXiv:2004.01468. [Google Scholar]
  19. Pan, S.; Hu, R.; Long, G.; Jiang, J.; Yao, L.; Zhang, C. Adversarially regularized graph autoencoder for graph embedding. In Proceedings of the 27th International Joint Conference on Artificial Intelligence (IJCAI 2018), Stockholm, Sweden, 13–19 July 2018; pp. 2609–2615. [Google Scholar]
  20. Barron, A.R. Entropy and the central limit theorem. J. Ann. Probab. 1986, 14, 336–342. [Google Scholar] [CrossRef]
  21. Bryson, M.C. Heavy-tailed distributions: Properties and tests. J. Technometrics 1974, 16, 61–68. [Google Scholar] [CrossRef]
  22. Ord, J.K. The Discrete Student’s t Distribution. Ann. Math. Stat. 1968, 39, 1513–1516. [Google Scholar] [CrossRef]
  23. Hill, G.W. ACM Algorithm 395: Student’s t-distribution. Commun. ACM 1970, 13, 617–619. [Google Scholar] [CrossRef]
  24. Zhang, Z.; Lai, K.; Lu, Z.; Tong, X. Bayesian inference and application of robust growth curve models using Student’s t distribution. Struct. Equ. Model. A Multidiscip. J. 2013, 20, 47–78. [Google Scholar] [CrossRef]
  25. Adamic, L.A.; Lukose, R.M.; Puniyani, A.R.; Huberman, B.A. Search in power-law networks. Phys. Rev. E 2001, 64, 046135. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Adamic, L.A.; Huberman, B.A.; Barabási, A.L.; Albert, R.; Jeong, H.; Bianconi, G. Power-law distribution of the world wide web. Science 2000, 287, 2115. [Google Scholar] [CrossRef] [Green Version]
  27. Faloutsos, M.; Faloutsos, P.; Faloutsos, C. On power-law relationships of the internet topology. ACM SIGCOMM Comput. Commun. Rev. 1999, 29, 251–262. [Google Scholar] [CrossRef]
  28. Schmidt, M.N.; Morup, M. Nonparametric Bayesian modeling of complex networks: An introduction. J. IEEE Signal Process. Mag. 2013, 30, 110–128. [Google Scholar] [CrossRef] [Green Version]
  29. Gershman, S.J.; Blei, D.M. A tutorial on Bayesian nonparametric models. J. Math. Psychol. 2012, 56, 1–12. [Google Scholar] [CrossRef] [Green Version]
  30. Hjort, N.L.; Holmes, C.; Müller, P.; Walker, S.G. (Eds.) Bayesian Nonparametrics; Cambridge University Press: Cambridge, UK, 2010; Volume 28. [Google Scholar]
  31. Müller, P.; Quintana, F.A.; Jara, A.; Hanson, T. Bayesian Nonparametric Data Analysis; Springer: New York, NY, USA, 2015. [Google Scholar]
  32. Teh, Y.W. Dirichlet Process. Available online: https://www.stats.ox.ac.uk/~teh/research/npbayes/Teh2010a.pdf (accessed on 11 March 2021).
  33. Escobar, M.D. Estimating normal means with a Dirichlet process prior. J. Am. Stat. Assoc. 1994, 89, 268–277. [Google Scholar] [CrossRef]
  34. Blei, D.M.; Jordan, M.I. Variational inference for Dirichlet process mixtures. J. Bayesian Anal. 2006, 1, 121–143. [Google Scholar] [CrossRef]
  35. Hoffman, M.D.; Johnson, M.J. Elbo surgery: Yet another way to carve up the variational evidence lower bound. In Proceedings of the Workshop in Advances in Approximate Bayesian Inference, NIPS, Barcelona, Spain, 9 December 2016; Volume 1, p. 2. [Google Scholar]
  36. Xu, H.; Chen, W.; Lai, J.; Li, Z.; Zhao, Y.; Pei, D. On the necessity and effectiveness of learning the prior of variational auto-encoder. arXiv 2019, arXiv:1905.13452. [Google Scholar]
  37. Nalisnick, E.; Smyth, P. Stick-breaking variational autoencoders. In Proceedings of the 5th International Conference on Learning Representations (ICLR 2017), Toulon, France, 24–26 April 2017. [Google Scholar]
  38. Xue, Y.; Liao, X.; Carin, L.; Krishnapuram, B. Multi-Task Learning for Classification with Dirichlet Process Priors. J. Mach. Learn. Res. 2007, 8, 35–63. [Google Scholar]
  39. Lawrence, A.; Ek, C.H.; Campbell, N. DP-GP-LVM: A Bayesian Non-Parametric Model for Learning Multivariate Dependency Structures. In Proceedings of the 36th International Conference on Machine Learning (ICML 2019), Long Beach, CA, USA, 9–15 June 2019; pp. 3682–3691. [Google Scholar]
  40. Teh, Y.W.; Grür, D.; Ghahramani, Z. Stick-breaking construction for the Indian buffet process. In Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics (AISTATS 2007), San Juan, Puerto Rico, 21–24 March 2007; pp. 556–563. [Google Scholar]
  41. Kotz, S.; Nadarajah, S. Multivariate t-Distributions and Their Applications; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  42. Forbes, C.; Evans, M.; Hastings, N.; Peacock, B. Statistical Distributions; John Wiley & Sons: Hoboken, NJ, USA, 2011. [Google Scholar]
  43. Huang, Y.; Zhang, Y.; Chambers, J.A. A novel Kullback-Leibler divergence minimization-based adaptive student’s t-filter. IEEE Trans. Signal Process. 2019, 67, 5417–5432. [Google Scholar] [CrossRef]
  44. Sen, P.; Namata, G.; Bilgic, M.; Getoor, L.; Galligher, B.; Eliassi-Rad, T. Collective classification in network data. AI Mag. 2008, 29, 93. [Google Scholar] [CrossRef] [Green Version]
  45. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  46. Zhang, Y.; Pal, S.; Coates, M.; Ustebay, D. Bayesian graph convolutional neural networks for semi-supervised classification. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI 2019), Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 5829–5836. [Google Scholar]
  47. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2018, arXiv:1710.10903. [Google Scholar]
Figure 1. Several probability distribution function curves of common probability distribution. They share the same mean and scale parameters, and there are two St distribution with dof parameter 5 and 50 respectively.
Figure 1. Several probability distribution function curves of common probability distribution. They share the same mean and scale parameters, and there are two St distribution with dof parameter 5 and 50 respectively.
Futureinternet 13 00075 g001
Figure 2. Several cumulative distribution function curves of common probability distribution. They share the same mean and scale parameters, and there are two St distribution with dof parameter 5 and 50, respectively.
Figure 2. Several cumulative distribution function curves of common probability distribution. They share the same mean and scale parameters, and there are two St distribution with dof parameter 5 and 50, respectively.
Futureinternet 13 00075 g002
Figure 3. The architecture of the DP-St GVAE. The encoder is a graph convolutional neural network that integrates the network structure A with node feature X and, by extension, generates the node representation Z. The decoder reconstructs a graph from the inner product of the node representation Z.
Figure 3. The architecture of the DP-St GVAE. The encoder is a graph convolutional neural network that integrates the network structure A with node feature X and, by extension, generates the node representation Z. The decoder reconstructs a graph from the inner product of the node representation Z.
Futureinternet 13 00075 g003
Figure 4. Latent representation visualization for Cora dataset. The horizontal axis is the dimension of the hidden variable, and the vertical axis is the number of nodes. (a) the latent variable learned by GVAE. (b) the latent variable learned by DP-St GVAE. The darker the higher probability nodes possess the latent features.
Figure 4. Latent representation visualization for Cora dataset. The horizontal axis is the dimension of the hidden variable, and the vertical axis is the number of nodes. (a) the latent variable learned by GVAE. (b) the latent variable learned by DP-St GVAE. The darker the higher probability nodes possess the latent features.
Futureinternet 13 00075 g004
Figure 5. Latent representation visualization for Citeseer dataset. The horizontal axis is the dimension of the hidden variable, and the vertical axis is the number of nodes. (a) the latent variable learned by GVAE. (b) the latent variable learned by DP-St GVAE. The darker the higher probability nodes possess the latent features.
Figure 5. Latent representation visualization for Citeseer dataset. The horizontal axis is the dimension of the hidden variable, and the vertical axis is the number of nodes. (a) the latent variable learned by GVAE. (b) the latent variable learned by DP-St GVAE. The darker the higher probability nodes possess the latent features.
Futureinternet 13 00075 g005
Table 1. Details of the datasets.
Table 1. Details of the datasets.
CoraCiteseerPubMed
Nodes2708332719,717
Edges5429473244,338
Features14333703500
Labels763
Table 2. Link prediction performance comparison of different dimensions for Cora.
Table 2. Link prediction performance comparison of different dimensions for Cora.
8 Dimensions16 Dimensions32 Dimensions64 Dimensions
AUCAPAUCAPAUCAPAUCAP
GVAE91.091.191.492.692.793.693.394.4
St GVAE85.583.491.089.992.189.795.095.2
DP-St GVAE92.393.994.195.194.495.593.594.8
Table 3. Link prediction performance comparison of different dimensions for Citeseer.
Table 3. Link prediction performance comparison of different dimensions for Citeseer.
8 Dimensions16 Dimensions32 Dimensions64 Dimensions
AUCAPAUCAPAUCAPAUCAP
GVAE89.590.590.892.091.092.191.192.5
St GVAE80.680.090.989.995.595.095.596.1
DP-St GVAE95.596.296.997.297.397.696.997.2
Table 4. Link prediction performance comparison of different dimensions for PubMed.
Table 4. Link prediction performance comparison of different dimensions for PubMed.
8 Dimensions16 Dimensions32 Dimensions64 Dimensions
AUCAPAUCAPAUCAPAUCAP
GVAE92.792.994.494.795.695.696.596.5
St GVAE89.987.897.497.397.897.597.897.6
DP-St GVAE96.497.197.197.097.397.497.997.8
Table 5. Semi-supervised node classification prediction accuracy for Cora.
Table 5. Semi-supervised node classification prediction accuracy for Cora.
5 Labels10 Labels20 Labels
GAT70.4 ± 3.776.6 ± 2.879.9 ± 1.8
GCN70.0 ± 3.776.0 ± 2.279.8 ± 1.8
GVAE68.4 ± 2.470.3 ± 2.771.6 ± 1.4
St GVAE69.8 ± 2.471.0 ± 1.572.2 ± 2.0
DP-St GVAE75.1 ± 2.676.9 ± 2.779.3 ± 2.4
Table 6. Semi-supervised node classification prediction accuracy for Citeseer.
Table 6. Semi-supervised node classification prediction accuracy for Citeseer.
5 Labels10 Labels20 Labels
GAT56.7 ± 5.164.1 ± 3.367.6 ± 2.3
GCN58.5 ± 4.765.4 ± 2.667.8 ± 2.3
GVAE49.1 ± 1.852.1 ± 1.452.4 ± 2.5
St GVAE52.5 ± 1.353.1 ± 2.953.8 ± 1.5
DP-St GVAE59.2 ± 2.564.5 ± 2.567.9 ± 1.8
Table 7. Semi-supervised node classification prediction accuracy for PubMed.
Table 7. Semi-supervised node classification prediction accuracy for PubMed.
5 Labels10 Labels20 Labels
GAT68.0 ± 4.872.6 ± 3.676.4 ± 3.0
GCN69.7 ± 4.573.9 ± 3.477.5 ± 2.5
GVAE74.6 ± 1.974.9 ± 2.375.4 ± 3.0
St GVAE76.2 ± 2.677.5 ± 3.379.0 ± 3.9
DP-St GVAE76.4 ± 3.978.4 ± 1.979.8 ± 1.2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, Y.; Huang, J. Dirichlet Process Prior for Student’s t Graph Variational Autoencoders. Future Internet 2021, 13, 75. https://doi.org/10.3390/fi13030075

AMA Style

Zhao Y, Huang J. Dirichlet Process Prior for Student’s t Graph Variational Autoencoders. Future Internet. 2021; 13(3):75. https://doi.org/10.3390/fi13030075

Chicago/Turabian Style

Zhao, Yuexuan, and Jing Huang. 2021. "Dirichlet Process Prior for Student’s t Graph Variational Autoencoders" Future Internet 13, no. 3: 75. https://doi.org/10.3390/fi13030075

APA Style

Zhao, Y., & Huang, J. (2021). Dirichlet Process Prior for Student’s t Graph Variational Autoencoders. Future Internet, 13(3), 75. https://doi.org/10.3390/fi13030075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop