Next Article in Journal
Symbolic Regulator Sets for a Weakly Nonlinear Discrete Control System with a Small Step
Next Article in Special Issue
The Application of Directed Hyper-Graphs for Analysis of Models of Information Systems
Previous Article in Journal
The Fuzzy Logic in the Problems of Test Control of a Bypass Turbojet Engine Gas Generator
Previous Article in Special Issue
Conceptual Coverage Driven by Essential Concepts: A Formal Concept Analysis Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Network Embedding Algorithm Taking in Variational Graph AutoEncoder

Software College, Northeastern University, Shenyang 110169, China
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(3), 485; https://doi.org/10.3390/math10030485
Submission received: 11 January 2022 / Revised: 29 January 2022 / Accepted: 1 February 2022 / Published: 2 February 2022
(This article belongs to the Special Issue Information Systems Modeling Based on Graph Theory)

Abstract

:
Complex networks with node attribute information are employed to represent complex relationships between objects. Research of attributed network embedding fuses the topology and the node attribute information of the attributed network in the common latent representation space, to encode the high-dimensional sparse network information to the low-dimensional dense vector representation, effectively improving the performance of the network analysis tasks. The current research on attributed network embedding is presently facing problems of high-dimensional sparsity of attribute eigenmatrix and underutilization of attribute information. In this paper, we propose a network embedding algorithm taking in a variational graph autoencoder (NEAT-VGA). This algorithm first pre-processes the attribute features, i.e., the attribute feature learning of the network nodes. Then, the feature learning matrix and the adjacency matrix of the network are fed into the variational graph autoencoder algorithm to obtain the Gaussian distribution of the potential vectors, which more easily generate high-quality node embedding representation vectors. Then, the embedding of the nodes obtained by sampling this Gaussian distribution is reconstructed with structural and attribute losses. The loss function is minimized by iterative training until the low-dimension vector representation, containing network structure information and attribute information of nodes, can be better obtained, and the performance of the algorithm is evaluated by link prediction experimental results.

1. Introduction

Attribute networks are widely employed to model connections between entities in the real world, where the connected edges of nodes denote the relationships between objects and the attribute information of the nodes in the description about the nodes themselves. For example, the citation network [1,2] describes citations between articles by academic authors, and the node attribute information describes the abstract and subject of the article. Social networks [3,4] describe interpersonal relationships between people, the node also contains a user’s area, email, and release of words, pictures, and video, etc. Protein interaction network [5]: In addition to the topological information about protein interactions, each node also features attribute information about proteins in biology. Performing data analysis and data mining on these attribute networks, using network information for prediction and decision making, and discovering useful information latent in them, have strong academic and commercial value. Common network analysis tasks include node classification [6], node clustering [7], community detection [8,9,10], link prediction [11,12], and anomaly detection [13]. The networks are growing in size in the big data era, some reaching hundreds of millions of nodes, with high and sparse dimensions that not only have complex structures but are also rich in information. If such large-scale networks are directly applied to data mining, not only will they consume a large amount of data storage resources, but also the time complexity of computation is high, which is not conducive to algorithm deployment and application.
Attribute network embedding [14] enables the topological structure information of an attribute network and node attribute information to be effectively fused in a common latent representation space, which preserves the information of both. The schematic diagram of attribute network embedding is shown in Figure 1. The accuracy of the downstream network analysis task is further improved by enhancing the network node vector representation by employing information about the attributes of the nodes, based on the previous consideration of only the network structure information.
Graph neural networks (GNNs) [15] are also employed to perform network embedding. The graph convolutional network (GCN) [16] and the GraphSAGE model [17] employ graph convolutional networks, sampling nodes, and local neighbors for feature aggregation to learn the vector representation of nodes. The graph autoencoders (GAEs) [18] algorithm is an attribute network embedding algorithm that introduces a deep autoencoder. Based on the idea of the autoencoder, after using GCN encoding, the structural error is reconstructed in the decoder to learn the vector representation of the node. Variational graph autoencoders (VGAEs) [18] build on GAEs by using graph variational autoencoders to learn the embedding representation of attribute networks. The co-embedding model for static attributed networks (CSAN) [19] is also an attribute network embedding algorithm based on graph autoencoder, but in the attribute reconstruction part, the matrix product of the GCN encoder of the graph convolutional network and the multilayer perceptron MLP processing is employed to reconstruct the attribute errors.
The GAE algorithm is the attribute feature matrix employed as the GCN encoder input when extracting the attribute features of the network. The node attribute feature matrix is generally obtained by processing the descriptive information of the nodes using a bag-of-words model or a TF-IDF model. If the word list is too large, it tends to be characterized by high-dimensional sparsity and great algorithmic complexity. Moreover, these feature matrices are directly employed as encoder input, which does not reflect or fully employ the node attribute information. In addition, the GAE algorithm takes a single mapping approach to generate node embedding vectors in the encoder part, while the variogram autoencoder can generate the distribution of latent vectors, which is more capable of generating high-quality potential vectors after sampling, and the GAE algorithm does not consider the reconstruction loss of attribute information.
To address the above problems, this paper proposes a network embedding algorithm taking in a variational graph autoencoder (NEAT-VGA). The main idea of the NEAT-VGA algorithm is to first pre-process the attribute features, and the obtained attribute feature learning vectors are inputted into the variational graph autoencoder, together with the adjacency matrix, to complete the attribute network embedding task. Initially, the MHRWAE algorithm is employed to learn the attribute features of the nodes for the attribute network to obtain a low-dimensional vector representation of the node attributes. Then, the feature learning matrix and the adjacency matrix of the network are employed as input to the variogram autoencoder algorithm to generate a Gaussian distribution of the latent vectors, sample the Gaussian distribution to obtain the embedding vectors of the nodes, reconstruct the structural and attribute losses, and iteratively train for attribute network embedding. The final experimental results show that the proposed algorithm in this paper has better link prediction results compared with the benchmark comparison algorithm, i.e., the proposed network representation learning algorithm in this paper has better representation performance.

2. Related Works

The DeepWalk [20] algorithm is inspired by the Word2Vec [21], which treats the sequence of nodes obtained by random walking in the network as sentences in the text and the nodes in the sequence as words in the text, to obtain the embedding representation of the nodes by maximizing the probability of predicting the node’s context by employing the skip-gram model. The LINE [22] algorithm learns the embedding representation of a node to design and optimize the objective function with the definition of first-order and second-order similarities of the nodes. The first-order similarity refers to the existence of contiguous edges between nodes, while the second-order similarity refers to the common neighbor nodes between nodes. With the introduction of second-order similarity, LINE greatly alleviates the sparsity problem in the network, and the obtained node embedding representation can also maintain the local and global structural information of the network to the greatest extent. Based on DeepWalk, Node2Vec [23] introduces parameters to control the path of random walk, allowing the choice of whether to bias the random walk towards depth or breadth, and the sequence of nodes generated by the random walk, thus enabling better preservation of local and global structural information of the network.
With the development of deep learning in the fields of natural language processing and computer vision, many algorithms have also emerged that employ deep learning for network embedding, which employs neural networks to extract the structure of attribute networks and attribute information to combine to obtain the embedding representation of network nodes. The self-translation network embedding (STNE) [24] algorithm employs a sequence model to map a sequence of nodes’ content information to a sequence of nodes for network embedding. The attributed social network embedding (ASNE) [25] algorithm combines the information about the attributes of nodes in social networks during the learning node embedding. The attributed network representation learning (ANRL) [26] algorithm is an algorithm that employs a neighborhood-enhanced autoencoder and an attribute-aware skip-gram model, employed to capture node attribute and network structure information for attribute network embedding algorithm, which employs the node’s attribute information as input to the autoencoder to reconstruct the attribute information of neighboring nodes.

3. Methodology

The NEAT-VGA algorithm consists of the 5 following components:
(1)
Node Attribute Feature Learning: We employ the Metropolis–Hastings random walk (MHRW) algorithm [27] sampling, node sequence generation corpus, and Doc2Vec [28] model training to preprocess the attribute information, which is the attribute feature learning, named the MHRWAE algorithm. High-dimensional sparse attribute feature matrix learning can obtain low-dimensionality and better reflect the attribute feature learning matrix.
(2)
Attribute Network Encoder: The adjacency matrix and the attribute feature learning matrix are employed as input to the variational graph autoencoder, which takes GCN for encoding and maps the attribute network to a Gaussian distribution.
(3)
Structure Reconstruction Decoder: Sampling the Gaussian distribution to obtain a vector representation of the nodes, the structure decoder is employed to reconstruct the adjacency matrix of the network, i.e., the structural information of the network.
(4)
Attribute Reconstruction Decoder: Sampling the Gaussian distribution to obtain a vector representation of the nodes and using the attribute decoder to reconstruct the attribute feature learning matrix of the encoder input.
(5)
Loss Function Definition: We define a novel loss function, and the matrix reconstructed by the decoder is compared with the input information of the variogram autoencoder to construct the loss function, considering the effects of structural and attribute reconstruction. The final learning yields vector representations that are as good as possible in low-dimensional space for both structure and attributes.
The framework of the NEAT-VGA algorithm is shown in Figure 2.

3.1. Preliminaries

The goal of the NEAT-VGA algorithm is to learn the feature vector of each node based on the attribute network graph. Symbols and descriptions are shown in Table 1.

3.2. Aggregation-MHRWAE

The random walk-based network embedding algorithms sample node sequences with random walks, making the sampled sequences biased toward nodes of a large degree, and such algorithms do not consider the attribute information of nodes during the training process. The stitching relationship between structure vectors and attribute vectors is too complicated to fuse the structure and attributes well by these methods directly. In this paper, we propose the MHRWAE algorithm, which is a module of the NEAT-VGA algorithm. Its main function is to preprocess the attribute features of the nodes of the attribute network.
The main idea of the MHRWAE algorithm is that the embeddings of some nodes in the network should be more similar if they have similar attributes and the distribution of attributes of neighborhood nodes similarly. The framework of this algorithm is shown in Figure 3. The MHRWAE algorithm employs MHRW to generate a fixed-length sequence of nodes and the structural information of the network is utilized to generate a corpus based on the aggregation of node neighborhood relationships and node attribute information. The corpus is then trained using skip-gram with negative sampling to fuse the network structure and attribute information of the nodes to obtain a low-dimensional representation vector for each node.
(1)
The unbiased random wandering is performed with the MHRW algorithm to generate a sequence of nodes. the MHRW algorithm samples nodes without bias towards nodes of a larger degree, and the node sequences generated by the sampling reflect the connectivity between nodes, i.e., the structural information of the network. The MHRW algorithm sampling process is where the transfer probability of the MH algorithm is employed in the Random Walk algorithm to determine the transfer probability when the current node is sampled to a neighboring node. We assume that the probability distribution given in the MH algorithm is π v = 1 / | V | , and the election probability, Q i , j , is assumed to be Q i , j = 1 / d i , then the transition probability of sampling from node i to its neighbor node j, as shown in Equation (1), as follows:
p i , j = 1 d i × min ( 1 , d i d j ) , i j ,
where p i , i = 1 p i , j , d i and d j respectively, represent the degrees of nodes i and j. p i , i indicates the probability of node i staying at the current node.
(2)
Corpus generation: In the process of corpus generation by the MHRWAE algorithm, unlike DeepWalk, which employs multiple sets of nodes and node context nodes as a corpus, it employs multiple sets of nodes and node neighborhood node attributes as the corpus for SGNS training. The corpus is generated employing neighborhood node attribute aggregation, in which a sequence of nodes of a given length is generated, and for each node in the sequence, the attributes of the node and its neighborhood nodes are paired and added to the multiset, with each node in the sequence completing an iterative node attribute aggregation operation to form the final corpus.
(3)
The Doc2Vec model trains the corpus: Doc2Vec is a model that generates vector representations of documents, and the PV-DBOW method in the model enables SGNS. using the corpus as input, the Doc2Vec model is employed to train the corpus and generate a vector representation of each document, i.e., a vector representation of each node is obtained.

3.3. NEAT-VGA

The training process of the NEAT-VGA algorithm is shown in Algorithm 1.
Algorithm 1. NEAT-VGA
  Input: Attribute network G = (V, E, X), parameterα, m, g, l, b, t, n, p, q, a1, a2.
  Output: Attribute network node embedding vector Z
1:
for n = 1: p do
2:
  Pick v i .
3:
  Generate a vertex sequence ( v i , v i , , v l ) of length l by MHRW on network G.
4:
  for j in 1: l-t do
5:
    for r in 1: t do
6:
      Add vertex–context–feature pair ( v j , f j + r ) to multiset D.
7:
      Add vertex–context–feature pair ( v j + r , f j ) to multiset D.
8:
      Update D.
9:
    end for
10:
  end for
11:
end for
12:
Doc2Vec (D, epochs = p, m, b, a 1 ).
13:
Create F R | V | × m .
14:
Calculate matrix D ~ 1 2 A ~ D ~ 1 2 .
15:
Initialization parameter   θ = [ W ( 0 ) , W ( 1 ) , W ( 2 ) , W ( 3 ) , W ( 4 ) ] .
16:
for j = 1: q do
17:
  Calculate μ and σ according to Equations (5) and (6).
18:
  Calculate Z according to Equation (7)
19:
  Calculate L s L a   L K L according to Equations (9)–(12).
20:
  Calculate L according to Equation (13)
21:
  Update θ by Adam algorithm and learning rate = a 2 .
22:
end for
23:
return  Z R | V | × n .

3.3.1. Node Attribute Feature Learning

Pre-processing of node attribute information uses the MHRWAE algorithm, which consists of the MHRW algorithm sampling, node sequence generation corpus, and Doc2Vec model training. The nodes include the attributes of their neighboring nodes in their attributes; as a result, each node’s attributes are a sentence of attributes, and the words of the sentence are the node’s attributes and the neighboring node’s attributes.
For each node, there is a corresponding sentence describing the attributes of the node, and the addition of the neighborhood attribute makes the node with the more similar structure, the sentence with the more similar node attribute. Finally, the node numbers and attribute sentences are trained on the Doc2Vec model to obtain the attribute feature learning vector for each node.
For the attribute network G = ( V , E , X ) , according to the MHRWAE algorithm, given the vector dimension, m, of the attribute embedding representation of each node, the number of sampled node sequences, g, the length of each sampled node sequence, l, the neighborhood size, t, when node attributes are aggregated, and negative sampling parameters, such as the number, b, the learning rate, a 1 , and the number of training, p, can be obtained: F R | V | × m .

3.3.2. Attribute Network Encoder

This algorithm is implemented by introducing a variational self-encoder into the network structure using a GCN, which is used to capture information about the underlying attribute network. The GCN encoder considers the proximity of high-order nodes, solves the problem of network sparsity, and captures the non-linear information of the data through multi-layer, non-linear conversion. The graph convolutional network of each layer can employ Equation (2) to express the specific convolution process, as follows:
H ( l + 1 ) = f ( H ( l ) , A | W ( l ) ) = σ ( D ~ 1 2 A ~ D ~ 1 2 H ( l ) W ( l ) ) ,
where H ( l ) is the input of the l layer convolutional layer, H ( l + 1 ) is the output of the l layer convolutional layer A ~ = A + I , which means that the adjacency matrix of the network and the identity matrix are added together. D ~ is a diagonal matrix, and the value of A ~ the diagonal is related. The specific calculation equation is D ~ i , i = j A ~ i , j . σ ( · ) is a non-linear activation function, for example, R e l u ( x ) = max ( 0 , x ) .
In the encoding stage, the NEAT-VGA algorithm obtains the Gaussian distribution of each node and samples the Gaussian distribution to obtain the attribute network embedding representation matrix, Z R | V | × n . The Gaussian distribution can be uniquely determined by the mean, μ, and variance, δ, and the encoder part is the process of obtaining μ and δ through the GCN encoder.
The NEAT-VGA algorithm first employs two layers of GCN to share weights W ( 0 ) . W ( 1 ) after extracting features, one layer of GCN is added to obtain μ , log σ , and weight W ( 2 ) , W ( 3 ) . Therefore, the generation of each μ and σ is obtained by three-layered GCN encoding, but in the first two, the network encoder that shares parameters when layering is employed as an attribute network encoder.
The activation function of each layer of graph convolutional network is R e l u ( x ) = max ( 0 , x ) , the specific attribute network encoder process is shown in Equations (3)–(6), as follows:
H ( 1 ) = f r e l u ( F , A | W ( 0 ) ) ,
H ( 2 ) = f r e l u ( H ( 1 ) , A | W ( 1 ) ) ,
μ = H ( 3 ) = f r e l u ( H ( 2 ) , A | W ( 2 ) ) ,
log σ = H ( 4 ) = f r e l u ( H ( 2 ) , A | W ( 3 ) ) ,
After μ and σ are obtained, which is after the mean vector covariance matrix of the Gaussian distribution of each node, further sampling can be performed to calculate the attribute network embedding representation matrix, Z R | V | × n . The specific equation for obtaining Z through the mean and variance is shown in Equation (7), as follows:
Z = μ + ε σ ,
where ε obeying the standard N ( 0 , 1 ) , Z is obeying N ( μ , σ 2 ) .
Therefore, first, randomly sample one ε from the standard distribution, and then calculate Z through the equation; then, you can further update the gradient through backpropagation, and iteratively train to obtain new μ and σ .

3.3.3. Structure Reconstruction Decoder

In the attribute network encoding part, the attribute network embedding representation matrix is obtained, Z R | V | × n . The reconstructed adjacency matrix, B, can be obtained, Z, by predicting whether there is an edge between two nodes through the matrix, and the calculation is shown in Equation (8), as follows:
B = s i g m o i d ( Z × Z T ) ,
The function can map the input variables to between (0, 1). It is an activation function often employed in neural networks, and Z T is the transpose matrix of matrix Z.
For structural errors, L s , calculating the error between the reconstructed adjacency matrix and the original adjacency matrix, which is Z, predicts whether there is a connection between the pair of nodes in the original network, which can be regarded as a two-classification problem in machine learning, using two-class prediction. As the loss function, the calculation Equation (9) is as follows:
L s = d s ( A , B ) = ( y log y ^ + ( 1 y ) log ( 1 y ^ ) ) / N ,
where y represents the value of the element in row i and column j in A, and its value is 0 or 1, which means whether there is an edge between node i and node j in the original network. y ^ represents the value of the element in B. After s i g m o i d ( · ) function, its value is between 0 and 1. N represents the number of predicted connecting edges. The purpose of constructing the loss function in this way is to make the adjacency matrix reconstructed by the structural decoder more similar to the adjacency matrix of the original network.

3.3.4. Attribute Reconstruction Decoder

Using GCN as the attribute decoder, after the encoder obtains Z, connect a layer of GCN, output the reconstructed attribute matrix Y, and predict the attribute information of each node in the original network. The specific calculation is shown in Equation (10).
Y = f r e l u ( Z , A | W ( 4 ) ) ,
where, to calculate the node attribute reconstruction error, the output dimension of this layer of GCN is m.
For the attribute error L a , calculating the error between the reconstructed attribute matrix and the original attribute matrix can be defined by calculating the distance between F and the reconstructed matrix Y in the vector space L a , as shown in Equation (11).
L a = d a ( F , Y ) = | | F - Y | | F ,
where represents | | F - Y | | F , the Frobenius norm of the difference between the matrices.

3.3.5. Loss Function Definition

The loss function is jointly determined by the structure loss function, L s , the attribute loss function, L a , and the relative entropy, L K L .
We use KL divergence to calculate the relative entropy, and its calculation equation is shown in Equation (12), as follows:
L K L = KL [ q ( Z | X , A ) | | p ( Z ) ] ,
KL divergence is employed to describe the measurement of the difference between two probability distributions. q ( Z | X , A ) is the probability distribution calculated by the previous GCN encoder, and p ( Z ) is a priori distribution and standard Gaussian distribution.
To control the contribution of L s and L a to the total loss function, L, during the training process, α [ 0 , 1 ] parameters are introduced to the weights of L s and L a . The calculation equation L is shown in Equation (13), as follows:
L = ( 1 - α ) L s + α L a + L K L ,
The larger the value of α , the greater the contribution of attribute reconstruction loss to the total loss function. The minimization process is to make the reconstructed adjacency matrix and attribute matrix closer to the input of the GCN encoder and the probability distribution obtained by the GCN closer to the standard Gaussian distribution.
In the training process, the Adam algorithm is employed in the gradient descent method to minimize the above loss function, L. The learning rate, a 2 , is updated after iterative q times. Finally, the NEAT-VGA model obtains Z that can better reflect the structure information and attribute information of the network.

4. Experiments

For the experiments in this section, three datasets commonly employed in attribute network experiments—Cora (Avaliable at https://neusncp.com/api/view_dataset?dataset_id=173 (accessed on 1 January 2022)), CiteSeer (Avaliable at https://neusncp.com/a-pi/view_dataset?dataset_id=82 (accessed on 1 January 2022)), and PubMed (Avaliable at https://linqs-data.soe.ucsc.edu/public/Pubmed-Diabetes (accessed on 1 January 2022))—were chosen [29]. Some basic information about the datasets is shown in Table 2. Link prediction was employed as a downstream application to validate the performance of the model and to analyze the experimental results. An implementation of the proposed method has been made available at https://github.com/mingshuonie/NEAT-VGA (accessed on 1 January 2022).

4.1. Experimental Setting

The performance of the algorithm’s attribute network embedding is evaluated based on whether the edges are correctly predicted in the experiment. The specific testing procedure is that the edges in the network are randomly divided into three parts, with 85% of the edges serving as the training set and 5% and 10% of the edges removed as the validation and test sets. The network embedding algorithm was employed to generate a low-dimensional vector of network nodes, logistic regression was employed as the classifier for the binary classification problem of link prediction, the AUC (area under the curve) and AP (average precision) metrics were employed for evaluation, and the AUC and AP values of the test results were calculated. Both AUC and AP can evaluate the excellent or poor performance of the classifier. Their values range from 0 to 1, and the closer the value is to 1, the more effective the classifier is.
The experimental data sets are Cora, CiteSeer, and PubMed. AANE [30], GraphSAGE [17], ANRL-WAN [26], VGAE [18], and CSAN [19] algorithms are comparison algorithms. To ensure the fairness of the experiment, each parameter is set according to the reference. The NEAT-VGA algorithm has 200 iterations of feature learning and dimensionality of 128 dimensions for F. The number of iterations for the variational autoencoder part is 400, the learning rate is 0.05, the learning rate is 0.01, the dimensions of the encoding part are all set to 64 dimensions, and the reconstruction weight parameter, α, is 0.5.

4.2. Results and Analysis

The link prediction results obtained from the experiment are shown in Table 3. Furthermore, the histogram and the link prediction experiment results for the AUC and AP value comparison are shown in Figure 4 and Figure 5.
As can be seen from Table 3 and Figure 4 and Figure 5, the network embedding is carried out on the three citation network data sets, and the link prediction experiment is carried out. The AUC and AP values are all in AANE, GraphSAGE, ANRL, VGAE, CSAN, and NEAT-VGA algorithm is gradually improved. All the compared algorithms are attributed to network embedding algorithms, which utilize the attribute information of the nodes. The NEAT-VGA algorithm proposed in this paper achieves the highest AUC and AP values for link prediction experiments by combining node attribute feature learning and variational graph autoencoder, both of which are improved compared with the other algorithms.
The NEAT-VGA algorithm holds the best performance based on two metrics—AUC and AP. The best performance among the other compared algorithms is the CSAN algorithm, which uses multiple perceptrons for feature extraction of the attributes of the nodes and considers attribute reconstruction on the decoder. The experimental results of the NEAT-VGA algorithm and the CSAN algorithm are compared and analyzed. On the Cora data, the NEAT-VGA algorithm improved the AUC and AP metrics by 0.2% and 0.1% compared with the CSAN algorithm. On the CiteSeer data, the NEAT-VGA algorithm improved the AUC and AP metrics by 1.3% and 1% over the CSAN algorithm, which is a large improvement. On the PubMed data, the NEAT-VGA algorithm improved the AUC and AP metrics by 0.6% and 0.8% over the CSAN algorithm, which was the highest of the other algorithms.
The NEAT-VGA algorithm is an improvement on the graph autoencoder algorithm, using node attribute feature learning, which results in a significant improvement in link prediction compared with the VGAE algorithm, using a three-layered GCN variational encoder in the encoder and a loss function that reconstructs the structure and attributes jointly. On the Cora data, the AUC and AP metrics on the link prediction task for the NEAT-VGA algorithm improved by 8.0% and 9.1% compared to the VGAE algorithm. On the CiteSeer data, the NEAT-VGA algorithm improved the AUC and AP metrics on the link prediction task by 7.0% and 5.2%, compared with the VGAE algorithm. On the PubMed data, the NEAT-VGA algorithm improved the AUC and AP metrics on the link prediction task by 4.4% and 4.0%, compared with the VGAE algorithm. In summary, the improved NEAT-VGA algorithm had a very significant improvement in results.
The reasons why the NEAT-VGA algorithm can achieve the best results in all three datasets and improve the embedding performance of the attribute network are analyzed from two aspects.
First, the employ of attribute features of the nodes. Except for the NEAT-VGA algorithm and the CSAN algorithm, several other algorithms do not learn the attribute features and directly employ the node vector representations processed by the bag-of-words model and the TF-IDF model. while these vectors can retain and make employ of the node information of the attribute network, these vectors do not extract information about the attribute features of the nodes. the GCN and MLP employed by the CSAN algorithm. The NEAT-VGA algorithm, on the other hand, employs the MHRWAE algorithm to obtain an embedding representation of the node attributes that better captures the attribute information using unbiased random wandering and node neighborhood attribute aggregation.
Secondly, in the attribute network embedding part, the AANE algorithm is implemented by matrix decomposition of the attribute similarity matrix, which is less effective than other algorithms that employ graph neural networks to capture the graph structure and attribute information, such as NEAT-VGA, GraphSAGE, ANRL-WAN, VGAE, and CSAN. Compared with the other graph neural network algorithms, NEAT-VGA also employs network structures and node attribute information well when performing network embedding, employs learned node features as input, effectively extracts higher-order nonlinear information from the network, and considers the contribution of node attribute reconstruction and structure reconstruction to the total loss function.
Overall, the NEAT-VGA algorithm achieves excellent performance on all three citation networks through node-attribute feature learning, variational graph autoencoders, and loss functions that reconstruct the joint structure and attributes, enabling the model to perform well for network embedding and obtain low-dimensional vector representations that retain the structure and attribute information.

5. Conclusions

In this paper, we propose the NEAT-VGA algorithm as a solution for the problem that the attribute information matrix in the attribute network embedding is often high-dimensional and sparse, which increases the complexity of the algorithm and does not reflect the attribute information of the nodes well; moreover, the GAE algorithm uses a single mapping method in the encoder part and does not consider the attribute reconstruction. The NEAT-VGA algorithm employs the MHRWAE algorithm as a pre-processing method for node attribute feature extraction. The experimental results of link prediction on the citation attribute network dataset verify that the proposed algorithm outperforms other comparison algorithms. However, it is mainly appropriate for the static network and lacks the mining of the dynamic change characteristics of the network, which is the future research direction.

Author Contributions

Conceptualization, D.C., H.Z. and D.W.; formal analysis, M.N. and Z.W.; funding acquisition, D.C. and D.W.; methodology, M.N. and H.Z.; project administration, D.C., M.N. and D.W.; resources, M.N.; software, H.Z.; supervision, D.C. and D.W.; validation, D.C.; visualization, M.N. and Z.W.; writing—original draft, M.N., H.Z. and Z.W.; writing—review and editing, D.C. and D.W. All authors have read and agreed to the published version of the manuscript.

Funding

The paper does not receive any fundings.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We thank anonymous reviewers for their careful reading and valuable comments, which are all valuable and helpful for revising and improving our paper, as well as the important guiding significance to our research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Asatani, K.; Mori, J.; Ochi, M.; Sakata, I. Detecting trends in academic research from a citation network using network representation learning. PLoS ONE 2018, 13, e0197260. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Huang, L.; Jiang, B.; Lv, S.; Liu, Y.; Li, D. Survey on deep learning based recommender systems. Chin. J. Comput. 2018, 41, 1619–1647. [Google Scholar]
  3. Su, S.; Sun, L.; Zhang, Z.; Li, G.; Qu, J. MASTER: Across Multiple social networks, integrate Attribute and STructure Embedding for Reconciliation. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 3863–3869. [Google Scholar]
  4. Wu, X.D.; Li, Y.; Li, L. Influence analysis of online social networks. Jisuanji Xuebao/Chin. J. Comput. 2014, 37, 735–752. [Google Scholar] [CrossRef]
  5. Haiyan Hong, W.L. Link Prediction Algorithm in Protein-Protein Interaction Network Based on Spatial Mapping. Comput. Sci. 2016, S1, 413–417, 434. [Google Scholar]
  6. Wang, X.; Cui, P.; Wang, J.; Pei, J.; Zhu, W.; Yang, S. Community preserving network embedding. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-2017), San Francisco, CA, USA, 4–9 February 2017; pp. 203–209. [Google Scholar]
  7. Rozemberczki, B.; Davies, R.; Sarkar, R.; Sutton, C. Gemsec: Graph embedding with self clustering. In Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM-2019), Vancouver, BC, Canada, 27–30 August 2019; pp. 65–72. [Google Scholar]
  8. Blondel, V.D.; Guillaume, J.-L.; Lambiotte, R.; Lefebvre, E. Fast unfolding of communities in large networks. J. Stat. Mech. Theory Exp. 2008, 2008, P10008. [Google Scholar] [CrossRef] [Green Version]
  9. Chen, D.; Nie, M.; Wang, J.; Kong, Y.; Wang, D.; Huang, X. Community Detection Based on Graph Representation Learning in Evolutionary Networks. Appl. Sci. 2021, 11, 4497. [Google Scholar] [CrossRef]
  10. Huang, X.; Chen, D.; Ren, T.; Wang, D. A survey of community detection methods in multilayer networks. Data Min. Knowl. Discov. 2021, 35, 1–45. [Google Scholar] [CrossRef]
  11. Gao, S.; Denoyer, L.; Gallinari, P. Temporal link prediction by integrating content and structure information. In Proceedings of the 20th ACM International Conference on Information and Knowledge Management (CIKM-2011), Glasgow, UK, 24–28 October 2011; pp. 1169–1174. [Google Scholar]
  12. Wang, D.; Nie, M.; Chen, D.; Wan, L.; Huang, X. Node Similarity Index and Community Identification in Bipartite Networks. J. Internet Technol. 2021, 22, 673–684. [Google Scholar]
  13. Bhuyan, M.H.; Bhattacharyya, D.K.; Kalita, J.K. Network anomaly detection: Methods, systems and tools. IEEE Commun. Surv. Tutor. 2013, 16, 303–336. [Google Scholar] [CrossRef]
  14. Bandyopadhyay, S.; Lokesh, N.; Murty, M.N. Outlier aware network embedding for attributed networks. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI-2019), Honolulu, HI, USA, 27 January–1 February 2019; pp. 12–19. [Google Scholar]
  15. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  17. Hamilton, W.L.; Ying, R.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS-2017), Long Beach, CA, USA, 4–9 December 2017; pp. 1025–1035. [Google Scholar]
  18. Kipf, T.N.; Welling, M. Variational graph auto-encoders. arXiv 2016, arXiv:1611.07308. [Google Scholar]
  19. Meng, Z.; Liang, S.; Zhang, X.; McCreadie, R.; Ounis, I. Jointly learning representations of nodes and attributes for attributed networks. ACM Trans. Inf. Syst. 2020, 38, 1–32. [Google Scholar] [CrossRef] [Green Version]
  20. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD-2014), New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  21. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the 26th International Conference on Neural Information Processing Systems (NIPS-2013), Lake Tahoe, CA, USA, 5–10 December 2013; pp. 3111–3119. [Google Scholar]
  22. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  23. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD-2016), San Francisco, CA, USA, 24–27 August 2016; pp. 855–864. [Google Scholar]
  24. Liu, J.; He, Z.; Wei, L.; Huang, Y. Content to node: Self-translation network embedding. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1794–1802. [Google Scholar]
  25. Liao, L.; He, X.; Zhang, H.; Chua, T.-S. Attributed social network embedding. IEEE Trans. Knowl. Data Eng. 2018, 30, 2257–2270. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, Z.; Yang, H.; Bu, J.; Zhou, S.; Yu, P.; Zhang, J.; Ester, M.; Wang, C. ANRL: Attributed Network Representation Learning via Deep Neural Networks. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 3155–3161. [Google Scholar]
  27. Gjoka, M.; Kurant, M.; Butts, C.T.; Markopoulou, A. Walking in facebook: A case study of unbiased sampling of osns. In Proceedings of the 2010 Proceedings IEEE Infocom, San Diego, CA, USA, 14–19 March 2010; pp. 1–9. [Google Scholar]
  28. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the 31st International Conference on International Conference on Machine Learning (ICML-2014), Beijing, China, 21–26 June 2014; pp. 1188–1196. [Google Scholar]
  29. Sen, P.; Namata, G.; Bilgic, M.; Getoor, L.; Galligher, B.; Eliassi-Rad, T. Collective classification in network data. AI Mag. 2008, 29, 93. [Google Scholar] [CrossRef] [Green Version]
  30. Huang, X.; Li, J.; Hu, X. Accelerated attributed network embedding. In Proceedings of the 2017 SIAM International Conference on Data Mining (SIAM-2017), Houston, TX, USA, 27–29 April 2017; pp. 633–641. [Google Scholar]
Figure 1. Diagram of attributed network embedding.
Figure 1. Diagram of attributed network embedding.
Mathematics 10 00485 g001
Figure 2. The framework of the NEAT-VGA algorithm.
Figure 2. The framework of the NEAT-VGA algorithm.
Mathematics 10 00485 g002
Figure 3. The framework of the MHRWAE algorithm.
Figure 3. The framework of the MHRWAE algorithm.
Mathematics 10 00485 g003
Figure 4. AUC of link prediction task.
Figure 4. AUC of link prediction task.
Mathematics 10 00485 g004
Figure 5. AP of link prediction task.
Figure 5. AP of link prediction task.
Mathematics 10 00485 g005
Table 1. Description of mathematical symbols.
Table 1. Description of mathematical symbols.
SymbolDescription
G Attribute network diagram
V , E Node set, edge set
V , E Number of nodes, number of edges
A The adjacency matrix of the topological structure of the attribute network graph
d Attribute vector dimension of each node
X R | V | × d Attribute matrix
m Attribute feature learning vector dimension
F R | V | × m Node attribute feature learning matrix
g The number of sample node sequences
l The length of each sample node sequence
t Neighborhood size when node attributes are aggregated
n The vector dimension represented by each node embedding
μ Node embedding represents the mean of Gaussian distribution
σ Node embedding represents the variance of Gaussian distribution
p Training times of the node attribute feature learning part
q The number of training times of the image autoencoder attribute network embedding
Z R | V | × n Attribute network embedding representation matrix
Table 2. Citation network dataset.
Table 2. Citation network dataset.
DatasetNodesEdgesFeaturesClasses
Cora2708542914337
CiteSeer3327473237036
PubMed19,71744,3385003
Table 3. Link prediction task performances.
Table 3. Link prediction task performances.
MethodCoraCiteSeerPubMed
AUCAPAUCAPAUCAP
AANE0.7670.7200.7850.7650.7830.754
GraphSAGE0.7950.7630.8020.7910.8400.829
ANRL-WAN0.8320.8430.8670.8480.9180.897
VGAE0.9140.9030.9080.9200.9440.947
CSAN0.9850.9840.9500.9580.9800.977
NEAT-VGA(Ours)0.9870.9850.9720.9680.9860.985
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, D.; Nie, M.; Zhang, H.; Wang, Z.; Wang, D. Network Embedding Algorithm Taking in Variational Graph AutoEncoder. Mathematics 2022, 10, 485. https://doi.org/10.3390/math10030485

AMA Style

Chen D, Nie M, Zhang H, Wang Z, Wang D. Network Embedding Algorithm Taking in Variational Graph AutoEncoder. Mathematics. 2022; 10(3):485. https://doi.org/10.3390/math10030485

Chicago/Turabian Style

Chen, Dongming, Mingshuo Nie, Hupo Zhang, Zhen Wang, and Dongqi Wang. 2022. "Network Embedding Algorithm Taking in Variational Graph AutoEncoder" Mathematics 10, no. 3: 485. https://doi.org/10.3390/math10030485

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop